Commit Graph

5250 Commits

Author SHA1 Message Date
Stas Bekman cbb2f75a16
[s2s tests] fix test_run_eval_search (#7297) 2020-09-21 14:00:40 -04:00
Suraj Patil 7a88ed6c2a
[model card] distlbart-mnli model cards (#7278) 2020-09-21 12:26:18 -04:00
Sylvain Gugger 63276b76d4
Fix #7284 (#7289) 2020-09-21 10:31:26 -04:00
Raphaël Bournhonesque 8d464374ba
Disable missing weight warning (#7282) 2020-09-21 09:14:48 -04:00
Stas Bekman 8ff88d25e9
[fsmt] rewrite SinusoidalPositionalEmbedding + USE_CUDA test fixes + new TranslationPipeline test (#7224)
* fix USE_CUDA, add pipeline

* USE_CUDA fix

* recode SinusoidalPositionalEmbedding into nn.Embedding subclass

was needed for torchscript to work - this is now part of the state_dict, so will have to remove these keys during save_pretrained

* back out (ci debug)

* restore

* slow last?

* facilitate not saving certain keys and test

* remove no longer used keys

* style

* fix logging import

* cleanup

* Update src/transformers/modeling_utils.py

Co-authored-by: Sam Shleifer <sshleifer@gmail.com>

* fix bug in max_positional_embeddings

* rename keys to keys_to_never_save per suggestion, improve the setup

* Update src/transformers/modeling_utils.py

Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>

Co-authored-by: Sam Shleifer <sshleifer@gmail.com>
Co-authored-by: Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
2020-09-21 09:13:35 -04:00
Dat Quoc Nguyen 67c4b0c517
Add model cards for new pre-trained BERTweet-COVID19 models (#7269)
Two new pre-trained models "vinai/bertweet-covid19-base-cased" and "vinai/bertweet-covid19-base-uncased" are resulted by further pre-training the pre-trained model "vinai/bertweet-base" on a  corpus of 23M COVID-19 English Tweets for 40 epochs.
2020-09-21 06:12:51 -04:00
Patrick von Platen 0cbe1139b1
Update README.md 2020-09-21 11:53:08 +02:00
Lysandre aae4edb5f0 Addressing review comment 2020-09-21 11:37:00 +02:00
Suraj Patil 43b9d93875
[example/glue] fix compute_metrics_fn for bart like models (#7248)
* fix compute_metrics_fn

* p.predictions -> preds

* apply suggestions
2020-09-21 05:34:20 -04:00
guillaume-be 39062d05f0
Fixed target_mapping preparation for XLNet when batch size > 1 (incl. beam search) (#7267) 2020-09-21 04:53:52 -04:00
Nadir El Manouzi 4b3e55bdcc
Add "Fine-tune ALBERT for sentence-pair classification" notebook to the community notebooks (#7255) 2020-09-21 04:25:22 -04:00
Stas Bekman 7cbf0f722d
examples/seq2seq/__init__.py mutates sys.path (#7194) 2020-09-20 16:54:42 -04:00
Manuel Romero a4faeceaed
Fix typo in model name (#7268) 2020-09-20 19:12:30 +02:00
Stas Bekman 47ab3e8262
@slow has to be last (#7251)
Found an issue when `@slow` isn't the last decorator (gets ignored!), so documenting this significance.
2020-09-20 09:17:29 -04:00
Stas Bekman 4f6e525742
model card improvements (#7221) 2020-09-19 17:02:05 -04:00
Stas Bekman eb074af75e
fsmt tiny model card + script (#7244) 2020-09-19 14:37:12 -04:00
Manuel Romero 1d90d0f386
Add title to model card (#7240) 2020-09-19 02:10:45 -04:00
Manuel Romero c9b7ef042f
Create README.md (#7239) 2020-09-19 02:09:29 -04:00
Sam Shleifer 83dba10b8f
[s2s] distributed_eval.py saves better speed info (#7242) 2020-09-18 15:46:01 -04:00
Dat Quoc Nguyen af2322c7a0
Add new pre-trained models BERTweet and PhoBERT (#6129)
* Add BERTweet and PhoBERT models

* Update modeling_auto.py

Re-add `bart` to LM_MAPPING

* Update tokenization_auto.py

Re-add `from .configuration_mobilebert import MobileBertConfig`
not sure why it's replaced by `from transformers.configuration_mobilebert import MobileBertConfig`

* Add BERTweet and PhoBERT to pretrained_models.rst

* Update tokenization_auto.py

Remove BertweetTokenizer and PhobertTokenizer out of tokenization_auto.py (they are currently not supported by AutoTokenizer.

* Update BertweetTokenizer - without nltk

* Update model card for BERTweet

* PhoBERT - with Auto mode - without import fastBPE

* PhoBERT - with Auto mode - without import fastBPE

* BERTweet - with Auto mode - without import fastBPE

* Add PhoBERT and BERTweet to TF modeling auto

* Improve Docstrings for PhobertTokenizer and BertweetTokenizer

* Update PhoBERT and BERTweet model cards

* Fixed a merge conflict in tokenization_auto

* Used black to reformat BERTweet- and PhoBERT-related files

* Used isort to reformat BERTweet- and PhoBERT-related files

* Reformatted BERTweet- and PhoBERT-related files based on flake8

* Updated test files

* Updated test files

* Updated tf test files

* Updated tf test files

* Updated tf test files

* Updated tf test files

* Update commits from huggingface

* Delete unnecessary files

* Add tokenizers to auto and init files

* Add test files for tokenizers

* Revised model cards

* Update save_vocabulary function in BertweetTokenizer and PhobertTokenizer and test files

* Revised test files

* Update orders of Phobert and Bertweet tokenizers in auto tokenization file
2020-09-18 13:16:43 -04:00
Patrick von Platen 9397436ea5
Create README.md 2020-09-18 16:52:00 +02:00
Patrick von Platen 7eeca4d399
Create README.md 2020-09-18 16:44:02 +02:00
Patrick von Platen 31516c776a
Update README.md 2020-09-18 16:37:14 +02:00
Patrick von Platen 4c14669a78
Update README.md 2020-09-18 16:35:11 +02:00
Yih-Dar 3a03bab9db
Fix a few countings (steps / epochs) in trainer_tf.py (#7175) 2020-09-18 09:28:56 -04:00
Stefan Schweter ee9eae4e06
token-classification: update url of GermEval 2014 dataset (#6571) 2020-09-18 06:18:06 -04:00
Julien Chaumond eef8d94d19 [model_cards]
We use ISO 639-1 cc @gentaiscool
2020-09-18 12:09:24 +02:00
Patrick von Platen afd6a9f827
Create README.md 2020-09-18 11:41:12 +02:00
Patrick von Platen 9f1544b9e0
Create README.md 2020-09-18 11:37:20 +02:00
Sameer Zahid 5c1d5ea667
Fixed typo in README (#7233) 2020-09-18 04:52:43 -04:00
Yuta Hayashibe 7719ecd19f
Fix a typo (#7225) 2020-09-18 04:23:33 -04:00
Manuel Romero 4a26e8ac5f
Create README.md (#7205) 2020-09-18 03:24:30 -04:00
Manuel Romero 94320c5b81
Add customized text to widget (#7204) 2020-09-18 03:24:23 -04:00
Manuel Romero 3aefb24b20
Create README.md (#7209) 2020-09-18 03:24:10 -04:00
Manuel Romero a22e7a8dd4
Create README.md (#7210) 2020-09-18 03:23:58 -04:00
Manuel Romero c028b26481
Create README.md (#7212) 2020-09-18 03:23:49 -04:00
Genta Indra Winata c7cdd7b4fd
Create README.md for indobert-lite-base-p1 (#7182) 2020-09-18 03:22:32 -04:00
Genta Indra Winata bfb9150b8f
Create README.md for indobert-lite-large-p1 (#7184)
* Create README.md

* Update README.md
2020-09-18 03:22:11 -04:00
Genta Indra Winata d193593403
Create README.md (#7183) 2020-09-18 03:21:54 -04:00
Genta Indra Winata e65d846674
Create README.md (#7185) 2020-09-18 03:21:39 -04:00
Genta Indra Winata e27d86d48d
Create README.md for indobert-large-p2 model card (#7181) 2020-09-18 03:21:28 -04:00
Genta Indra Winata 881c0783e9
Create README.md for indobert-large-p1 model card (#7180) 2020-09-18 03:21:16 -04:00
Genta Indra Winata e0d58a5c87
Create README.md (#7179) 2020-09-18 03:20:59 -04:00
Genta Indra Winata 1313a1d2a8
Create README.md for indobert-base-p2 (#7178) 2020-09-18 03:20:29 -04:00
tuner007 cf24f43e76
Create README.md (#7095)
Create model card for Pegasus QA
2020-09-18 03:19:45 -04:00
Sam Shleifer 67d9fc50d9
[s2s] remove double assert (#7223) 2020-09-17 18:32:31 -04:00
Stas Bekman edbaad2c5c
[model cards] fix metadata - 3rd attempt (#7218) 2020-09-17 16:57:06 -04:00
Stas Bekman 999a1c957a
skip failing FSMT CUDA tests until investigated (#7220) 2020-09-17 16:53:14 -04:00
Stas Bekman 51c4adf54c
[model cards] fix dataset yaml (#7216) 2020-09-17 15:29:39 -04:00
Sam Shleifer a5638b2b3a
[s2s] dynamic batch size with --max_tokens_per_batch (#7030) 2020-09-17 15:19:34 -04:00