Commit Graph

803 Commits

Author SHA1 Message Date
VictorSanh 122d5c52ac distinguish was is not trained 2019-06-06 17:02:51 +02:00
VictorSanh 2647ac3294 forgot bertForPreTraining 2019-06-06 16:57:40 +02:00
VictorSanh cf44d98392 Add more examples to BERT models for torchhub 2019-06-06 16:36:02 +02:00
Thomas Wolf 2a329c6186
Merge pull request #651 from huggingface/gpt_torchhub
Add GPT* compatibility to torchhub
2019-05-31 14:44:52 +02:00
VictorSanh 45d21502f0 update doc 2019-05-31 01:04:16 -04:00
VictorSanh 98f5c7864f decorelate dependencies + fix bug 2019-05-31 01:00:29 -04:00
VictorSanh c8bd026ef6 move dependecies list to hubconf 2019-05-31 00:36:58 -04:00
VictorSanh 19ef2b0a66 Fix typo in hubconf 2019-05-31 00:33:33 -04:00
VictorSanh d0f591051c gpt_hubconf 2019-05-31 00:28:10 -04:00
VictorSanh 4a210c9fc6 Move bert_hubconf to hubconfs 2019-05-31 00:28:00 -04:00
VictorSanh 0c5a4fe9c9 modify from_pretrained for OpenAIGPT 2019-05-31 00:27:18 -04:00
VictorSanh 372a5c1cee Hubconf doc - Specia case loading 2019-05-30 16:06:21 -04:00
Victor SANH 96592b544b
default in __init__s for classification BERT models (#650) 2019-05-30 15:53:13 -04:00
VictorSanh 4cda86b08f Update hubconf for torchhub: paths+examples+doc 2019-05-30 18:38:00 +00:00
Thomas Wolf 3fc63f126d
Merge pull request #598 from burcturkoglu/master
Updating learning rate with special warm up in examples
2019-05-10 13:48:12 +02:00
burcturkoglu 00c7fd2b79 Division to num_train_optimizer of global_step in lr_this_step is removed. 2019-05-09 10:57:03 +03:00
burcturkoglu fa37b4da77 Merge branch 'master' of https://github.com/huggingface/pytorch-pretrained-BERT 2019-05-09 10:55:24 +03:00
burcturkoglu 5289b4b9e0 Division to num_train_optimizer of global_step in lr_this_step is removed. 2019-05-09 10:51:38 +03:00
Thomas Wolf 701bd59b8b
Merge pull request #585 from huntzhan/master
Make the epsilon of LayerNorm configurable.
2019-05-08 16:56:38 +02:00
Thomas Wolf 303b5e2b92
Merge pull request #545 from ailzhang/cache_dir
move pytroch_pretrained_bert cache folder under same path as torch
2019-05-08 16:55:27 +02:00
Thomas Wolf 0198399d84
Merge pull request #570 from MottoX/fix-1
Create optimizer only when args.do_train is True
2019-05-08 16:07:50 +02:00
Thomas Wolf 50fa92c026
Merge pull request #571 from MottoX/patch-1
Fix documentation typo
2019-05-08 16:06:13 +02:00
huntzhan 101ab4dd8e Make the epsilon of LayerNorm configurable. 2019-05-06 00:26:21 +08:00
MottoX 18c8aef9d3 Fix documentation typo 2019-05-02 19:23:36 +08:00
MottoX 74dbba64bc Prepare optimizer only when args.do_train is True 2019-05-02 19:09:29 +08:00
Thomas Wolf 3ae8c8be1e
Merge pull request #562 from apappu97/roc_stories_lmlabels_fix
Small fix to remove shifting of lm labels during pre process of RocStories.
2019-05-01 11:20:17 +02:00
Thomas Wolf e89520175d
Merge pull request #564 from 8enmann/patch-2
Fix #537
2019-05-01 11:18:46 +02:00
Ben Mann 74f7906db4
Fix #537 2019-04-30 19:48:22 -07:00
Aneesh Pappu 365fb34c6c small fix to remove shifting of lm labels during pre process of roc stories, as this shifting happens interanlly in the model 2019-04-30 13:53:04 -07:00
Thomas Wolf 2dee86319d
Merge pull request #527 from Mathieu-Prouveur/fix_value_training_loss
Update example files so that tr_loss is not affected by args.gradient…
2019-04-30 11:12:55 +02:00
Mathieu Prouveur 87b9ec3843 Fix tr_loss rescaling factor using global_step 2019-04-29 12:58:29 +02:00
Ailing Zhang 3963d57c89 move pytroch_pretrained_bert cache folder under same path as torch 2019-04-27 11:09:11 -07:00
thomwolf b832d5bb8a Release: 0.6.2 2019-04-25 21:37:47 +02:00
Thomas Wolf e6cf62d499
Merge pull request #488 from dhpollack/fix_multichoice
fixed BertForMultipleChoice model init and forward pass
2019-04-25 21:04:16 +02:00
Thomas Wolf 1cc1c3c344
Merge pull request #533 from lukovnikov/master
Docs for new learning rate code
2019-04-25 21:02:35 +02:00
Thomas Wolf dee8af4e46
Merge pull request #518 from huggingface/schedules_in_examples
Fix training schedules in examples to match new API
2019-04-25 21:01:04 +02:00
lukovnikov 56a47ce2b7 - replaced OpenAIGPTAdam with OpenAIAdam in docs 2019-04-25 16:05:28 +02:00
lukovnikov 331a46ff04 - replaced OpenAIGPTAdam with OpenAIAdam in docs 2019-04-25 16:04:37 +02:00
lukovnikov 704037ad51 - updated docs for new LR API
- added some images for illustration
- updated comments in optimization
2019-04-25 15:59:39 +02:00
Thomas Wolf d76a57b0ba
Merge pull request #506 from ailzhang/hubconf
Hubconf
2019-04-24 20:59:21 +02:00
thomwolf 80f995a141 revert BertForMultipleChoice linear classifier 2019-04-24 16:51:54 +02:00
Mathieu Prouveur ed8fad7390 Update example files so that tr_loss is not affected by args.gradient_accumulation_step 2019-04-24 14:07:00 +02:00
thomwolf d94c6b0144 fix training schedules in examples to match new API 2019-04-23 11:17:06 +02:00
Thomas Wolf c36cca075a
Merge pull request #515 from Rocketknight1/master
Fix --reduce_memory in finetune_on_pregenerated
2019-04-23 10:30:23 +02:00
Thomas Wolf 99e02c3415
Merge pull request #512 from cynthia/master
Fix indentation weirdness in GPT-2 example.
2019-04-23 10:29:01 +02:00
Thomas Wolf 98cb7b2c51
Merge pull request #445 from lukovnikov/master
Learning rate schedules improvement + extension
2019-04-23 10:27:38 +02:00
Matthew Carrigan b8e2a9c584 Made --reduce_memory actually do something in finetune_on_pregenerated 2019-04-22 14:01:48 +01:00
Matt af8a0384fc
Merge pull request #1 from huggingface/master
Pulling commits from main repo
2019-04-22 13:56:47 +01:00
Sangwhan Moon 14b1f719f4 Fix indentation weirdness in GPT-2 example. 2019-04-22 02:20:22 +09:00
lukovnikov 69850b4011 python 2 compat 2019-04-21 14:02:38 +02:00