- 18 Jul, 2020 2 commits
-
-
Sam Shleifer authored
-
Nathan Raw authored
Co-authored-by:Sam Shleifer <sshleifer@gmail.com>
-
- 17 Jul, 2020 1 commit
-
-
Sam Shleifer authored
-
- 16 Jul, 2020 1 commit
-
-
Sam Shleifer authored
-
- 15 Jul, 2020 2 commits
-
-
Sam Shleifer authored
-
Sam Shleifer authored
-
- 14 Jul, 2020 1 commit
-
-
Boris Dayma authored
* docs(wandb): explain how to use W&B integration fix #5262 * Also mention TensorBoard Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 10 Jul, 2020 1 commit
-
-
Julien Chaumond authored
Co-Authored-By:
Suraj Patil <surajp815@gmail.com> Co-Authored-By:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 09 Jul, 2020 1 commit
-
-
Lysandre Debut authored
* Test XLA examples * Style * Using `require_torch_tpu` * Style * No need for pytest
-
- 08 Jul, 2020 1 commit
-
-
Ji Xin authored
* Add deebert code * Add readme of deebert * Add test for deebert Update test for Deebert * Update DeeBert (README, class names, function refactoring); remove requirements.txt * Format update * Update test * Update readme and model init methods
-
- 07 Jul, 2020 5 commits
-
-
Patrick von Platen authored
-
Sam Shleifer authored
improve unittests for finetuning, especially w.r.t testing frozen parameters fix freeze_embeds for T5 add streamlit setup.cfg
-
Patrick von Platen authored
[Almost all TF models] TF clean up: add missing CLM / MLM loss; fix T5 naming and keras compile (#5395) * add first version of clm tf * make style * add more tests for bert * update tf clm loss * fix tests * correct tf ner script * add mlm loss * delete bogus file * clean tf auto model + add tests * finish adding clm loss everywhere * fix training in distilbert * fix flake8 * save intermediate * fix tf t5 naming * remove prints * finish up * up * fix tf gpt2 * fix new test utils import * fix flake8 * keep backward compatibility * Update src/transformers/modeling_tf_albert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_auto.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_electra.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_roberta.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_mobilebert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_auto.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_bert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/modeling_tf_distilbert.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * apply sylvains suggestions Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Suraj Patil authored
* add SquadDataset * add DataCollatorForQuestionAnswering * update __init__ * add run_squad with trainer * add DataCollatorForQuestionAnswering in __init__ * pass data_collator to trainer * doc tweak * Update run_squad_trainer.py * Update __init__.py * Update __init__.py Co-authored-by:
Julien Chaumond <chaumond@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Shashank Gupta authored
* Added data collator for XLNet language modeling and related calls Added DataCollatorForXLNetLanguageModeling in data/data_collator.py to generate necessary inputs for language modeling training with XLNetLMHeadModel. Also added related arguments, logic and calls in examples/language-modeling/run_language_modeling.py. Resolves: #4739, #2008 (partially) * Changed name to `DataCollatorForPermutationLanguageModeling` Changed the name of `DataCollatorForXLNetLanguageModeling` to the more general `DataCollatorForPermutationLanguageModelling`. Removed the `--mlm` flag requirement for the new collator and defined a separate `--plm_probability` flag for its use. CTRL uses a CLM loss just like GPT and GPT-2, so should work out of the box with this script (provided `past` is taken care of similar to `mems` for XLNet). Changed calls and imports appropriately. * Added detailed comments, changed variable names Added more detailed comments to `DataCollatorForPermutationLanguageModeling` in `data/data_collator.py` to explain working. Also cleaned up variable names and made them more informative. * Added tests for new data collator Added tests in `tests/test_trainer.py` for DataCollatorForPermutationLanguageModeling based on those in DataCollatorForLanguageModeling. A specific test has been added to check for odd-length sequences. * Fixed styling issues
-
- 06 Jul, 2020 1 commit
-
-
Lysandre Debut authored
-
- 01 Jul, 2020 3 commits
-
-
Sylvain Gugger authored
* Cleanup and unify Trainer/TFTrainer * Forgot to adapt TFTrainingArgs * In tf scripts n_gpu -> n_replicas * Update src/transformers/training_args.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Address review comments * Formatting * Fix typo Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Sam Shleifer authored
-
Sylvain Gugger authored
-
- 30 Jun, 2020 4 commits
-
-
Hong Xu authored
Otherwise, if label is not specified, the following error occurs: Traceback (most recent call last): File "run_ner.py", line 303, in <module> main() File "run_ner.py", line 101, in main model_args, data_args, training_args = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1])) File "/home/user/anaconda3/envs/bert/lib/python3.7/site-packages/transformers/hf_argparser.py", line 159, in parse_json_file obj = dtype(**inputs) TypeError: __init__() missing 1 required positional argument: 'labels'
-
Sam Shleifer authored
-
Kevin Canwen Xu authored
-
MichaelJanz authored
* Fix the bug 'Attempted relative import with no known parent package' when using the bertabs example. Also change the used model from bertabs-finetuned-cnndm, since it seems not be accessible anymore * Update run_summarization.py Co-authored-by:Kevin Canwen Xu <canwenxu@126.com>
-
- 29 Jun, 2020 2 commits
-
-
Sam Shleifer authored
-
Patrick von Platen authored
* first doc version * add benchmark docs * fix typos * improve README * Update docs/source/benchmarks.rst Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * fix naming and docs Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 28 Jun, 2020 2 commits
-
-
Sam Shleifer authored
* all save_pretrained methods mkdir if not os.path.exists
-
Suraj Patil authored
-
- 26 Jun, 2020 4 commits
-
-
Sam Shleifer authored
-
Sam Shleifer authored
-
Thomas Wolf authored
* remove references to old API in docstring - update data processors * style * fix tests - better type checking error messages * better type checking * include awesome fix by @LysandreJik for #5310 * updated doc and examples
-
Patrick von Platen authored
* improve plotting * better labels * fix time plot
-
- 25 Jun, 2020 3 commits
-
-
Lysandre Debut authored
-
Sam Shleifer authored
-
Sam Shleifer authored
-
- 24 Jun, 2020 5 commits
-
-
Victor SANH authored
* fix weirdness in roberta/bart for mnli trained checkpoints * black compliance * isort code check
-
Patrick von Platen authored
* add benchmark for all kinds of models * improved import * delete bogus files * make style
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Kevin Canwen Xu authored
* Fix PABEE division by zero error * patience=0 by default
-
- 23 Jun, 2020 1 commit
-
-
Sam Shleifer authored
-