- 22 Feb, 2021 1 commit
-
-
Stas Bekman authored
-
- 18 Feb, 2021 1 commit
-
-
Stas Bekman authored
* memory tracker metrics * go back to eval for somewhat consistency * handle no-gpu case * deal with stackable eval calls * restore callback order * style * simplify the API * add test * docs * consistently use eval_ prefix * improve docs * Update src/transformers/trainer_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * rename method * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 16 Feb, 2021 1 commit
-
-
Zhang Cheng authored
-
- 15 Feb, 2021 2 commits
-
-
Suraj Patil authored
* move old s2s scripts to legacy * add the tests back * proper rename * restore * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas@stason.org> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Stas Bekman authored
* fix run_seq2seq.py; porting DeepSpeed tests to it * unrefactor * defensive programming * defensive programming 2 * port the rest of the trainer tests * style * a cleaner scripts dir finder * cleanup
-
- 12 Feb, 2021 1 commit
-
-
Suraj Patil authored
* fix rouge metrics and task specific params * fix typo * round metrics * typo * remove task_specific_params
-
- 09 Feb, 2021 1 commit
-
-
Suraj Patil authored
* add do_predict, pass eval_beams durig eval * update help * apply suggestions from code review
-
- 08 Feb, 2021 5 commits
-
-
Stas Bekman authored
-
Stas Bekman authored
* deepspeed bug fixes and tests * manual wrap?
-
Olivier authored
* replace -100 token ids with the tokenizer pad_id for compute_metrics * fixed typo for label_ids
-
Stas Bekman authored
-
Stas Bekman authored
-
- 05 Feb, 2021 2 commits
-
-
Stas Bekman authored
* make executable * make executable * same for the template * cleanup
-
Suraj Patil authored
* add prepare_decoder_input_ids_from_labels in s2s models * support lbl smoothing and enc/emb freezing * fix freezing * use pad_token_id from config * remove embed freezing and add warning * prepare decoder_input_ids inside DataCollatorForSeq2Seq
-
- 01 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
* Remove subclass for sortish sampler * Use old Seq2SeqTrainer in script * Styling
-
Stas Bekman authored
-
- 29 Jan, 2021 1 commit
-
-
Stas Bekman authored
-
- 28 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 27 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 26 Jan, 2021 2 commits
-
-
Magdalena Biesialska authored
-
Andrea Cappelli authored
* Pad to 8x for fp16 multiple choice example (#9752) * Pad to 8x for fp16 squad trainer example (#9752) * Pad to 8x for fp16 ner example (#9752) * Pad to 8x for fp16 swag example (#9752) * Pad to 8x for fp16 qa beam search example (#9752) * Pad to 8x for fp16 qa example (#9752) * Pad to 8x for fp16 seq2seq example (#9752) * Pad to 8x for fp16 glue example (#9752) * Pad to 8x for fp16 new ner example (#9752) * update script template #9752 * Update examples/multiple-choice/run_swag.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update examples/question-answering/run_qa.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update examples/question-answering/run_qa_beam_search.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * improve code quality #9752 Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 25 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Auto-resume training from checkpoint * Update examples/text-classification/run_glue.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Roll out to other examples Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 22 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Fixes to run_seq2seq and instructions * Add more defaults for summarization
-
- 21 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Fix memory regression in Seq2Seq example * Fix test and properly deal with -100 * Easier condition with device safety * Patch for MBartTokenzierFast
-
- 19 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
* New run_seq2seq script * Add tests * Mark as slow * Update examples/seq2seq/run_seq2seq.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/data/data_collator.py Co-authored-by:
Suraj Patil <surajp815@gmail.com> * Update src/transformers/data/data_collator.py Co-authored-by:
Suraj Patil <surajp815@gmail.com> * Address review comments Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Suraj Patil <surajp815@gmail.com>
-
Sylvain Gugger authored
-
- 15 Jan, 2021 1 commit
-
-
Stas Bekman authored
-
- 14 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Upstream (and rename) sortish sampler * Use proper sampler * Update src/transformers/trainer_pt_utils.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 13 Jan, 2021 1 commit
-
-
Stas Bekman authored
* deepspeed integration * style * add test * ds wants to do its own backward * fp16 assert * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * style * for clarity extract what args are being passed to deepspeed * introduce the concept of self.wrapped_model * s/self.wrapped_model/self.model_wrapped/ * complete transition to self.wrapped_model / self.model * fix * doc * give ds its own init * add custom overrides, handle bs correctly * fix test * clean up model_init logic, fix small bug * complete fix * collapse --deepspeed_config into --deepspeed * style * start adding doc notes * style * implement hf2ds optimizer and scheduler configuration remapping * oops * call get_num_training_steps absolutely when needed * workaround broken auto-formatter * deepspeed_config arg is no longer needed - fixed in deepspeed master * use hf's fp16 args in config * clean * start on the docs * rebase cleanup * finish up --fp16 * clarify the supported stages * big refactor thanks to discovering deepspeed.init_distributed * cleanup * revert fp16 part * add checkpoint-support * more init ds into integrations * extend docs * cleanup * unfix docs * clean up old code * imports * move docs * fix logic * make it clear which file it's referring to * document nodes/gpus * style * wrong format * style * deepspeed handles gradient clipping * easier to read * major doc rewrite * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * docs * switch to AdamW optimizer * style * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * clarify doc Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 05 Jan, 2021 1 commit
-
-
Patrick von Platen authored
* first try * remove old template * finish bart * finish mbart * delete unnecessary line * init pegasus * save intermediate * correct pegasus * finish pegasus * remove cookie cutter leftover * add marian * finish blenderbot * replace in file * correctly split blenderbot * delete "old" folder * correct "add statement" * adapt config for tf comp * correct configs for tf * remove ipdb * fix more stuff * fix mbart * push pegasus fix * fix mbart * more fixes * fix research projects code * finish docs for bart, mbart, and marian * delete unnecessary file * correct attn typo * correct configs * remove pegasus for seq class * correct peg docs * correct peg docs * finish configs * further improve docs * add copied from statements to mbart * fix copied from in mbart * add copy statements to marian * add copied from to marian * add pegasus copied from * finish pegasus * finish copied from * Apply suggestions from code review * make style * backward comp blenderbot * apply lysandres and sylvains suggestions * apply suggestions * push last fixes * fix docs * fix tok tests * fix imports code style * fix doc
-
- 23 Dec, 2020 1 commit
-
-
Sylvain Gugger authored
-
- 22 Dec, 2020 3 commits
-
-
Sylvain Gugger authored
-
Manuel Romero authored
-
Sylvain Gugger authored
* Add label smoothing in Trainer * Add options for scheduler and Adafactor in Trainer * Put Seq2SeqTrainer in the main lib * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Address review comments and adapt scripts * Documentation * Move test not using script to tests folder Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 20 Dec, 2020 1 commit
-
-
Stas Bekman authored
-
- 18 Dec, 2020 2 commits
-
-
Sylvain Gugger authored
* Add timing inside Trainer * Fix tests * Add n_objs for train * Sort logs
-
Stas Bekman authored
-
- 17 Dec, 2020 1 commit
-
-
Stas Bekman authored
-
- 16 Dec, 2020 1 commit
-
-
Sylvain Gugger authored
* Experimental stupport for fairscale ShardedDDP * Add import error if fairscale not available * Address review comments * Fix seq2seq trainer
-
- 15 Dec, 2020 1 commit
-
-
Stas Bekman authored
update README with good news that the leak fix has been applied to pytorch-1.7.1.
-