- 08 Mar, 2021 3 commits
-
-
Sylvain Gugger authored
This reverts commit b35e7b68.
-
Sylvain Gugger authored
-
Stas Bekman authored
-
- 05 Mar, 2021 1 commit
-
-
Joakim Warholm authored
-
- 04 Mar, 2021 2 commits
-
-
Sylvain Gugger authored
* Rework TPU checkpointing in Trainer * Wraps the barrier in a dist test * Address review comments * Remove line
-
Philipp Schmid authored
* removed overwrites * remove default value for output_dir * adjusted typing
-
- 03 Mar, 2021 2 commits
-
-
Sylvain Gugger authored
* Fix gradient accumulation for SM Model Parallelism * Style and divide loss by grad accum steps
-
Stas Bekman authored
* remap classes to strings * missing new util * style * doc * move the autogenerated file * Trigger CI
-
- 01 Mar, 2021 1 commit
-
-
Patrick von Platen authored
* add encode labels function to tokenizer * start adding finetuning * init dropout * upload * correct convert script * apply changes * fix second typo * make first dummy training run * adapt convert script * push confg for comparison * remove conf * finish training * adapt data collator * add research folder * update according to fairseq feedback * some minor corrections * refactor masking indices a bit * some minor changes * clean tokenizer * finish clean-up * remove previous logic * update run script * correct training * finish changes * finish model * correct bug * fix training a bit more * add some tests * finish gradient checkpointing * finish example * correct gradient checkpointing * improve tokenization method * revert changes in tokenizer * revert general change * adapt fine-tuning * update * save intermediate test * Update README.md * finish finetuning * delete conversion script * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py * Update src/transformers/models/wav2vec2/processing_wav2vec2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * finish wav2vec2 script * finish wav2vec2 fine-tuning * finalize test * correct test * adapt tests * finish * remove test file Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 27 Feb, 2021 2 commits
-
-
Stas Bekman authored
* refactors * typo
-
Amog Kamsetty authored
* fixes * update resources * formatting * remove import * add log statement * use fstring * add period * Update src/transformers/integrations.py
-
- 25 Feb, 2021 1 commit
-
-
Sylvain Gugger authored
* Ass support for ZeRO-2/3 and ZeRO-offload in fairscale * Quality * Rework from review comments * Add doc * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Address review comments Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
- 24 Feb, 2021 2 commits
-
-
Stas Bekman authored
* move secondary methods into a separate file * cleanup * style
-
Stas Bekman authored
* handle get_last_lr() before first step() * abstract away the lr getting logic * cleanup * add test * move to utils
-
- 22 Feb, 2021 4 commits
-
-
Sylvain Gugger authored
-
Stas Bekman authored
* make logging and saving trainer built-in * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Tanmay Garg authored
Enhance resume_from_checkpoint argument of Trainer.train to accept bool type. If True given, last saved checkpoint in self.args.output_dir will be loaded. (#10280)
-
Stas Bekman authored
* implement gradient_accumulation_steps support in DeepSpeed integration * typo * cleanup * cleanup
-
- 19 Feb, 2021 1 commit
-
-
Stas Bekman authored
* implement --fp16_full_eval * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * style * add test Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 18 Feb, 2021 2 commits
-
-
Stas Bekman authored
* memory tracker metrics * go back to eval for somewhat consistency * handle no-gpu case * deal with stackable eval calls * restore callback order * style * simplify the API * add test * docs * consistently use eval_ prefix * improve docs * Update src/transformers/trainer_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * rename method * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Tanmay Garg authored
Introduce warmup_ratio training argument in both TrainingArguments and TFTrainingArguments classes (#6673)
-
- 17 Feb, 2021 1 commit
-
-
Stas Bekman authored
* refactor place_model_on_device logic, add deepspeed * doc * style
-
- 16 Feb, 2021 2 commits
-
-
Stas Bekman authored
* [trainer] fix ignored columns logger This PR fixes a confusing log entry that says: ``` The following columns in the evaluation set don't have a corresponding argument in `T5ForConditionalGeneration.forward` and have been ignored: . ``` when everything is in order. * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
-
- 15 Feb, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 11 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
* Refactor things out of main train * Store signature * Add SageMakerTrainer * Init + Copyright * Address review comments
-
Stas Bekman authored
* init devices/setup explicitly * docs + test * simplify * cleanup * cleanup * cleanup * correct the required dist setup * derive local_rank from env LOCAL_RANK
-
- 10 Feb, 2021 1 commit
-
-
Stas Bekman authored
* free up memory at the end of train * rework tests * consistent formatting * correction
-
- 08 Feb, 2021 1 commit
-
-
Stas Bekman authored
* deepspeed bug fixes and tests * manual wrap?
-
- 04 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
-
Stas Bekman authored
* trainer fixes * don't switch the model just for deepspeed and mp * correct the fix
-
- 03 Feb, 2021 1 commit
-
-
yylun authored
* fix steps_in_epoch variable when using max_steps * redundant sentence * Revert "redundant sentence" This reverts commit ad5c0e9b6e66d65732dee2239cdc9c76dfa0dc5a. * remove redundant sentence Co-authored-by:wujindou <wujindou@sogou-inc.com>
-
- 02 Feb, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 29 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* When on sagemaker use their env variables for saves * Address review comments * Quality
-
- 28 Jan, 2021 4 commits
-
-
abhishek thakur authored
-
abhishek thakur authored
-
Sylvain Gugger authored
-
abhishek thakur authored
Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
- 27 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
* Whenresuming training from checkpoint, Trainer loads model * Finish cleaning tests * Address review comment * Use global_step from state
-
Sylvain Gugger authored
* Add a flag for find_unused_parameters * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Remove negation Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-