- 26 Aug, 2020 1 commit
-
-
Lysandre authored
-
- 17 Aug, 2020 1 commit
-
-
Sam Shleifer authored
-
- 11 Aug, 2020 2 commits
-
-
Stas Bekman authored
* [wip] add get_polynomial_decay_schedule_with_warmup * style * add assert * change lr_end to a much smaller default number * check for exact equality * [model_cards] electra-base-turkish-cased-ner (#6350) * for electra-base-turkish-cased-ner * Add metadata Co-authored-by:
Julien Chaumond <chaumond@gmail.com> * Temporarily de-activate TPU CI * Update modeling_tf_utils.py (#6372) fix typo: ckeckpoint->checkpoint * the test now works again (#6371) * correct pl link in readme (#6364) * refactor almost identical tests (#6339) * refactor almost identical tests * important to add a clear assert error message * make the assert error even more descriptive than the original bt * Small docfile fixes (#6328) * Patch models (#6326) * TFAlbertFor{TokenClassification, MultipleChoice} * Patch models * BERT and TF BERT info s * Update check_repo * Ci GitHub caching (#6382) * Cache Github Actions CI * Remove useless file * Colab button (#6389) * Add colab button * Add colab link for tutorials * Fix links for open in colab (#6391) * Update src/transformers/optimization.py consistently use lr_end=1e-7 default Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * [wip] add get_polynomial_decay_schedule_with_warmup * style * add assert * change lr_end to a much smaller default number * check for exact equality * Update src/transformers/optimization.py consistently use lr_end=1e-7 default Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * remove dup (leftover from merge) * convert the test into the new refactored format * stick to using the current_step as is, without ++ Co-authored-by:
M. Yusuf Sarıgöz <yusufsarigoz@gmail.com> Co-authored-by:
Julien Chaumond <chaumond@gmail.com> Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Alexander Measure <ameasure@gmail.com> Co-authored-by:
Rohit Gupta <rohitgr1998@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Stas Bekman authored
-
- 09 Aug, 2020 1 commit
-
-
Sam Shleifer authored
-
- 06 Aug, 2020 1 commit
-
-
Bhashithe Abeysinghe authored
Co-authored-by:Sam Shleifer <sshleifer@gmail.com>
-
- 05 Aug, 2020 1 commit
-
-
Stas Bekman authored
* support --lr_scheduler with multiple possibilities * correct the error message * add a note about supported schedulers * cleanup * cleanup2 * needs the argument default * style * add another assert in the test * implement requested changes * cleanups * fix relative import * cleanup
-
- 03 Aug, 2020 1 commit
-
-
Sam Shleifer authored
-
- 30 Jul, 2020 1 commit
-
-
Stas Bekman authored
-
- 18 Jul, 2020 1 commit
-
-
Nathan Raw authored
Co-authored-by:Sam Shleifer <sshleifer@gmail.com>
-
- 26 Jun, 2020 1 commit
-
-
Sam Shleifer authored
-
- 23 Jun, 2020 2 commits
-
-
Sam Shleifer authored
-
Sam Shleifer authored
-
- 17 Jun, 2020 1 commit
-
-
Sam Shleifer authored
-
- 07 May, 2020 1 commit
-
-
Julien Chaumond authored
* Created using Colaboratory * [examples] reorganize files * remove run_tpu_glue.py as superseded by TPU support in Trainer * Bugfix: int, not tuple * move files around
-
- 22 Apr, 2020 1 commit
-
-
Julien Chaumond authored
* doc * [tests] Add sample files for a regression task * [HUGE] Trainer * Feedback from @sshleifer * Feedback from @thomwolf + logging tweak * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes * [glue] Use default max_seq_length of 128 like before * [glue] move DataTrainingArguments around * [ner] Change interface of InputExample, and align run_{tf,pl} * Re-align the pl scripts a little bit * ner * [ner] Add integration test * Fix language_modeling with API tweak * [ci] Tweak loss target * Don't break console output * amp.initialize: model must be on right device before * [multiple-choice] update for Trainer * Re-align to 827d6d6e
-
- 20 Apr, 2020 1 commit
-
-
Sam Shleifer authored
-
- 15 Apr, 2020 1 commit
-
-
Sam Shleifer authored
- adds pytorch-lightning dependency
-
- 07 Apr, 2020 1 commit
-
-
Sam Shleifer authored
-
- 25 Mar, 2020 1 commit
-
-
Andre Carrera authored
-
- 17 Mar, 2020 1 commit
-
-
Nathan Raw authored
*
✨ Alter base pl transformer to use automodels *🐛 Add batch size env variable to function call *💄 Apply black code style from Makefile *🚚 Move lightning base out of ner directory *✨ Add lightning glue example *💄 self * move _feature_file to base class *✨ Move eval logging to custom callback *💄 Apply black code style *🐛 Add parent to pythonpath, remove copy command *🐛 Add missing max_length kwarg
-
- 27 Feb, 2020 1 commit
-
-
srush authored
* changes to allow for tpu training * black * tpu * tpu
-
- 20 Feb, 2020 2 commits
-
-
srush authored
-
srush authored
* initial pytorch lightning commit * tested multigpu * Fix learning rate schedule * black formatting * fix flake8 * isort * isort * . Co-authored-by:Check your git settings! <chris@chris-laptop>
-