"vscode:/vscode.git/clone" did not exist on "f1f23ad1710953e75b53a85953b018b8caceb427"
- 30 Apr, 2021 1 commit
-
-
bonniehyeon authored
* Fix do_eval default value in training_args.py * Update PULL_REQUEST_TEMPLATE.md
-
- 29 Apr, 2021 1 commit
-
-
Sylvain Gugger authored
* Split checkpoint from model_name_or_path in examples * Address review comments * Address review comments
-
- 26 Apr, 2021 1 commit
-
-
Stas Bekman authored
* adding Z-inf * revamp config process * up version requirement * wip * massive rewrite * cleanup * cleanup * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * consistent json commas * act on suggestions * leave this feature for 0.3.16 * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 23 Apr, 2021 1 commit
-
-
Sylvain Gugger authored
* Initial support for upload to hub * push -> upload * Fixes + examples * Fix torchhub test * Torchhub test I hate you * push_model_to_hub -> push_to_hub * Apply mixin to other pretrained models * Remove ABC inheritance * Add tests * Typo * Run tests * Install git-lfs * Change approach * Add push_to_hub to all * Staging test suite * Typo * Maybe like this? * More deps * Cache * Adapt name * Quality * MOAR tests * Put it in testing_utils * Docs + torchhub last hope * Styling * Wrong method * Typos * Update src/transformers/file_utils.py Co-authored-by:
Julien Chaumond <julien@huggingface.co> * Address review comments * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Julien Chaumond <julien@huggingface.co> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
- 16 Apr, 2021 1 commit
-
-
Sylvain Gugger authored
* Bulk of the work * Polish and tests * Update QA Trainer * Avoid breaking the predict method * Deprecation warnings * Store real eval dataloder * Get eval dataset reference before wrap
-
- 13 Apr, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 08 Apr, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 31 Mar, 2021 2 commits
-
-
JohnnyC08 authored
In the group by length documentation length is misspelled as legnth
-
Sylvain Gugger authored
* Replace is_sagemaker_distributed_available * Merge SageMakerTrainer into Trainer * Test with shorter condition * Put back deleted line * Deprecate SageMakerTrainer and SageMakerTrainingArguments * Apply suggestions from code review Co-authored-by:
Philipp Schmid <32632186+philschmid@users.noreply.github.com> Co-authored-by:
Philipp Schmid <32632186+philschmid@users.noreply.github.com>
-
- 29 Mar, 2021 1 commit
-
-
pcuenca authored
A new argument `length_column_name` has been added to `TrainingArguments`, with default value `"length"`. If this column exists and `group_by_length` is `True`, the train sampler will use it for grouping rather than computing it before training starts. This is an optimization that allows the user to prepare data for fast processing, preventing sequential access to the dataset as described in issue #10909.
-
- 24 Mar, 2021 1 commit
-
-
Sidd Karamcheti authored
-
- 16 Mar, 2021 2 commits
-
-
Cheng Li authored
* pass hf optimizer and scheduler to deepspeed if not specified in ds config * pass hf optimizer and scheduler to deepspeed if not specified in ds config * update * make init_deepspeed support config dict * fix docstring formatting * clean up trainer's comments * add new tests * fix type * composit argparse doesn't work * style * add a new test, rename others * document new functionality * complete tests, add docs * style * correct level * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * add new methods to the doc * must tell DS we are using a non-native optimizer * add protection against cpu_offload + HF optimizer combo * fix the cli overrides * sync docs + tests * restore AdamW * better docs * need new version * no longer needed * remove outdate information * refactor duplicated code Co-authored-by:
Stas Bekman <stas@stason.org> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
* Add DistributedSamplerWithLoop * Fix typo * Test and small fix
-
- 15 Mar, 2021 2 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Handle save differently * Missing imports * Fix typo * Adapt to recent changes in save_pretrained * Forgotten brackets * Optimizer load * Fix world size * Deal wth None * Remove needless self
-
- 12 Mar, 2021 2 commits
-
-
PaulLerner authored
* fix: #10628 expanduser path in TrainingArguments * docs: explain why we expand paths in TrainingArguments * Style Co-authored-by:Sylvain Gugger <sylvain.gugger@gmail.com>
-
Sylvain Gugger authored
* Add auto_wrap option in fairscale integration * Style
-
- 04 Mar, 2021 1 commit
-
-
Philipp Schmid authored
* removed overwrites * remove default value for output_dir * adjusted typing
-
- 03 Mar, 2021 1 commit
-
-
Sylvain Gugger authored
* Fix gradient accumulation for SM Model Parallelism * Style and divide loss by grad accum steps
-
- 28 Feb, 2021 1 commit
-
-
Tanmay Garg authored
* Introduce save_strategy training argument * deprecate EvaluationStrategy * collapse EvaluationStrategy and LoggingStrategy into a single IntervalStrategy enum * modify tests to use modified enum
-
- 25 Feb, 2021 1 commit
-
-
Sylvain Gugger authored
* Ass support for ZeRO-2/3 and ZeRO-offload in fairscale * Quality * Rework from review comments * Add doc * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Address review comments Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
- 19 Feb, 2021 2 commits
-
-
Tanmay Garg authored
Introduce logging_strategy training argument in TrainingArguments and TFTrainingArguments. (#9838)
-
Stas Bekman authored
* implement --fp16_full_eval * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * style * add test Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 18 Feb, 2021 2 commits
-
-
Stas Bekman authored
* memory tracker metrics * go back to eval for somewhat consistency * handle no-gpu case * deal with stackable eval calls * restore callback order * style * simplify the API * add test * docs * consistently use eval_ prefix * improve docs * Update src/transformers/trainer_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * rename method * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Tanmay Garg authored
Introduce warmup_ratio training argument in both TrainingArguments and TFTrainingArguments classes (#6673)
-
- 11 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
* Refactor things out of main train * Store signature * Add SageMakerTrainer * Init + Copyright * Address review comments
-
Stas Bekman authored
* init devices/setup explicitly * docs + test * simplify * cleanup * cleanup * cleanup * correct the required dist setup * derive local_rank from env LOCAL_RANK
-
- 09 Feb, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 31 Jan, 2021 1 commit
-
-
lewtun authored
* Clarify definition of seed argument in Trainer * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/training_args_tf.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Fix style * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 29 Jan, 2021 2 commits
-
-
Stas Bekman authored
-
Sylvain Gugger authored
* When on sagemaker use their env variables for saves * Address review comments * Quality
-
- 28 Jan, 2021 2 commits
-
-
abhishek thakur authored
-
abhishek thakur authored
Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
- 27 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Add a flag for find_unused_parameters * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> * Remove negation Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
- 26 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Add a debug print * Adapt Trainer to use smdistributed if available * Forgotten parenthesis * Real check for sagemaker * Donforget to define device... * Woopsie, local)rank is defined differently * Update since local_rank has the proper value * Remove debug statement * More robust check for smdistributed * Quality * Deal with key not present error
-
- 22 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 20 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
-
Gunjan Chhablani authored
* Fix Trainer and Args to mention AdamW, not Adam. * Update the docs for Training Arguments. * Change arguments adamw_* to adam_* * Fixed links to AdamW in TrainerArguments docs * Fix line length in Training Args docs.
-
- 14 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
* Upstream (and rename) sortish sampler * Use proper sampler * Update src/transformers/trainer_pt_utils.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Sylvain Gugger authored
* Fix Trainer with a parallel model * More clean up
-