- 18 Feb, 2021 1 commit
-
-
Stas Bekman authored
* memory tracker metrics * go back to eval for somewhat consistency * handle no-gpu case * deal with stackable eval calls * restore callback order * style * simplify the API * add test * docs * consistently use eval_ prefix * improve docs * Update src/transformers/trainer_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * rename method * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 17 Feb, 2021 1 commit
-
-
Stas Bekman authored
* fix invalid port * missing requirements
-
- 16 Feb, 2021 1 commit
-
-
Zhang Cheng authored
-
- 15 Feb, 2021 2 commits
-
-
Suraj Patil authored
* move old s2s scripts to legacy * add the tests back * proper rename * restore * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas@stason.org> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Stas Bekman authored
* fix run_seq2seq.py; porting DeepSpeed tests to it * unrefactor * defensive programming * defensive programming 2 * port the rest of the trainer tests * style * a cleaner scripts dir finder * cleanup
-
- 12 Feb, 2021 1 commit
-
-
Suraj Patil authored
* fix rouge metrics and task specific params * fix typo * round metrics * typo * remove task_specific_params
-
- 11 Feb, 2021 2 commits
-
-
Stas Bekman authored
* init devices/setup explicitly * docs + test * simplify * cleanup * cleanup * cleanup * correct the required dist setup * derive local_rank from env LOCAL_RANK
-
Qbiwan authored
* remove xnli_compute_metrics, add load_dataset, load_metric, set_seed,metric.compute,load_metric * fix * fix * fix * push * fix * everything works * fix init * fix * special treatment for sepconv1d * style *
馃檹 馃徑 * add doc and cleanup * fix doc * fix doc again * fix doc again * Apply suggestions from code review * make style * Proposal that should work * Remove needless code * Fix test * Apply suggestions from code review * remove xnli_compute_metrics, add load_dataset, load_metric, set_seed,metric.compute,load_metric * amend README * removed data_args.task_name and replaced with task_name = "xnli"; use split function to load train and validation dataset separately; remove __post_init__; remove flag --task_name from README. * removed dict task_to_keys, use str "xnli" instead of variable task_name, change preprocess_function to use examples["premise"], examples["hypothesis"] directly, remove sentence1_key and sentence2_key, change compute_metrics function to cater only to accuracy metric, add condition for train_langauge is None when using dataset.load_dataset() * removed `torch.distributed.barrier()` and `import torch` as `from_pretrained` is able to do the work; amend README
-
- 10 Feb, 2021 2 commits
-
-
Stas Bekman authored
* free up memory at the end of train * rework tests * consistent formatting * correction
-
Lysandre Debut authored
-
- 09 Feb, 2021 2 commits
-
-
Boris Dayma authored
* doc: update W&B related doc * doc(wandb): mention report_to * doc(wandb): commit suggestion Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * doc(wandb): fix typo * doc(wandb): remove WANDB_DISABLED Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Suraj Patil authored
* add do_predict, pass eval_beams durig eval * update help * apply suggestions from code review
-
- 08 Feb, 2021 6 commits
-
-
Stas Bekman authored
-
Stas Bekman authored
* deepspeed bug fixes and tests * manual wrap?
-
Olivier authored
* replace -100 token ids with the tokenizer pad_id for compute_metrics * fixed typo for label_ids
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Stas Bekman authored
-
- 05 Feb, 2021 2 commits
-
-
Stas Bekman authored
* make executable * make executable * same for the template * cleanup
-
Suraj Patil authored
* add prepare_decoder_input_ids_from_labels in s2s models * support lbl smoothing and enc/emb freezing * fix freezing * use pad_token_id from config * remove embed freezing and add warning * prepare decoder_input_ids inside DataCollatorForSeq2Seq
-
- 03 Feb, 2021 2 commits
-
-
Suraj Patil authored
-
Stas Bekman authored
Looks like a vulnerability and it's not really used anywhere in the code, so just as well remove it completely from deps. https://github.com/huggingface/transformers/security/dependabot/examples/research_projects/lxmert/requirements.txt/bleach/open
-
- 02 Feb, 2021 1 commit
-
-
Patrick von Platen authored
* change tokenizer requirement * split line * Correct typo from list to str * improve style * make other function pretty as well * add comment * correct typo * add new test * pass tests for tok without padding token * Apply suggestions from code review
-
- 01 Feb, 2021 3 commits
-
-
Sylvain Gugger authored
* Remove subclass for sortish sampler * Use old Seq2SeqTrainer in script * Styling
-
wlhgtc authored
* MOD: fit chinese wwm to new datasets * MOD: move wwm to new folder * MOD: formate code * Styling * MOD add param and recover trainer Co-authored-by:Sylvain Gugger <sylvain.gugger@gmail.com>
-
Stas Bekman authored
-
- 29 Jan, 2021 1 commit
-
-
Stas Bekman authored
-
- 28 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 27 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
-
- 26 Jan, 2021 3 commits
-
-
Yusuke Mori authored
-
Magdalena Biesialska authored
-
Andrea Cappelli authored
* Pad to 8x for fp16 multiple choice example (#9752) * Pad to 8x for fp16 squad trainer example (#9752) * Pad to 8x for fp16 ner example (#9752) * Pad to 8x for fp16 swag example (#9752) * Pad to 8x for fp16 qa beam search example (#9752) * Pad to 8x for fp16 qa example (#9752) * Pad to 8x for fp16 seq2seq example (#9752) * Pad to 8x for fp16 glue example (#9752) * Pad to 8x for fp16 new ner example (#9752) * update script template #9752 * Update examples/multiple-choice/run_swag.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update examples/question-answering/run_qa.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update examples/question-answering/run_qa_beam_search.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * improve code quality #9752 Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 25 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Auto-resume training from checkpoint * Update examples/text-classification/run_glue.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Roll out to other examples Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
- 23 Jan, 2021 1 commit
-
-
Wilfried L. Bounsi authored
-
- 22 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
* Fixes to run_seq2seq and instructions * Add more defaults for summarization
-
Stefan Schweter authored
-
- 21 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* Fix memory regression in Seq2Seq example * Fix test and properly deal with -100 * Easier condition with device safety * Patch for MBartTokenzierFast
-
- 20 Jan, 2021 2 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Restrain tokenizer.model_max_length default * Fix indent
-
- 19 Jan, 2021 1 commit
-
-
Sylvain Gugger authored
* New run_seq2seq script * Add tests * Mark as slow * Update examples/seq2seq/run_seq2seq.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/data/data_collator.py Co-authored-by:
Suraj Patil <surajp815@gmail.com> * Update src/transformers/data/data_collator.py Co-authored-by:
Suraj Patil <surajp815@gmail.com> * Address review comments Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Suraj Patil <surajp815@gmail.com>
-