- 07 May, 2020 4 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
* README * Update README.md
-
Julien Chaumond authored
* Created using Colaboratory * [examples] reorganize files * remove run_tpu_glue.py as superseded by TPU support in Trainer * Bugfix: int, not tuple * move files around
-
- 22 Apr, 2020 1 commit
-
-
Julien Chaumond authored
* doc * [tests] Add sample files for a regression task * [HUGE] Trainer * Feedback from @sshleifer * Feedback from @thomwolf + logging tweak * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes * [glue] Use default max_seq_length of 128 like before * [glue] move DataTrainingArguments around * [ner] Change interface of InputExample, and align run_{tf,pl} * Re-align the pl scripts a little bit * ner * [ner] Add integration test * Fix language_modeling with API tweak * [ci] Tweak loss target * Don't break console output * amp.initialize: model must be on right device before * [multiple-choice] update for Trainer * Re-align to 827d6d6e
-
- 10 Apr, 2020 2 commits
-
-
Jin Young Sohn authored
* Initial commit to get BERT + run_glue.py on TPU * Add README section for TPU and address comments. * Cleanup TPU bits from run_glue.py (#3) TPU runner is currently implemented in: https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py. We plan to upstream this directly into `huggingface/transformers` (either `master` or `tpu`) branch once it's been more thoroughly tested. * Cleanup TPU bits from run_glue.py TPU runner is currently implemented in: https://github.com/pytorch-tpu/transformers/blob/tpu/examples/run_glue_tpu.py . We plan to upstream this directly into `huggingface/transformers` (either `master` or `tpu`) branch once it's been more thoroughly tested. * No need to call `xm.mark_step()` explicitly (#4) Since for gradient accumulation we're accumulating on batches from `ParallelLoader` instance which on next() marks the step itself. * Resolve R/W conflicts from multiprocessing (#5) * Add XLNet in list of models for `run_glue_tpu.py` (#6) * Add RoBERTa to list of models in TPU GLUE (#7) * Add RoBERTa and DistilBert to list of models in TPU GLUE (#8) * Use barriers to reduce duplicate work/resources (#9) * Shard eval dataset and aggregate eval metrics (#10) * Shard eval dataset and aggregate eval metrics Also, instead of calling `eval_loss.item()` every time do summation with tensors on device. * Change defaultdict to float * Reduce the pred, label tensors instead of metrics As brought up during review some metrics like f1 cannot be aggregated via averaging. GLUE task metrics depends largely on the dataset, so instead we sync the prediction and label tensors so that the metrics can be computed accurately on those instead. * Only use tb_writer from master (#11) * Apply huggingface black code formatting * Style * Remove `--do_lower_case` as example uses cased * Add option to specify tensorboard logdir This is needed for our testing framework which checks regressions against key metrics writtern by the summary writer. * Using configuration for `xla_device` * Prefix TPU specific comments. * num_cores clarification and namespace eval metrics * Cache features file under `args.cache_dir` Instead of under `args.data_dir`. This is needed as our test infra uses data_dir with a read-only filesystem. * Rename `run_glue_tpu` to `run_tpu_glue` Co-authored-by:
LysandreJik <lysandre.debut@reseau.eseo.fr>
-
Julien Chaumond authored
-
- 09 Mar, 2020 1 commit
-
-
Lysandre authored
closes #3183
-
- 25 Feb, 2020 1 commit
-
-
Jhuo IH authored
-
- 22 Feb, 2020 1 commit
-
-
saippuakauppias authored
-
- 20 Feb, 2020 1 commit
-
-
srush authored
* initial pytorch lightning commit * tested multigpu * Fix learning rate schedule * black formatting * fix flake8 * isort * isort * . Co-authored-by:Check your git settings! <chris@chris-laptop>
-
- 17 Feb, 2020 1 commit
-
-
VictorSanh authored
-
- 07 Feb, 2020 2 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
-
- 30 Jan, 2020 1 commit
-
-
Jared Nielsen authored
-
- 24 Jan, 2020 1 commit
-
-
VictorSanh authored
-
- 16 Jan, 2020 3 commits
- 06 Jan, 2020 2 commits
-
-
alberduris authored
-
alberduris authored
-
- 24 Dec, 2019 1 commit
-
-
Aymeric Augustin authored
Use -e only in docs targeted at contributors. If a user copy-pastes command line with [--editable], they will hit an error. If they don't know the --editable option, we're giving them a choice to make before they can move forwards, but this isn't a choice they need to make right now.
-
- 21 Dec, 2019 1 commit
-
-
thomwolf authored
-
- 19 Dec, 2019 1 commit
-
-
Ejar authored
Updated documentation due to typo
-
- 10 Dec, 2019 4 commits
-
-
Rémi Louf authored
-
Suvrat Bhooshan authored
-
Rémi Louf authored
-
Rémi Louf authored
-
- 05 Dec, 2019 2 commits
-
-
Julien Plu authored
-
thomwolf authored
-
- 27 Nov, 2019 4 commits
-
-
VictorSanh authored
-
VictorSanh authored
-
VictorSanh authored
-
Julien Chaumond authored
-
- 23 Nov, 2019 1 commit
-
-
Julien Chaumond authored
-
- 21 Nov, 2019 1 commit
-
-
Rémi Louf authored
-
- 15 Nov, 2019 1 commit
-
-
Xu Hongshen authored
-
- 14 Nov, 2019 1 commit
-
-
Thomas Wolf authored
-
- 04 Nov, 2019 1 commit
-
-
thomwolf authored
-
- 30 Oct, 2019 1 commit
-
-
Rémi Louf authored
-