- 19 May, 2020 2 commits
-
-
Patrick von Platen authored
* fix gpu slow tests in pytorch * change model to device syntax
-
Julien Chaumond authored
* Distributed eval: SequentialDistributedSampler + gather all results * For consistency only write to disk from world_master Close https://github.com/huggingface/transformers/issues/4272 * Working distributed eval * Hook into scripts * Fix #3721 again * TPU.mesh_reduce: stay in tensor space Thanks @jysohn23 * Just a small comment * whitespace * torch.hub: pip install packaging * Add test scenarii
-
- 18 May, 2020 2 commits
-
-
Boris Dayma authored
-
Julien Chaumond authored
see https://github.com/huggingface/transformers/pull/4367#discussion_r426356693 Hat/tip @girishponkiya
-
- 15 May, 2020 3 commits
-
-
Julien Chaumond authored
-
Lysandre Debut authored
-
Julien Chaumond authored
-
- 14 May, 2020 2 commits
-
-
Julien Chaumond authored
* Fix: unpin flake8 and fix cs errors * Ok we still need to quote those
-
Julien Chaumond authored
see context in https://github.com/huggingface/transformers/pull/4223
-
- 13 May, 2020 2 commits
-
-
Julien Plu authored
* Add QA trainer example for TF * Make data_dir optional * Fix parameter logic * Fix feature convert * Update the READMEs to add the question-answering task * Apply style * Change 'sequence-classification' to 'text-classification' and prefix with 'eval' all the metric names * Apply style * Apply style
-
Julien Chaumond authored
* Improvements to the wandb integration * small reorg + no global necessary * feat(trainer): log epoch and final metrics * Simplify logging a bit * Fixup * Fix crash when just running eval Co-authored-by:
Chris Van Pelt <vanpelt@gmail.com> Co-authored-by:
Boris Dayma <boris.dayma@gmail.com>
-
- 12 May, 2020 1 commit
-
-
Viktor Alm authored
* catch gpu len 1 set to gpu0 * Add mpc to trainer * Add MPC for TF * fix TF automodel for MPC and add Albert * Apply style * Fix import * Note to self: double check * Make shape None, None for datasetgenerator output shapes * Add from_pt bool which doesnt seem to work * Original checkpoint dir * Fix docstrings for automodel * Update readme and apply style * Colab should probably not be from users * Colabs should probably not be from users * Add colab * Update README.md * Update README.md * Cleanup __intit__ * Cleanup flake8 trailing comma * Update src/transformers/training_args_tf.py * Update src/transformers/modeling_tf_auto.py Co-authored-by:
Viktor Alm <viktoralm@pop-os.localdomain> Co-authored-by:
Julien Chaumond <chaumond@gmail.com>
-
- 11 May, 2020 1 commit
-
-
Stefan Schweter authored
* docs: fix link to token classification (NER) example * examples: fix links to NER scripts
-
- 08 May, 2020 1 commit
-
-
Julien Chaumond authored
* [TPU] Doc, fix xla_spawn.py, only preprocess dataset once * Update examples/README.md * [xla_spawn] Add `_mp_fn` to other Trainer scripts * [TPU] Fix: eval dataloader was None
-
- 07 May, 2020 5 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
* README * Update README.md
-
Julien Chaumond authored
* Created using Colaboratory * [examples] reorganize files * remove run_tpu_glue.py as superseded by TPU support in Trainer * Bugfix: int, not tuple * move files around
-
Lysandre Debut authored
* wip * wip * a last wip * Better logging when using TPUs * Correct argument name * Tests * fix * Metrics in evaluation * Update src/transformers/training_args.py * [tpu] Use launcher script instead * [tpu] lots of tweaks * Fix formatting Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 06 May, 2020 2 commits
-
-
Julien Plu authored
* First commit to add a TF version of the trainer. * Make the TF trainer closer to what looks the PT trainer * Refactoring common code between the PT and TF trainer into an util file. * Some bugfix + better similarity with the PT trainer * Add missing class in transformers init * Bugfix over prediction + use classification report instead of simple metrics * Fix name error * Fix optimization tests + style * Apply style * Several bugfix for multi-gpu training * Apply style * Apply style * Add glue example for the TF trainer * Several bugix + address the reviews * Fix on the TF training args file * Add a debug mode * Bugfix in utils_ner.py when segment_ids is None * Apply style * Apply style * Add TPU strategy * Fix selection strategy
-
Simone Primarosa authored
-
- 02 May, 2020 3 commits
-
-
William Falcon authored
-
William Falcon authored
-
Stefan Schweter authored
* ner: parse args from .args file or JSON * examples: mention json-based configuration file support for run_ner script
-
- 01 May, 2020 1 commit
-
-
Julien Chaumond authored
[qol] example scripts: parse args from .args file or JSON
-
- 29 Apr, 2020 1 commit
-
-
Julien Chaumond authored
* [file_utils] use_cdn + documentation * Move to cdn. urls for weights * [urls] Hotfix for bert-base-japanese
-
- 28 Apr, 2020 2 commits
-
-
Sam Shleifer authored
* add known 3rd party to setup.cfg * comment * Update CONTRIBUTING.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Patrick von Platen authored
* fix empty prompt * fix length in generation pipeline
-
- 24 Apr, 2020 2 commits
-
-
Julien Chaumond authored
Close #3921
-
Cola authored
* Shuffle train subset * Cleaner shuffle
-
- 22 Apr, 2020 2 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
* doc * [tests] Add sample files for a regression task * [HUGE] Trainer * Feedback from @sshleifer * Feedback from @thomwolf + logging tweak * [file_utils] when downloading concurrently, get_from_cache will use the cached file for subsequent processes * [glue] Use default max_seq_length of 128 like before * [glue] move DataTrainingArguments around * [ner] Change interface of InputExample, and align run_{tf,pl} * Re-align the pl scripts a little bit * ner * [ner] Add integration test * Fix language_modeling with API tweak * [ci] Tweak loss target * Don't break console output * amp.initialize: model must be on right device before * [multiple-choice] update for Trainer * Re-align to 827d6d6e
-
- 20 Apr, 2020 3 commits
-
-
Andrey Kulagin authored
-
Jared T Nielsen authored
* Add qas_id * Fix incorrect name in squad.py * Make output files optional for squad eval
-
Sam Shleifer authored
-
- 18 Apr, 2020 1 commit
-
-
Thomas Wolf authored
* First pass on utility classes and python tokenizers * finishing cleanup pass * style and quality * Fix tests * Updating following @mfuntowicz comment * style and quality * Fix Roberta * fix batch_size/seq_length inBatchEncoding * add alignement methods + tests * Fix OpenAI and Transfo-XL tokenizers * adding trim_offsets=True default for GPT2 et RoBERTa * style and quality * fix tests * add_prefix_space in roberta * bump up tokenizers to rc7 * style * unfortunately tensorfow does like these - removing shape/seq_len for now * Update src/transformers/tokenization_utils.py Co-Authored-By:
Stefan Schweter <stefan@schweter.it> * Adding doc and docstrings * making flake8 happy Co-authored-by:
Stefan Schweter <stefan@schweter.it>
-
- 16 Apr, 2020 3 commits
-
-
Sam Shleifer authored
renames `run_bart_sum.py` to `finetune.py`
-
Patrick von Platen authored
* Refactored use of newstest2013 to newstest2014. Fixed bug where argparse consumed first command line argument as model_size argument rather than using default model_size by forcing explicit --model_size flag inclusion * More pythonic file handling through 'with' context * COSMETIC - ran Black and isort * Fixed reference to number of lines in newstest2014 * Fixed failing test. More pythonic file handling * finish PR from tholiao * remove outcommented lines * make style * make isort happy Co-authored-by:Thomas Liao <tholiao@gmail.com>
-
Davide Fiocco authored
-
- 15 Apr, 2020 1 commit
-
-
Sam Shleifer authored
- adds pytorch-lightning dependency
-