- 02 Apr, 2020 1 commit
-
-
Patrick von Platen authored
* replace heavy t5 models with tiny random models as was done by sshleifer * fix isort
-
- 01 Apr, 2020 1 commit
-
-
Julien Chaumond authored
* Start cleaning examples * Fixup
-
- 31 Mar, 2020 1 commit
-
-
Patrick von Platen authored
* fix conflicts * add model size argument to summarization * correct wrong import * fix isort * correct imports * other isort make style * make style
-
- 30 Mar, 2020 3 commits
-
-
Ethan Perez authored
* Using loaded checkpoint with --do_predict Without this fix, I'm getting near-random validation performance for a trained model, and the validation performance differs per validation run. I think this happens since the `model` variable isn't set with the loaded checkpoint, so I'm using a randomly initialized model. Looking at the model activations, they differ each time I run evaluation (but they don't with this fix). * Update checkpoint loading * Fixing model loading
-
Sam Shleifer authored
-
Julien Plu authored
* Update the NER TF script to remove the softmax and make the pad token label id to -1 * Reformat the quality and style Co-authored-by:Julien Plu <julien.plu@adevinta.com>
-
- 29 Mar, 2020 1 commit
-
-
Sam Shleifer authored
-
- 27 Mar, 2020 4 commits
-
-
Patrick von Platen authored
* force bleu * fix wrong file name * rename file * different filenames for each example test * test files should clean up after themselves * test files should clean up after themselves * do not force bleu * correct typo * fix isort
-
Funtowicz Morgan authored
* Use tokenizer.num_added_tokens to count number of added special_tokens instead of hardcoded numbers. Signed-off-by:
Morgan Funtowicz <morgan@huggingface.co> * run_ner.py - Do not add a label to the labels_ids if word_tokens is empty. This can happen when using bert-base-multilingual-cased with an input containing an unique space. In this case, the tokenizer will output just an empty word_tokens thus leading to an non-consistent behavior over the labels_ids tokens adding one more tokens than tokens vector. Signed-off-by:
Morgan Funtowicz <morgan@huggingface.co>
-
Patrick von Platen authored
-
Lysandre Debut authored
T5-small in test isort
-
- 26 Mar, 2020 3 commits
-
-
Patrick von Platen authored
* add translation example * make style * adapt docstring * add gpu device as input for example * small renaming * better README
-
Patrick von Platen authored
* rebase to master * change tf to pytorch * change to pytorch * small fix * renaming * add gpu training possibility * renaming * improve README * incoorporate collins feedback * better Readme * better README.md
-
Lysandre Debut authored
-
- 25 Mar, 2020 1 commit
-
-
Andre Carrera authored
-
- 24 Mar, 2020 3 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
-
Julien Chaumond authored
-
- 23 Mar, 2020 1 commit
-
-
Julien Chaumond authored
-
- 20 Mar, 2020 3 commits
-
-
Julien Chaumond authored
-
Elijah Rippeth authored
For more details, see https://pytorch.org/docs/stable/optim.html#how-to-adjust-learning-rate
-
Patrick von Platen authored
* make style * fix conflicts
-
- 19 Mar, 2020 3 commits
-
-
Nitish Shirish Keskar authored
torch.cuda.empty_cache() was being called from a TF function (even when torch is unavailable) not sure any replacement is needed if TF OOMs
-
Julien Chaumond authored
-
mataney authored
* solving bug where for small epochs and large gradient_accumulation_steps we never train * black formatting * no need to change these files
-
- 17 Mar, 2020 4 commits
-
-
J.P Lee authored
* Update examples/ner/run_ner.py to use AutoModel * Fix missing code and apply `make style` command
-
Nathan Raw authored
*
✨ Alter base pl transformer to use automodels *🐛 Add batch size env variable to function call *💄 Apply black code style from Makefile *🚚 Move lightning base out of ner directory *✨ Add lightning glue example *💄 self * move _feature_file to base class *✨ Move eval logging to custom callback *💄 Apply black code style *🐛 Add parent to pythonpath, remove copy command *🐛 Add missing max_length kwarg -
Patrick von Platen authored
* change do_samples back * None better default as boolean * adapt do_sample to True in test example * make style
-
Thomas Wolf authored
* memory benchmark rss * have both forward pass and line-by-line mem tracing * cleaned up tracing * refactored and cleaning up API * no f-strings yet... * add GPU mem logging * fix GPU memory monitoring * style and quality * clean up and doc * update with comments * Switching to python 3.6+ * fix quality
-
- 16 Mar, 2020 1 commit
-
-
Sam Shleifer authored
* Remove unused kwargs * dont call forward in tests
-
- 13 Mar, 2020 3 commits
-
-
Patrick von Platen authored
-
Patrick von Platen authored
-
dependabot[bot] authored
Bumps [psutil](https://github.com/giampaolo/psutil) from 5.6.3 to 5.6.6. - [Release notes](https://github.com/giampaolo/psutil/releases) - [Changelog](https://github.com/giampaolo/psutil/blob/master/HISTORY.rst) - [Commits](https://github.com/giampaolo/psutil/compare/release-5.6.3...release-5.6.6 ) Signed-off-by:
dependabot[bot] <support@github.com>
-
- 12 Mar, 2020 1 commit
-
-
Sam Shleifer authored
* Update bart example docs
-
- 11 Mar, 2020 1 commit
-
-
Patrick von Platen authored
-
- 10 Mar, 2020 1 commit
-
-
Shubham Agarwal authored
* 1. seqeval required by ner pl example. install from examples/requirements. 2. unrecognized arguments: save_steps * pl checkpoint callback filenotfound error: make directory and pass * #3159 pl checkpoint path difference * 1. Updated Readme for pl 2. pl script now also correct displays logs 3. pass gpu ids compared to number of gpus * Updated results in readme * 1. updated readme 2. removing deprecated pl methods 3. finalizing scripts * comment length check * using deprecated validation_end for stable results * style related changes
-
- 09 Mar, 2020 2 commits
-
-
Sam Shleifer authored
-
Lysandre authored
closes #3183
-
- 05 Mar, 2020 2 commits
-
-
Sam Shleifer authored
* improved documentation
-
sshleifer authored
-