- 24 Jun, 2020 6 commits
-
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
flozi00 authored
* Create README.md * Update model_cards/a-ware/roberta-large-squad-classification/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Adriano Diniz authored
Fix/add information in README.md
-
ahotrod authored
electra_large_discriminator_squad2_512 Question Answering LM
-
Kevin Canwen Xu authored
* Fix PABEE division by zero error * patience=0 by default
-
- 23 Jun, 2020 11 commits
-
-
Sylvain Gugger authored
* Only put tensors on a device * Type hint and unpack list comprehension
-
Sylvain Gugger authored
* Add version control menu * Constify things Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Apply suggestions from code review Co-authored-by:
Julien Chaumond <chaumond@gmail.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Julien Chaumond <chaumond@gmail.com>
-
Sam Shleifer authored
-
Julien Chaumond authored
-
Sam Shleifer authored
-
Thomas Wolf authored
-
Thomas Wolf authored
* Add return lengths * make pad a bit more flexible so it can be used as collate_fn * check all kwargs sent to encoding method are known * fixing kwargs in encodings * New AddedToken class in python This class let you specify specifique tokenization behaviors for some special tokens. Used in particular for GPT2 and Roberta, to control how white spaces are stripped around special tokens. * style and quality * switched to hugginface tokenizers library for AddedTokens * up to tokenizer 0.8.0-rc3 - update API to use AddedToken state * style and quality * do not raise an error on additional or unused kwargs for tokenize() but only a warning * transfo-xl pretrained model requires torch * Update src/transformers/tokenization_utils.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Patrick von Platen authored
* improve mem handling * improve mem for pos ax encodings
-
Sam Shleifer authored
-
Sam Shleifer authored
-
Sam Shleifer authored
-
- 22 Jun, 2020 23 commits
-
-
flozi00 authored
* [Modelcard] bart-squadv2 * Update README.md * Update README.md
-
flozi00 authored
-
Fran Martinez authored
* Create README.md * changes in model usage section * minor changes in output visualization * minor errata in readme
-
furunkel authored
* Create README.md * Update README.md
-
bogdankostic authored
-
Adriano Diniz authored
-
Adriano Diniz authored
-
Adriano Diniz authored
* Create README.md * Apply suggestions from code review Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Micha毛l Benesty authored
* Add link to new comunity notebook (optimization) related to https://github.com/huggingface/transformers/issues/4842#event-3469184635 This notebook is about benchmarking model training with/without dynamic padding optimization. https://github.com/ELS-RD/transformers-notebook Using dynamic padding on MNLI provides a **4.7 times training time reduction**, with max pad length set to 512. The effect is strong because few examples are >> 400 tokens in this dataset. IRL, it will depend of the dataset, but it always bring improvement and, after more than 20 experiments listed in this [article](https://towardsdatascience.com/divide-hugging-face-transformers-training-time-by-2-or-more-21bf7129db9q-21bf7129db9e?source=friends_link&sk=10a45a0ace94b3255643d81b6475f409 ), it seems to not hurt performance. Following advice from @patrickvonplaten I do the PR myself :-) * Update notebooks/README.md Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Lee Haau-Sing authored
* nyu-mll: roberta on smaller datasets * Update README.md * Update README.md Co-authored-by:Alex Warstadt <alexwarstadt@gmail.com>
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Quicktour part 1 * Update * All done * Typos Co-authored-by:
Thomas Wolf <thomwolf@users.noreply.github.com> * Address comments in quick tour * Update docs/source/quicktour.rst Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update from feedback Co-authored-by:
Thomas Wolf <thomwolf@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Thomas Wolf authored
* Cleaner warning when loading pretrained models This make more explicit logging messages when using the various `from_pretrained` methods. It also make these messages as `logging.warning` because it's a common source of silent mistakes. * Update src/transformers/modeling_utils.py Co-authored-by:
Julien Chaumond <chaumond@gmail.com> * Update src/transformers/modeling_utils.py Co-authored-by:
Julien Chaumond <chaumond@gmail.com> * style and quality Co-authored-by:
Julien Chaumond <chaumond@gmail.com>
-
Lysandre Debut authored
* Have documentation fail on warning * Force ci failure * Revert "Force ci failure" This reverts commit f0a4666ec2eb4cd00a4da48af3357defc63324a0.
-
Sylvain Gugger authored
-
Adriano Diniz authored
-
Manuel Romero authored
* Create README.md @julien-c check out that dataset meta tag is right * Fix typo Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Manuel Romero authored
-
Patrick von Platen authored
-
Thomas Wolf authored
* fix #5081 and improve backward compatibility (slightly) * add nlp to setup.cfg - style and quality * align default to previous default * remove test that doesn't generalize
-
Malte authored
Fix for https://github.com/huggingface/transformers/issues/3809
-
Iz Beltagy authored
* add support for gradient checkpointing in BERT * fix unit tests * isort * black * workaround for `torch.utils.checkpoint.checkpoint` not accepting bool * Revert "workaround for `torch.utils.checkpoint.checkpoint` not accepting bool" This reverts commit 5eb68bb804f5ffbfc7ba13c45a47717f72d04574. * workaround for `torch.utils.checkpoint.checkpoint` not accepting bool Co-authored-by:Lysandre Debut <lysandre@huggingface.co>
-
Joseph Liu authored
* Configure all models to use output_hidden_states as argument passed to foward() * Pass all tests * Remove cast_bool_to_primitive in TF Flaubert model * correct tf xlnet * add pytorch test * add tf test * Fix broken tests * Configure all models to use output_hidden_states as argument passed to foward() * Pass all tests * Remove cast_bool_to_primitive in TF Flaubert model * correct tf xlnet * add pytorch test * add tf test * Fix broken tests * Refactor output_hidden_states for mobilebert * Reset and remerge to master Co-authored-by:
Joseph Liu <joseph.liu@coinflex.com> Co-authored-by:
patrickvonplaten <patrick.v.platen@gmail.com>
-