- 26 Jun, 2020 8 commits
-
-
Kevin Canwen Xu authored
-
Sam Shleifer authored
-
Patrick von Platen authored
* add notebook * Cr茅茅 avec Colaboratory * move notebook to correct folder * correct link * correct filename * correct filename * better name
-
Patrick von Platen authored
* fix docs * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_tf_utils.py * Update src/transformers/modeling_tf_utils.py * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_tf_utils.py * Update src/transformers/modeling_utils.py
-
Patrick von Platen authored
* improve plotting * better labels * fix time plot
-
Sylvain Gugger authored
* Bert base model card * Add metadata * Adapt examples * GPT2 model card * Remove the BERT model card * Change language code
-
Sylvain Gugger authored
* Bert base model card * Add metadata * Adapt examples * Comment on text generation * Update model_cards/bert-base-uncased-README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Funtowicz Morgan authored
* Add new parameter `pad_to_multiple_of` on tokenizers. * unittest for pad_to_multiple_of * Add .name when logging enum. * Fix missing .items() on dict in tests. * Add special check + warning if the tokenizer doesn't have proper pad_token. * Use the correct logger format specifier. * Ensure tokenizer with no pad_token do not modify the underlying padding strategy. * Skip test if tokenizer doesn't have pad_token * Fix RobertaTokenizer on empty input * Format. Signed-off-by:
Morgan Funtowicz <funtowiczmo@gmail.com> * fix and updating to simpler API Co-authored-by:
Thomas Wolf <thomwolf@users.noreply.github.com>
-
- 25 Jun, 2020 11 commits
-
-
Lysandre Debut authored
-
Joe Davison authored
* add initial fine-tuning guide * split code blocks to smaller segments * fix up trianer section of fine-tune doc * a few last typos * Update usage -> task summary link Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Lysandre Debut authored
* Refactor code samples * Test docstrings * Style * Tokenization examples * Run rust of tests * First step to testing source docs * Style and BART comment * Test the remainder of the code samples * Style * let to const * Formatting fixes * Ready for merge * Fix fixture + Style * Fix last tests * Update docs/source/quicktour.rst Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Addressing @sgugger's comments + Fix MobileBERT in TF Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Thomas Wolf authored
* avoid recursion in id checks for fast tokenizers * better typings and fix #5232 * align slow and fast tokenizers behaviors for Roberta and GPT2 * style and quality * fix tests - improve typings
-
Sylvain Gugger authored
-
Thomas Wolf authored
[Tokenization] Fix #5181 - make #5155 more explicit - move back the default logging level in tests to WARNING (#5252) * fix-5181 Padding to max sequence length while truncation to another length was wrong on slow tokenizers * clean up and fix #5155 * fix XLM test * Fix tests for Transfo-XL * logging only above WARNING in tests * switch slow tokenizers tests in @slow * fix Marian truncation tokenization test * style and quality * make the test a lot faster by limiting the sequence length used in tests
-
Sam Shleifer authored
-
Julien Chaumond authored
-
Anthony MOI authored
-
Moumeneb1 authored
-
Sam Shleifer authored
-
- 24 Jun, 2020 21 commits
-
-
Sylvain Gugger authored
* All done * Link to the tutorial * Typo fixes Co-authored-by:
Thomas Wolf <thomwolf@users.noreply.github.com> * Add metnion of the return_xxx args Co-authored-by:
Thomas Wolf <thomwolf@users.noreply.github.com>
-
Thomas Wolf authored
* update tests for fast tokenizers + fix small bug in saving/loading * better tests on serialization * fixing serialization * comment cleanup
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Use _static from master everywhere * Copy to existing too
-
Victor SANH authored
* fix weirdness in roberta/bart for mnli trained checkpoints * black compliance * isort code check
-
Julien Chaumond authored
-
Setu Shah authored
-
Sylvain Gugger authored
-
Sai Saketh Aluru authored
* Add dehatebert-mono-arabic readme card * Update dehatebert-mono-arabic model card * model cards for Hate-speech-CNERG models
-
Lysandre Debut authored
* Cleaning TensorFlow models Update all classes stylr * Don't average loss
-
Sylvain Gugger authored
-
Ali Modarressi authored
-
Sylvain Gugger authored
* Try with the same command * Try like this
-
Sylvain Gugger authored
-
Patrick von Platen authored
* fix use cache * add bart use cache * fix bart * finish bart
-
Sylvain Gugger authored
-
Patrick von Platen authored
-
Patrick von Platen authored
* add benchmark for all kinds of models * improved import * delete bogus files * make style
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
flozi00 authored
* Create README.md * Update model_cards/a-ware/roberta-large-squad-classification/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-