- 03 Apr, 2020 9 commits
-
-
Lysandre Debut authored
* Electra wip * helpers * Electra wip * Electra v1 * ELECTRA may be saved/loaded * Generator & Discriminator * Embedding size instead of halving the hidden size * ELECTRA Tokenizer * Revert BERT helpers * ELECTRA Conversion script * Archive maps * PyTorch tests * Start fixing tests * Tests pass * Same configuration for both models * Compatible with base + large * Simplification + weight tying * Archives * Auto + Renaming to standard names * ELECTRA is uncased * Tests * Slight API changes * Update tests * wip * ElectraForTokenClassification * temp * Simpler arch + tests Removed ElectraForPreTraining which will be in a script * Conversion script * Auto model * Update links to S3 * Split ElectraForPreTraining and ElectraForTokenClassification * Actually test PreTraining model * Remove num_labels from configuration * wip * wip * From discriminator and generator to electra * Slight API changes * Better naming * TensorFlow ELECTRA tests * Accurate conversion script * Added to conversion script * Fast ELECTRA tokenizer * Style * Add ELECTRA to README * Modeling Pytorch Doc + Real style * TF Docs * Docs * Correct links * Correct model intialized * random fixes * style * Addressing Patrick's and Sam's comments * Correct links in docs
-
Yohei Tamura authored
* BertJapaneseTokenizer accept options for mecab * black * fix mecab_option to Option[str]
-
HUSEIN ZOLKEPLI authored
* add bert bahasa readme * update readme * update readme * added xlnet * added tiny-bert and fix xlnet readme * added albert base
-
ahotrod authored
Update AutoModel & AutoTokernizer loading.
-
ahotrod authored
-
HenrykBorzymowski authored
* added model_cards for polish squad models * corrected mistake in polish design cards Co-authored-by:Henryk Borzymowski <henryk.borzymowski@pwc.com>
-
redewiedergabe authored
* Create README.md * added meta block (language: german) * Added additional information about test data
-
ahotrod authored
-
Henryk Borzymowski authored
-
- 02 Apr, 2020 5 commits
-
-
Patrick von Platen authored
-
Nicolas authored
* Resizing embedding matrix after sending it to the optimizer prevents from updating the newly resized matrix. * Remove space for style matter
-
Mark Kockerbeck authored
-
Patrick von Platen authored
* solve conflicts * improve comments
-
Patrick von Platen authored
* replace heavy t5 models with tiny random models as was done by sshleifer * fix isort
-
- 01 Apr, 2020 7 commits
-
-
Patrick von Platen authored
* change tf t5 argument naming for TF 2.2 * correct bug in testing
-
Patrick von Platen authored
-
Anirudh Srinivasan authored
-
Patrick von Platen authored
[T5, Testst] Add extensive hard-coded integration tests and make sure PT and TF give equal results (#3550) * add some t5 integration tests * finish summarization and translation integration tests for T5 - results loook good * add tf test * fix == vs is bug * fix tf beam search error and make tf t5 tests pass
-
HUSEIN ZOLKEPLI authored
* add bert bahasa readme * update readme * update readme * added xlnet * added tiny-bert and fix xlnet readme
-
Manuel Romero authored
Create model card for: distilbert-multi-finetuned-for-xqua-on-tydiqa
-
Julien Chaumond authored
* Start cleaning examples * Fixup
-
- 31 Mar, 2020 16 commits
-
-
Patrick von Platen authored
* add bad words list * make style * add bad_words_tokens * make style * better naming * make style * fix typo
-
Patrick von Platen authored
* fix conflicts * add model size argument to summarization * correct wrong import * fix isort * correct imports * other isort make style * make style
-
Manuel Romero authored
- Show that the last uploaded version was trained on more data (custom_license files)
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Manuel Romero authored
Fix typo
-
Manuel Romero authored
-
Branden Chan authored
* Create README.md * Update README.md
-
Manuel Romero authored
-
Gabriele Sarti authored
-
Manuel Romero authored
-
Manuel Romero authored
-
Sho Arora authored
-
Leandro von Werra authored
* feat: add model card bert-imdb * feat: add model card gpt2-imdb-pos * feat: add model card gpt2-imdb
-
- 30 Mar, 2020 3 commits
-
-
Ethan Perez authored
* Using loaded checkpoint with --do_predict Without this fix, I'm getting near-random validation performance for a trained model, and the validation performance differs per validation run. I think this happens since the `model` variable isn't set with the loaded checkpoint, so I'm using a randomly initialized model. Looking at the model activations, they differ each time I run evaluation (but they don't with this fix). * Update checkpoint loading * Fixing model loading
-
Sam Shleifer authored
-
dougian authored
Co-authored-by:Ioannis Douratsos <ioannisd@amazon.com>
-