- 18 Nov, 2020 15 commits
-
-
Vishal Singh authored
Modified Model in Action section. The class `AutoModelWithLMHead` is deprecated so changed it to `AutoModelForSeq2SeqLM` for encoder-decoder models. Removed duplicate eos token.
-
smanjil authored
* replace performance table with markdown * Update model_cards/smanjil/German-MedBERT/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
hhou435 authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Patrick von Platen authored
* improve summary * small fixes * cleaned line length * correct "" formatting * apply sylvains suggestions
-
Nicola De Cao authored
* Adding PrefixConstrainedLogitsProcessor * fixing RAG and style_doc * fixing black (v20 instead of v19) * Improving doc in generation_logits_process.py * Improving docs and typing in generation_utils.py * docs improvement * adding test and fixing doc typo * fixing doc_len * isort on test * fixed test * improve docstring a bit Co-authored-by:Patrick von Platen <patrick.v.platen@gmail.com>
-
Julien Plu authored
* New TF loading weights * apply style * Better naming * Largely comment the loading method * Apply style * Address Patrick's comments * Remove useless line of code * Update Docstring * Address Sylvain's and Lysandre's comments * Simplify the names computation * Typos
-
Ratthachat (Jung) authored
(one line typo)
-
Stas Bekman authored
-
Stas Bekman authored
-
Micha艂 Pogoda authored
Multiline string informing about missing PyTorch/TensorFlow had missing space.
-
Sylvain Gugger authored
-
Benjamin Minixhofer authored
* make tr_loss regular float * Revert "make tr_loss regular float" This reverts commit c9d7ccfaf0c4387187b0841694f01ec0ffd5f4ba. * reset loss at each logging step * keep track of total loss with _total_loss_scalar * add remaining tr_loss at the end
-
cronoik authored
-
- 17 Nov, 2020 14 commits
-
-
Sylvain Gugger authored
-
Caitlin Ostroff authored
* Add Harry Potter Model * Update model_cards/ceostroff/harry-potter-gpt2-fanfiction/README.md * Update model_cards/ceostroff/harry-potter-gpt2-fanfiction/README.md * Update model_cards/ceostroff/harry-potter-gpt2-fanfiction/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
Sylvain Gugger authored
* Remove old deprecated arguments Co-authored-by:
LysandreJik <lysandre.debut@reseau.eseo.fr> * Remove needless imports * Fix tests Co-authored-by:
LysandreJik <lysandre.debut@reseau.eseo.fr>
-
Lysandre Debut authored
* Tokenizers should be framework agnostic * Run the slow tests * Not testing * Fix documentation * Apply suggestions from code review Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Sylvain Gugger authored
-
Stas Bekman authored
-
Sylvain Gugger authored
* First fixes * Fix imports and add init * Fix typo * Move init to final dest * Fix tokenization import * More fixes * Styling
-
Julien Chaumond authored
* <small>tiny typo</small> * Tokenizers: ability to load from model subfolder * use subfolder for local files as well * Uniformize model shortcut name => model id * from s3 => from huggingface.co Co-authored-by:Quentin Lhoest <lhoest.q@gmail.com>
-
Sylvain Gugger authored
-
sgugger authored
-
Patrick von Platen authored
* add docs * make style
-
Patrick von Platen authored
* add mt5 and t5v1_1 model * fix tests * correct some imports * add tf model * finish tf t5 * improve examples * fix copies * clean doc
-
fajri91 authored
-
Sylvain Gugger authored
* Put models in subfolders * Styling * Fix imports in tests * More fixes in test imports * Sneaky hidden imports * Fix imports in doc files * More sneaky imports * Finish fixing tests * Fix examples * Fix path for copies * More fixes for examples * Fix dummy files * More fixes for example * More model import fixes * Is this why you're unhappy GitHub? * Fix imports in conver command
-
- 16 Nov, 2020 10 commits
-
-
Julien Plu authored
* Fix mixed precision issue for GPT2 * Forgot one cast * oops * Forgotten casts
-
Sylvain Gugger authored
* Use the CI to identify failing tests * Remove from all examples and tests * More default switch * Fixes * More test fixes * More fixes * Last fixes hopefully * Use the CI to identify failing tests * Remove from all examples and tests * More default switch * Fixes * More test fixes * More fixes * Last fixes hopefully * Run on the real suite * Fix slow tests
-
Sylvain Gugger authored
-
LSinev authored
* Fix passing token_type_ids during GPT2DoubleHeadsModel.generate() if used and for GPT2LMHeadModel too * Update tests to check token_type_ids usage in GPT2 models
-
Yusuke Mori authored
* Simply insert T5Tokenizer's prepare_seq2seq_batch * Update/Add some 'import' * fix RunTimeError caused by '.view' * Moves .view related error avoidance from seq2seq_trainer to inside prophetnet * Update test_tokenization_prophetnet.py * Format the test code with black * Re-format the test code * Update test_tokenization_prophetnet.py * Add importing require_torch in the test code * Add importing BatchEncoding in the test code * Re-format the test code on Colab
-
Stas Bekman authored
* [doc] typo fix @sgugger * Update src/transformers/modeling_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Branden Chan authored
-
Mehrdad Farahani authored
-
Mehrdad Farahani authored
-
zhezhaoa authored
* Create README.md * Update model_cards/uer/chinese_roberta_L-2_H-128/README.md Co-authored-by:Julien Chaumond <chaumond@gmail.com>
-
- 15 Nov, 2020 1 commit
-
-
Thomas Wolf authored
[breaking|pipelines|tokenizers] Adding slow-fast tokenizers equivalence tests pipelines - Removing sentencepiece as a required dependency (#8073) * Fixing roberta for slow-fast tests * WIP getting equivalence on pipelines * slow-to-fast equivalence - working on question-answering pipeline * optional FAISS tests * Pipeline Q&A * Move pipeline tests to their own test job again * update tokenizer to add sequence id methods * update to tokenizers 0.9.4 * set sentencepiecce as optional * clean up squad * clean up pipelines to use sequence_ids * style/quality * wording * Switch to use_fast = True by default * update tests for use_fast at True by default * fix rag tokenizer test * removing protobuf from required dependencies * fix NER test for use_fast = True by default * fixing example tests (Q&A examples use slow tokenizers for now) * protobuf in main deps extras["sentencepiece"] and example deps * fix protobug install test * try to fix seq2seq by switching to slow tokenizers for now * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-