- 24 Feb, 2021 1 commit
-
-
abhishek thakur authored
* convbert conversion test * fin * fin * fin * clean up tf<->pt conversion * remove from_pt Co-authored-by:patrickvonplaten <patrick.v.platen@gmail.com>
-
- 22 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
* Deprecate prepare_seq2seq_batch * Fix last tests * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Suraj Patil <surajp815@gmail.com> * More review comments Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Suraj Patil <surajp815@gmail.com>
-
Julien Plu authored
* AMP * Add LED * Apply style * Fix longformer
-
- 19 Feb, 2021 9 commits
-
-
Pengcheng He authored
* Integrate DeBERTa v2(the 1.5B model surpassed human performance on SuperGLUE); Add DeBERTa v2 900M,1.5B models; * DeBERTa-v2 * Fix v2 model loading issue (#10129) * Doc members * Update src/transformers/models/deberta/modeling_deberta.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Address Sylvain's comments * Address Patrick's comments Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Style Co-authored-by:
Lysandre <lysandre.debut@reseau.eseo.fr> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Julien Plu authored
* Fix AMP and XLA * Remove useless var
-
Julien Plu authored
* Fix AMP * Apply style * Remove unused import
-
Julien Plu authored
-
Julien Plu authored
* Fix XLA * Rework cast * Apply style
-
Julien Plu authored
* Fix AMP * Trigger CI * Rework cast
-
Julien Plu authored
* Fix AMP * Rework cast * Apply style
-
Stas Bekman authored
* implement --fp16_full_eval * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * style * add test Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Stas Bekman authored
-
- 18 Feb, 2021 4 commits
-
-
Stas Bekman authored
* memory tracker metrics * go back to eval for somewhat consistency * handle no-gpu case * deal with stackable eval calls * restore callback order * style * simplify the API * add test * docs * consistently use eval_ prefix * improve docs * Update src/transformers/trainer_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * rename method * style Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Julien Plu authored
* rework savedmodel slow test * Improve savedmodel tests * Remove useless content
-
Julien Plu authored
-
Julien Plu authored
* Fix XLA and AMP * Fix AMP and XLA * Apply style * Apply Patrick's comment
-
- 17 Feb, 2021 4 commits
-
-
Julien Plu authored
* Fix XLA and AMP * Apply style * Remove useless cast
-
Julien Plu authored
* Fix Flaubert and XLM * Remove useless cast * Tiny fix * Tiny fix
-
Julien Plu authored
* Update BART * Update Blenderbot * Update BlenderbotSmall * Update Marian * Update MBart * Update MBart * Update Pegasus * Update template * Fix Marian and Pegasus * Apply style * Default initializer * Default initializer * Default initializer * Remove int32 casts * Fix template * Remove more cast
-
Daniel Stancl authored
* Fix head_mask and decoder_head_mask in TFT5 models * Enable test_headmasking both fot TFT5 tester and TFT5EncoderOnly tester Co-authored-by:patrickvonplaten <patrick.v.platen@gmail.com>
-
- 16 Feb, 2021 2 commits
-
-
Sylvain Gugger authored
-
Julien Plu authored
-
- 15 Feb, 2021 5 commits
-
-
Lysandre Debut authored
Co-authored-by:
Quentin Lhoest <lhoest.q@gmail.com> Co-authored-by:
Quentin Lhoest <lhoest.q@gmail.com>
-
Julien Plu authored
-
Suraj Patil authored
* add tokenizer for mBART-50 * update tokenizers * make src_lang and tgt_lang optional * update tokenizer test * add setter * update docs * update conversion script * update docs * update conversion script * update tokenizer * update test * update docs * doc * address Sylvain's suggestions * fix test * fix formatting * nits
-
Julien Plu authored
* Add check-ops script * Finish to implement check_tf_ops and start the test * Make the test mandatory only for BERT * Update tf_ops folder * Remove useless classes * Add the ONNX test for GPT2 and BART * Add a onnxruntime slow test + better opset flexibility * Fix test + apply style * fix tests * Switch min opset from 12 to 10 * Update src/transformers/file_utils.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * Fix GPT2 * Remove extra shape_list usage * Fix GPT2 * Address Morgan's comments Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Nicolas Patry authored
Fixes #10168
-
- 13 Feb, 2021 1 commit
-
-
Nicolas Patry authored
* Conversion from slow to fast for BPE spm vocabs contained an error. - There is only 1 test currently (tokenizers + slow) that used the modified path and it's reformer, which does not contain any ids modification so the bug was silent for now. - The real issue is that vocab variable was overloaded by SentencePieceExtractor, leading to Slow specific vocab oddities to be completely ignored - The bug was reported here https://github.com/huggingface/transformers/issues/9518 - Ran the complete tokenization test suite with slow without error (`RUN_SLOW=1 pytest -sv tests/test_tokenization_*`) * Remove rebase error. * Adding the fixture.
-
- 12 Feb, 2021 2 commits
-
-
Julien Chaumond authored
-
Julien Chaumond authored
* [hf_api] delete deprecated methods and tests cc @lhoestq * Update test_hf_api.py
-
- 11 Feb, 2021 1 commit
-
-
Patrick von Platen authored
* save intermediate * finish batch the same as fairseq * add normalization * fix batched input * add better comment * Update src/transformers/models/wav2vec2/modeling_wav2vec2.py * add nice docstring * add tokenizer tests * make all slow tests pass * finish PR * correct import
-
- 10 Feb, 2021 3 commits
-
-
Suraj Patil authored
* add forced logits processors * delete adjust_logits method * add forced_eos_token_id argument in config * add tests for forced logits processors * update gen utils tests * add forced option to tf generate * remove adjust_logits method from tf models * update adjust_logits for marian * delete _force_token_id_to_be_generated method * style * import warnings * pass max_length to _get_logits_processor * set forced_eos_token_id to None * set forced attributes in conf utils * typo * fix rag generate * add forced_eos_token_id in rag config * remove force_bos_token_to_be_generated from BartConfig * remove _force_token_ids_generation from FSMT * nit * fix negative constant * apply suggestions from code review
-
Julien Plu authored
* Fix test * Remove commented test * Fix name * Apply style * Fix check copies * Remove prints * Restore boolean * Fix reshape
-
Lysandre Debut authored
-
- 09 Feb, 2021 3 commits
-
-
abhishek thakur authored
-
Daniel Stancl authored
* Add head masking to TF LED * Add head_mask to Longformer + one doc piece to LED * Fix integration tests
-
Patrick von Platen authored
* add wav2vec2CTC and deprecate for maskedlm * remove from docs
-
- 08 Feb, 2021 3 commits
-
-
sandip authored
-
demSd authored
* claiming this issue * Integration test for BertGeneration(Encoder and Decoder) * fix code quality
-
Patrick von Platen authored
-