1. 11 Nov, 2021 1 commit
  2. 05 Aug, 2021 1 commit
  3. 13 Jul, 2021 1 commit
  4. 23 Jun, 2021 1 commit
  5. 01 Jun, 2021 1 commit
  6. 26 May, 2021 1 commit
    • Daniel Stancl's avatar
      Fix usage of head masks by TF encoder-decoder models' `generate()` function (#11775) · 0b933584
      Daniel Stancl authored
      * Fix Bart
      
      * Fix Blenderbot{,_small}
      
      * Fix LED
      
      * Fix Marian
      
      * Fix MBart
      
      * Fix Pegasus
      
      * Fix T5
      
      * Add test for generation with head_mask
      
      * Add a common TF test
      
      * Override a test for the LED model as head masking is not yet properly implemented
      
      * Remove all head_masks from input preparation for LED
      
      * Drop masking for T5 as it needs a bit of refactor
      0b933584
  7. 19 Feb, 2021 1 commit
  8. 18 Feb, 2021 1 commit
  9. 17 Feb, 2021 1 commit
  10. 15 Feb, 2021 1 commit
    • Julien Plu's avatar
      Check TF ops for ONNX compliance (#10025) · c8d3fa0d
      Julien Plu authored
      
      
      * Add check-ops script
      
      * Finish to implement check_tf_ops and start the test
      
      * Make the test mandatory only for BERT
      
      * Update tf_ops folder
      
      * Remove useless classes
      
      * Add the ONNX test for GPT2 and BART
      
      * Add a onnxruntime slow test + better opset flexibility
      
      * Fix test + apply style
      
      * fix tests
      
      * Switch min opset from 12 to 10
      
      * Update src/transformers/file_utils.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Fix GPT2
      
      * Remove extra shape_list usage
      
      * Fix GPT2
      
      * Address Morgan's comments
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      c8d3fa0d
  11. 29 Jan, 2021 1 commit
  12. 27 Jan, 2021 1 commit
  13. 26 Jan, 2021 1 commit
    • Daniel Stancl's avatar
      Add head_mask/decoder_head_mask for TF BART models (#9639) · 1867d9a8
      Daniel Stancl authored
      * Add head_mask/decoder_head_mask for TF BART models
      
      * Add head_mask and decoder_head_mask input arguments for TF BART-based
      models as a TF counterpart to the PR #9569
      
      * Add test_headmasking functionality to tests/test_modeling_tf_common.py
      
      * TODO: Add a test to verify that we can get a gradient back for
      importance score computation
      
      * Remove redundant #TODO note
      
      Remove redundant #TODO note from tests/test_modeling_tf_common.py
      
      * Fix assertions
      
      * Make style
      
      * Fix ...Model input args and adjust one new test
      
      * Add back head_mask and decoder_head_mask to BART-based ...Model
      after the last commit
      
      * Remove head_mask ande decoder_head_mask from input_dict
      in TF test_train_pipeline_custom_model as these two have different
      shape than other input args (Necessary for passing this test)
      
      * Revert adding global_rng in test_modeling_tf_common.py
      1867d9a8
  14. 11 Jan, 2021 1 commit
    • Julien Plu's avatar
      Full rework of the TF input/output embeddings and bias resizing (#9193) · 1243ee7d
      Julien Plu authored
      * Start rework resizing
      
      * Rework bias/decoder resizing
      
      * Full resizing rework
      
      * Full resizing rework
      
      * Start to update the models with the new approach
      
      * Finish to update the models
      
      * Update all the tests
      
      * Update the template
      
      * Fix tests
      
      * Fix tests
      
      * Test a new approach
      
      * Refactoring
      
      * Refactoring
      
      * Refactoring
      
      * New rework
      
      * Rework BART
      
      * Rework bert+blenderbot
      
      * Rework CTRL
      
      * Rework Distilbert
      
      * Rework DPR
      
      * Rework Electra
      
      * Rework Flaubert
      
      * Rework Funnel
      
      * Rework GPT2
      
      * Rework Longformer
      
      * Rework Lxmert
      
      * Rework marian+mbart
      
      * Rework mobilebert
      
      * Rework mpnet
      
      * Rework openai
      
      * Rework pegasus
      
      * Rework Roberta
      
      * Rework T5
      
      * Rework xlm+xlnet
      
      * Rework template
      
      * Fix TFT5EncoderOnly + DPRs
      
      * Restore previous methods
      
      * Fix Funnel
      
      * Fix CTRL and TransforXL
      
      * Apply style
      
      * Apply Sylvain's comments
      
      * Restore a test in DPR
      
      * Address the comments
      
      * Fix bug
      
      * Apply style
      
      * remove unused import
      
      * Fix test
      
      * Forgot a method
      
      * missing test
      
      * Trigger CI
      
      * naming update
      
      * Rebase
      
      * Trigger CI
      1243ee7d
  15. 07 Jan, 2021 1 commit
    • Julien Plu's avatar
      New serving (#9419) · 812045ad
      Julien Plu authored
      * Add a serving method
      
      * Add albert
      
      * Add serving for BERT and BART
      
      * Add more models
      
      * Finish the serving addition
      
      * Temp fix
      
      * Restore DPR
      
      * Fix funnel attribute
      
      * Fix attributes GPT2
      
      * Fix OpenAIGPT attribute
      
      * Fix T5 attributes
      
      * Fix Bart attributes
      
      * Fix TransfoXL attributes
      
      * Add versioning
      
      * better test
      
      * Update template
      
      * Fix Flaubert
      
      * Fix T5
      
      * Apply style
      
      * Remove unused imports
      
      * Deactivate extra parameters
      
      * Remove too long test + saved_model default to False
      
      * Ignore the saved model test for some models
      
      * Fix some inputs
      
      * Fix mpnet serving
      
      * Trigger CI
      
      * Address all comments
      812045ad
  16. 22 Dec, 2020 1 commit
  17. 14 Dec, 2020 2 commits
    • Julien Plu's avatar
      Fix T5 and BART for TF (#9063) · df3f4d2a
      Julien Plu authored
      * Fix T5 for graphe compilation+execution
      
      * Fix BART
      
      * Fix import
      
      * Fix naming
      
      * fix attribute name
      
      * Oops
      
      * fix import
      
      * fix tests
      
      * fix tests
      
      * Update test
      
      * Add mising import
      
      * Address Patrick's comments
      
      * Style
      
      * Address Patrick's comment
      df3f4d2a
    • Julien Plu's avatar
      Fix embeddings resizing in TF models (#8657) · 51d9c569
      Julien Plu authored
      * Resize the biases in same time than the embeddings
      
      * Trigger CI
      
      * Biases are not reset anymore
      
      * Remove get_output_embeddings + better LM model detection in generation utils
      
      * Apply style
      
      * First test on BERT
      
      * Update docstring + new name
      
      * Apply the new resizing logic to all the models
      
      * fix tests
      
      * Apply style
      
      * Update the template
      
      * Fix naming
      
      * Fix naming
      
      * Apply style
      
      * Apply style
      
      * Remove unused import
      
      * Revert get_output_embeddings
      
      * Trigger CI
      
      * Update num parameters
      
      * Restore get_output_embeddings in TFPretrainedModel and add comments
      
      * Style
      
      * Add decoder resizing
      
      * Style
      
      * Fix tests
      
      * Separate bias and decoder resize
      
      * Fix tests
      
      * Fix tests
      
      * Apply style
      
      * Add bias resizing in MPNet
      
      * Trigger CI
      
      * Apply style
      51d9c569
  18. 07 Dec, 2020 1 commit
  19. 30 Nov, 2020 1 commit
    • Ahmed Elnaggar's avatar
      Add T5 Encoder for Feature Extraction (#8717) · 40ecaf0c
      Ahmed Elnaggar authored
      
      
      * Add T5 Encoder class for feature extraction
      
      * fix T5 encoder add_start_docstrings indent
      
      * update init with T5 encoder
      
      * update init with TFT5ModelEncoder
      
      * remove TFT5ModelEncoder
      
      * change T5ModelEncoder order in init
      
      * add T5ModelEncoder to transformers init
      
      * clean T5ModelEncoder
      
      * update init with TFT5ModelEncoder
      
      * add TFModelEncoder for Tensorflow
      
      * update init with TFT5ModelEncoder
      
      * Update src/transformers/models/t5/modeling_t5.py
      
      change output from Seq2SeqModelOutput to BaseModelOutput
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * remove encoder_outputs
      
      1. remove encoder_outputs from the function call.
      2. remove the encoder_outputs If statement.
      3. remove isinstance from return_dict.
      
      * Authorize missing decoder keys
      
      * remove unnecessary input parameters
      
      remove pask_key_values and use_cache
      
      * remove use_cache
      
      remove use_cache from the forward method
      
      * add doctoring for T5 encoder
      
      add doctoring for T5 encoder with T5_ENCODER_INPUTS_DOCSTRING
      
      * change return_dict to dot access
      
      * add T5_ENCODER_INPUTS_DOCSTRING for TF T5
      
      * change TFT5Encoder output type to BaseModelOutput
      
      * remove unnecessary parameters for TFT5Encoder
      
      * remove unnecessary if statement
      
      * add import BaseModelOutput
      
      * fix BaseModelOutput typo to TFBaseModelOutput
      
      * update T5 doc with T5ModelEncoder
      
      * add T5ModelEncoder to tests
      
      * finish pytorch
      
      * finish docs and mt5
      
      * add mtf to init
      
      * fix init
      
      * remove n_positions
      
      * finish PR
      
      * Update src/transformers/models/mt5/modeling_mt5.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/models/t5/modeling_t5.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/models/t5/modeling_tf_t5.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/models/mt5/modeling_tf_mt5.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * make style
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      40ecaf0c
  20. 17 Nov, 2020 1 commit
    • Patrick von Platen's avatar
      T5 & mT5 (#8552) · 86822a35
      Patrick von Platen authored
      * add mt5 and t5v1_1 model
      
      * fix tests
      
      * correct some imports
      
      * add tf model
      
      * finish tf t5
      
      * improve examples
      
      * fix copies
      
      * clean doc
      86822a35
  21. 16 Nov, 2020 1 commit
    • Sylvain Gugger's avatar
      Switch `return_dict` to `True` by default. (#8530) · 1073a2bd
      Sylvain Gugger authored
      * Use the CI to identify failing tests
      
      * Remove from all examples and tests
      
      * More default switch
      
      * Fixes
      
      * More test fixes
      
      * More fixes
      
      * Last fixes hopefully
      
      * Use the CI to identify failing tests
      
      * Remove from all examples and tests
      
      * More default switch
      
      * Fixes
      
      * More test fixes
      
      * More fixes
      
      * Last fixes hopefully
      
      * Run on the real suite
      
      * Fix slow tests
      1073a2bd
  22. 13 Nov, 2020 1 commit
  23. 30 Oct, 2020 1 commit
    • Lysandre Debut's avatar
      Ci test tf super slow (#8007) · 10f8c636
      Lysandre Debut authored
      * Test TF GPU CI
      
      * Change cache
      
      * Fix missing torch requirement
      
      * Fix some model tests
      
      
      Style
      
      * LXMERT
      
      * MobileBERT
      
      * Longformer skip test
      
      * XLNet
      
      * The rest of the tests
      
      * RAG goes OOM in multi gpu setup
      
      * YAML test files
      
      * Last fixes
      
      * Skip doctests
      
      * Fill mask tests
      
      * Yaml files
      
      * Last test fix
      
      * Style
      
      * Update cache
      
      * Change ONNX tests to slow + use tiny model
      10f8c636
  24. 21 Oct, 2020 1 commit
    • Sam Shleifer's avatar
      Add TFBartForConditionalGeneration (#5411) · 82984215
      Sam Shleifer authored
      
      
      * half done
      
      * doc improvement
      
      * Cp test file
      
      * brokedn
      
      * broken test
      
      * undo some mess
      
      * ckpt
      
      * borked
      
      * Halfway
      
      * 6 passing
      
      * boom boom
      
      * Much progress but still 6
      
      * boom boom
      
      * merged master
      
      * 10 passing
      
      * boom boom
      
      * Style
      
      * no t5 changes
      
      * 13 passing
      
      * Integration test failing, but not gibberish
      
      * Frustrated
      
      * Merged master
      
      * 4 fail
      
      * 4 fail
      
      * fix return_dict
      
      * boom boom
      
      * Still only 4
      
      * prepare method
      
      * prepare method
      
      * before delete classif
      
      * Skip tests to avoid adding boilerplate
      
      * boom boom
      
      * fast tests passing
      
      * style
      
      * boom boom
      
      * Switch to supporting many input types
      
      * remove FIXMENORM
      
      * working
      
      * Fixed past_key_values/decoder_cached_states confusion
      
      * new broken test
      
      * Fix attention mask kwarg name
      
      * undo accidental
      
      * Style and reviewers
      
      * style
      
      * Docs and common tests
      
      * Cleaner assert messages
      
      * copy docs
      
      * style issues
      
      * Sphinx fix
      
      * Simplify caching logic
      
      * test does not require torch
      
      * copy _NoLayerEmbedTokens
      
      * Update src/transformers/modeling_tf_bart.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update tests/test_modeling_tf_bart.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/modeling_tf_bart.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/modeling_tf_bart.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/modeling_tf_bart.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Line length and dont document None
      
      * Add pipeline test coverage
      
      * assert msg
      
      * At parity
      
      * Assert messages
      
      * mark slow
      
      * Update compile test
      
      * back in init
      
      * Merge master
      
      * Fix tests
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      82984215
  25. 18 Oct, 2020 1 commit
    • Thomas Wolf's avatar
      [Dependencies|tokenizers] Make both SentencePiece and Tokenizers optional dependencies (#7659) · ba8c4d0a
      Thomas Wolf authored
      * splitting fast and slow tokenizers [WIP]
      
      * [WIP] splitting sentencepiece and tokenizers dependencies
      
      * update dummy objects
      
      * add name_or_path to models and tokenizers
      
      * prefix added to file names
      
      * prefix
      
      * styling + quality
      
      * spliting all the tokenizer files - sorting sentencepiece based ones
      
      * update tokenizer version up to 0.9.0
      
      * remove hard dependency on sentencepiece 馃帀
      
      * and removed hard dependency on tokenizers 馃帀
      
      
      
      * update conversion script
      
      * update missing models
      
      * fixing tests
      
      * move test_tokenization_fast to main tokenization tests - fix bugs
      
      * bump up tokenizers
      
      * fix bert_generation
      
      * update ad fix several tokenizers
      
      * keep sentencepiece in deps for now
      
      * fix funnel and deberta tests
      
      * fix fsmt
      
      * fix marian tests
      
      * fix layoutlm
      
      * fix squeezebert and gpt2
      
      * fix T5 tokenization
      
      * fix xlnet tests
      
      * style
      
      * fix mbart
      
      * bump up tokenizers to 0.9.2
      
      * fix model tests
      
      * fix tf models
      
      * fix seq2seq examples
      
      * fix tests without sentencepiece
      
      * fix slow => fast  conversion without sentencepiece
      
      * update auto and bert generation tests
      
      * fix mbart tests
      
      * fix auto and common test without tokenizers
      
      * fix tests without tokenizers
      
      * clean up tests lighten up when tokenizers + sentencepiece are both off
      
      * style quality and tests fixing
      
      * add sentencepiece to doc/examples reqs
      
      * leave sentencepiece on for now
      
      * style quality split hebert and fix pegasus
      
      * WIP Herbert fast
      
      * add sample_text_no_unicode and fix hebert tokenization
      
      * skip FSMT example test for now
      
      * fix style
      
      * fix fsmt in example tests
      
      * update following Lysandre and Sylvain's comments
      
      * Update src/transformers/testing_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/testing_utils.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      ba8c4d0a
  26. 01 Oct, 2020 1 commit
  27. 01 Sep, 2020 1 commit
    • Patrick von Platen's avatar
      [Generate] Facilitate PyTorch generate using `ModelOutputs` (#6735) · afc4ece4
      Patrick von Platen authored
      * fix generate for GPT2 Double Head
      
      * fix gpt2 double head model
      
      * fix  bart / t5
      
      * also add for no beam search
      
      * fix no beam search
      
      * fix encoder decoder
      
      * simplify t5
      
      * simplify t5
      
      * fix t5 tests
      
      * fix BART
      
      * fix transfo-xl
      
      * fix conflict
      
      * integrating sylvains and sams comments
      
      * fix tf past_decoder_key_values
      
      * fix enc dec test
      afc4ece4
  28. 26 Aug, 2020 1 commit
  29. 24 Aug, 2020 1 commit
  30. 13 Aug, 2020 1 commit
    • Stas Bekman's avatar
      cleanup tf unittests: part 2 (#6260) · e983da0e
      Stas Bekman authored
      * cleanup torch unittests: part 2
      
      * remove trailing comma added by isort, and which breaks flake
      
      * one more comma
      
      * revert odd balls
      
      * part 3: odd cases
      
      * more ["key"] -> .key refactoring
      
      * .numpy() is not needed
      
      * more unncessary .numpy() removed
      
      * more simplification
      e983da0e
  31. 05 Aug, 2020 1 commit
    • Sylvain Gugger's avatar
      Tf model outputs (#6247) · c67d1a02
      Sylvain Gugger authored
      * TF outputs and test on BERT
      
      * Albert to DistilBert
      
      * All remaining TF models except T5
      
      * Documentation
      
      * One file forgotten
      
      * TF outputs and test on BERT
      
      * Albert to DistilBert
      
      * All remaining TF models except T5
      
      * Documentation
      
      * One file forgotten
      
      * Add new models and fix issues
      
      * Quality improvements
      
      * Add T5
      
      * A bit of cleanup
      
      * Fix for slow tests
      
      * Style
      c67d1a02
  32. 07 Jul, 2020 1 commit
  33. 03 Jul, 2020 1 commit
  34. 01 Jul, 2020 1 commit
  35. 26 Jun, 2020 1 commit
  36. 24 Jun, 2020 1 commit
  37. 16 Jun, 2020 1 commit
  38. 01 May, 2020 1 commit
    • Julien Chaumond's avatar
      [ci] Load pretrained models into the default (long-lived) cache · f54dc3f4
      Julien Chaumond authored
      There's an inconsistency right now where:
      - we load some models into CACHE_DIR
      - and some models in the default cache
      - and often, in both for the same models
      
      When running the RUN_SLOW tests, this takes a lot of disk space, time, and bandwidth.
      
      I'd rather always use the default cache
      f54dc3f4
  39. 17 Apr, 2020 1 commit