1. 05 Apr, 2022 1 commit
  2. 04 Apr, 2022 1 commit
  3. 01 Apr, 2022 1 commit
  4. 30 Mar, 2022 1 commit
  5. 25 Mar, 2022 1 commit
  6. 23 Mar, 2022 2 commits
  7. 22 Mar, 2022 1 commit
  8. 21 Mar, 2022 1 commit
  9. 16 Mar, 2022 1 commit
  10. 08 Mar, 2022 1 commit
  11. 04 Mar, 2022 1 commit
  12. 25 Feb, 2022 1 commit
    • Yih-Dar's avatar
      Fix tf.concatenate + test past_key_values for TF models (#15774) · 8635407b
      Yih-Dar authored
      
      
      * fix wrong method name tf.concatenate
      
      * add tests related to causal LM / decoder
      
      * make style and quality
      
      * clean-up
      
      * Fix TFBertModel's extended_attention_mask when past_key_values is provided
      
      * Fix tests
      
      * fix copies
      
      * More tf.int8 -> tf.int32 in TF test template
      
      * clean-up
      
      * Update TF test template
      
      * revert the previous commit + update the TF test template
      
      * Fix TF template extended_attention_mask when past_key_values is provided
      
      * Fix some styles manually
      
      * clean-up
      
      * Fix ValueError: too many values to unpack in the test
      
      * Fix more: too many values to unpack in the test
      
      * Add a comment for extended_attention_mask when there is past_key_values
      
      * Fix TFElectra extended_attention_mask when past_key_values is provided
      
      * Add tests to other TF models
      
      * Fix for TF Electra test: add prepare_config_and_inputs_for_decoder
      
      * Fix not passing training arg to lm_head in TFRobertaForCausalLM
      
      * Fix tests (with past) for TF Roberta
      
      * add testing for pask_key_values for TFElectra model
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      8635407b
  13. 23 Feb, 2022 2 commits
  14. 15 Feb, 2022 1 commit
    • Patrick von Platen's avatar
      TF generate refactor - Greedy Search (#15562) · 2e12b907
      Patrick von Platen authored
      
      
      * TF generate start refactor
      
      * Add tf tests for sample generate
      
      * re-organize
      
      * boom boom
      
      * Apply suggestions from code review
      
      * re-add
      
      * add all code
      
      * make random greedy pass
      
      * make encoder-decoder random work
      
      * further improvements
      
      * delete bogus file
      
      * make gpt2 and t5 tests work
      
      * finish logits tests
      
      * correct logits processors
      
      * correct past / encoder_outputs drama
      
      * refactor some methods
      
      * another fix
      
      * refactor shape_list
      
      * fix more shape list
      
      * import shape
      _list
      
      * finish docs
      
      * fix imports
      
      * make style
      
      * correct tf utils
      
      * Fix TFRag as well
      
      * Apply Lysandre's and Sylvais suggestions
      
      * Update tests/test_generation_tf_logits_process.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Update src/transformers/tf_utils.py
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * remove cpu according to gante
      
      * correct logit processor
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      2e12b907
  15. 08 Feb, 2022 1 commit
  16. 01 Feb, 2022 2 commits
    • SaulLu's avatar
      fix the `tokenizer_config.json` file for the slow tokenizer when a fast... · 7b8bdd86
      SaulLu authored
      fix the `tokenizer_config.json` file for the slow tokenizer when a fast version is available (#15319)
      
      * add new test
      
      * update test
      
      * remove `tokenizer_file` from `additional_files_names` in `tokenization_utils_base.py`
      
      * add `tokenizer_file` for the fast only tokenizer
      
      * change global variables layoutxml
      
      * remove `"tokenizer_file"` from DPR tokenizer's Global variables
      
      * remove `tokenizer_file` from herbert slow tokenizer init
      
      * `"tokenizer_file"` from LED tokenizer's Global variables
      
      * remove `tokenizer_file` from mbart slow tokenizer init
      
      * remove `tokenizer_file` from slow tokenizer template
      
      * adapt to versioning
      
      * adapt the `test_tokenizer_mismatch_warning` test
      
      * clean test
      
      * clarify `VOCAB_FILES_NAMES` in tokenization_utils_fast.py
      
      * Revert "remove `tokenizer_file` from mbart slow tokenizer init"
      
      This reverts commit 0dbb723fa9c7599d4640fe30b3647a74eb4a64e1.
      
      * Revert "`"tokenizer_file"` from LED tokenizer's Global variables"
      
      This reverts commit 5a3f879bdd651233f3d74a3d1146c34cde82b0c2.
      
      * Revert "remove `tokenizer_file` from herbert slow tokenizer init"
      
      This reverts commit f5e10007b7b0ec5345e015b9de7ffec72c5407fd.
      
      * Revert "remove `"tokenizer_file"` from DPR tokenizer's Global variables"
      
      This reverts commit da0895330bedfafc81ae3073470a9348c669f032.
      
      * set `tokenizer_file` in super `__init__` of mbart
      7b8bdd86
    • Yih-Dar's avatar
      Fix TF Causal LM models' returned logits (#15256) · dc05dd53
      Yih-Dar authored
      
      
      * Fix TF Causal LM models' returned logits
      
      * Fix expected shape in the tests
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      dc05dd53
  17. 31 Jan, 2022 2 commits
  18. 24 Jan, 2022 1 commit
  19. 21 Jan, 2022 1 commit
  20. 19 Jan, 2022 1 commit
    • Matt's avatar
      Rename compute_loss in TF models (#15207) · 2708bfa1
      Matt authored
      * Rename compute_loss to hf_compute_loss to avoid conflicts with the new Keras method
      
      * make style
      
      * Adding deprecation warning to `compute_loss`
      
      * Fix sneaky reference to compute_loss
      
      * Replace logger.warning with warnings.warn
      
      * Clarifying warning and deprecation timeline
      2708bfa1
  21. 14 Jan, 2022 1 commit
  22. 11 Jan, 2022 2 commits
  23. 10 Jan, 2022 2 commits
  24. 22 Dec, 2021 1 commit
  25. 21 Dec, 2021 2 commits
    • Sylvain Gugger's avatar
      Mass conversion of documentation from rst to Markdown (#14866) · 27b3031d
      Sylvain Gugger authored
      * Convert docstrings of all configurations and tokenizers
      
      * Processors and fixes
      
      * Last modeling files and fixes to models
      
      * Pipeline modules
      
      * Utils files
      
      * Data submodule
      
      * All the other files
      
      * Style
      
      * Missing examples
      
      * Style again
      
      * Fix copies
      
      * Say bye bye to rst docstrings forever
      27b3031d
    • Sylvain Gugger's avatar
      Convert docstrings of modeling files (#14850) · 7af80f66
      Sylvain Gugger authored
      * Convert file_utils docstrings to Markdown
      
      * Test on BERT
      
      * Return block indent
      
      * Temporarily disable doc styler
      
      * Remove from quality checks as well
      
      * Remove doc styler mess
      
      * Remove check from circleCI
      
      * Fix typo
      
      * Convert file_utils docstrings to Markdown
      
      * Test on BERT
      
      * Return block indent
      
      * Temporarily disable doc styler
      
      * Remove from quality checks as well
      
      * Remove doc styler mess
      
      * Remove check from circleCI
      
      * Fix typo
      
      * Let's go on all other model files
      
      * Add templates too
      
      * Styling and quality
      7af80f66
  26. 17 Dec, 2021 1 commit
    • Daniel Stancl's avatar
      Implement head_mask for Flax BERT and other models copied from BERT (#14620) · ff066119
      Daniel Stancl authored
      * Implement head_mask for Flax BERT and other models copied from BERT
      
      * Remove `from jax._src.nn.functions import sigmoid`
      
      Remove `from jax._src.nn.functions import sigmoid` unintentionally added by IDE
      
      * Remove no more valid copy statement
      
      * Apply patil-suraj's suggestions from code review
      
      * Apply suggestions from the code review
      
      * Update Flax template
      
      * Fix a typo
      
      * Also update template for CausalLM modules
      ff066119
  27. 16 Dec, 2021 1 commit
  28. 13 Dec, 2021 2 commits
  29. 10 Dec, 2021 1 commit
  30. 30 Nov, 2021 1 commit
  31. 18 Nov, 2021 1 commit
  32. 11 Nov, 2021 1 commit
    • Suraj Patil's avatar
      Fix Flax params dtype (#13098) · e92190c0
      Suraj Patil authored
      
      
      * fix inits
      
      * fix embed dtype
      
      * fix embed dtype
      
      * add test to check default dtype
      
      * quality
      
      * add type conversion methods for flax models
      
      * more robust casting
      
      * cast sinusoidal positions
      
      * update pegasus
      
      * update albert
      
      * update test
      
      * make sure dtype is passed to every module
      
      * style
      
      * fix electra dense
      
      * fix t5
      
      * quality
      
      * add more tests
      
      * better name
      
      * use the dtype for lm head computation
      
      * fix albert
      
      * style
      
      * fix albert embed dtype
      
      * more tests
      
      * fix vision enc-dec
      
      * cleanup
      
      * fix embed dtype pegasus
      
      * fix default param test
      
      * doc
      
      * update template
      
      * fix final_logits_bias dtype
      
      * Apply suggestions from code review
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      
      * fix doc
      
      * fix doc
      
      * add detailed docstring for dtype parameter
      
      * remove un-necessary import
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      e92190c0