1. 09 Apr, 2020 1 commit
  2. 06 Apr, 2020 1 commit
  3. 31 Mar, 2020 1 commit
  4. 26 Mar, 2020 1 commit
  5. 19 Mar, 2020 1 commit
    • Patrick von Platen's avatar
      Support T5 Generation (#3228) · bbf26c4e
      Patrick von Platen authored
      
      
      * fix conflicts
      
      * update bart max length test
      
      * correct spelling mistakes
      
      * implemented model specific encode function
      
      * fix merge conflicts
      
      * better naming
      
      * save intermediate state -> need to rethink strucuture a bit
      
      * leave tf problem as it is for now
      
      * current version
      
      * add layers.pop
      
      * remove ipdb
      
      * make style
      
      * clean return cut decoding
      
      * remove ipdbs
      
      * Fix restoring layers in the decoders that doesnt exists.
      
      * push good intermediate solution for now
      
      * fix conflicts
      
      * always good to refuse to merge conflicts when rebasing
      
      * fix small bug
      
      * improve function calls
      
      * remove unused file
      
      * add correct scope behavior for t5_generate
      Co-authored-by: default avatarMorgan Funtowicz <funtowiczmo@gmail.com>
      bbf26c4e
  6. 17 Mar, 2020 1 commit
  7. 11 Mar, 2020 4 commits
  8. 05 Mar, 2020 1 commit
  9. 03 Mar, 2020 1 commit
    • Patrick von Platen's avatar
      Add generate() functionality to TF 2.0 (#3063) · 41341003
      Patrick von Platen authored
      * add first copy past test to tf 2 generate
      
      * add tf top_k_top_p_filter fn
      
      * add generate function for TF
      
      * add generate function for TF
      
      * implemented generate for all models expect transfoXL
      
      * implemented generate for all models expect transfoXL
      
      * implemented generate for all models expect transfoXL
      
      * make style
      
      * change permission of test file to correct ones
      
      * delete ipdb
      
      * delete ipdb
      
      * fix bug and finish simple gpt2 integration test
      
      * clean test file
      
      * clean test file
      
      * make style
      
      * make style
      
      * make style
      
      * make style
      
      * change import style
      
      * change import style
      
      * make style
      
      * make style
      
      * add decorators
      
      * add decorators
      
      * fix tf ctrl bug dim => axis in TF
      
      * make style
      
      * make style
      
      * refactored test file
      
      * refactored test file
      
      * take out test_torch_tf_conversion if nothing is defined
      
      * take out test_torch_tf_conversion if nothing is defined
      
      * remove useless files
      
      * remove useless files
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * solve conflicts
      
      * solve conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * merge conflicts
      
      * delete ipdb
      
      * exposed top_k_top_p_filtering fns
      
      * delete weirdly created w! file
      
      * add comment to test tf common modeling
      
      * fix conflicts
      
      * fix conflicts
      
      * make style
      
      * merge conflicts
      
      * make style
      
      * change tf.tensor.shape to shape_list(tensor)
      41341003
  10. 02 Mar, 2020 1 commit
  11. 26 Feb, 2020 1 commit
  12. 24 Feb, 2020 1 commit
    • Patrick von Platen's avatar
      Add slow generate tests for pretrained lm models (#2909) · 17c45c39
      Patrick von Platen authored
      * add slow generate lm_model tests
      
      * fix conflicts
      
      * merge conflicts
      
      * fix conflicts
      
      * add slow generate lm_model tests
      
      * make style
      
      * delete unused variable
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * delete unused variable
      
      * fix conflicts
      
      * finished hard coded tests
      17c45c39
  13. 21 Feb, 2020 1 commit
    • Patrick von Platen's avatar
      Improve special_token_id logic in run_generation.py and add tests (#2885) · fc38d4c8
      Patrick von Platen authored
      
      
      * improving generation
      
      * finalized special token behaviour for no_beam_search generation
      
      * solved modeling_utils merge conflict
      
      * solve merge conflicts in modeling_utils.py
      
      * add run_generation improvements from PR #2749
      
      * adapted language generation to not use hardcoded -1 if no padding token is available
      
      * remove the -1 removal as hard coded -1`s are not necessary anymore
      
      * add lightweight language generation testing for randomely initialized models - just checking whether no errors are thrown
      
      * add slow language generation tests for pretrained models using hardcoded output with pytorch seed
      
      * delete ipdb
      
      * check that all generated tokens are valid
      
      * renaming
      
      * renaming Generation -> Generate
      
      * make style
      
      * updated so that generate_beam_search has same token behavior than generate_no_beam_search
      
      * consistent return format for run_generation.py
      
      * deleted pretrain lm generate tests -> will be added in another PR
      
      * cleaning of unused if statements and renaming
      
      * run_generate will always return an iterable
      
      * make style
      
      * consistent renaming
      
      * improve naming, make sure generate function always returns the same tensor, add docstring
      
      * add slow tests for all lmhead models
      
      * make style and improve example comments modeling_utils
      
      * better naming and refactoring in modeling_utils
      
      * improving generation
      
      * finalized special token behaviour for no_beam_search generation
      
      * solved modeling_utils merge conflict
      
      * solve merge conflicts in modeling_utils.py
      
      * add run_generation improvements from PR #2749
      
      * adapted language generation to not use hardcoded -1 if no padding token is available
      
      * remove the -1 removal as hard coded -1`s are not necessary anymore
      
      * add lightweight language generation testing for randomely initialized models - just checking whether no errors are thrown
      
      * add slow language generation tests for pretrained models using hardcoded output with pytorch seed
      
      * delete ipdb
      
      * check that all generated tokens are valid
      
      * renaming
      
      * renaming Generation -> Generate
      
      * make style
      
      * updated so that generate_beam_search has same token behavior than generate_no_beam_search
      
      * consistent return format for run_generation.py
      
      * deleted pretrain lm generate tests -> will be added in another PR
      
      * cleaning of unused if statements and renaming
      
      * run_generate will always return an iterable
      
      * make style
      
      * consistent renaming
      
      * improve naming, make sure generate function always returns the same tensor, add docstring
      
      * add slow tests for all lmhead models
      
      * make style and improve example comments modeling_utils
      
      * better naming and refactoring in modeling_utils
      
      * changed fast random lm generation testing design to more general one
      
      * delete in old testing design in gpt2
      
      * correct old variable name
      
      * temporary fix for encoder_decoder lm generation tests - has to be updated when t5 is fixed
      
      * adapted all fast random generate tests to new design
      
      * better warning description in modeling_utils
      
      * better comment
      
      * better comment and error message
      Co-authored-by: default avatarThomas Wolf <thomwolf@users.noreply.github.com>
      fc38d4c8
  14. 20 Feb, 2020 1 commit
    • Sam Shleifer's avatar
      New BartModel (#2745) · 53ce3854
      Sam Shleifer authored
      * Results same as fairseq
      * Wrote a ton of tests
      * Struggled with api signatures
      * added some docs
      
      53ce3854
  15. 04 Feb, 2020 3 commits
  16. 16 Jan, 2020 1 commit
  17. 14 Jan, 2020 1 commit
    • Lysandre's avatar
      Bias should be resized with the weights · 100e3b6f
      Lysandre authored
      Created a link between the linear layer bias and the model attribute bias. This does not change anything for the user nor for the conversion scripts, but allows the `resize_token_embeddings` method to resize the bias as well as the weights of the decoder.
      
      Added a test.
      100e3b6f
  18. 11 Jan, 2020 2 commits
  19. 10 Jan, 2020 1 commit
  20. 06 Jan, 2020 2 commits
  21. 23 Dec, 2019 1 commit
  22. 22 Dec, 2019 10 commits
  23. 21 Dec, 2019 2 commits
    • Aymeric Augustin's avatar
      Reformat source code with black. · fa84ae26
      Aymeric Augustin authored
      This is the result of:
      
          $ black --line-length 119 examples templates transformers utils hubconf.py setup.py
      
      There's a lot of fairly long lines in the project. As a consequence, I'm
      picking the longest widely accepted line length, 119 characters.
      
      This is also Thomas' preference, because it allows for explicit variable
      names, to make the code easier to understand.
      fa84ae26
    • Aymeric Augustin's avatar
      Take advantage of the cache when running tests. · b670c266
      Aymeric Augustin authored
      Caching models across test cases and across runs of the test suite makes
      slow tests somewhat more bearable.
      
      Use gettempdir() instead of /tmp in tests. This makes it easier to
      change the location of the cache with semi-standard TMPDIR/TEMP/TMP
      environment variables.
      
      Fix #2222.
      b670c266