1. 17 Apr, 2020 1 commit
  2. 16 Apr, 2020 1 commit
    • Patrick von Platen's avatar
      [TFT5, Cache] Add cache to TFT5 (#3772) · 38f7461d
      Patrick von Platen authored
      * correct gpt2 test inputs
      
      * make style
      
      * delete modeling_gpt2 change in test file
      
      * translate from pytorch
      
      * correct tests
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * make tensorflow t5 caching work
      
      * make style
      
      * clean reorder cache
      
      * remove unnecessary spaces
      
      * fix test
      38f7461d
  3. 14 Apr, 2020 1 commit
  4. 09 Apr, 2020 1 commit
  5. 01 Apr, 2020 1 commit
  6. 30 Mar, 2020 1 commit
    • Patrick von Platen's avatar
      [T5] make decoder input ids optional for t5 training (#3521) · 75ec6c9e
      Patrick von Platen authored
      * make decoder input ids optional for t5 training
      
      * lm_lables should not be shifted in t5
      
      * add tests
      
      * finish shift right functionality for PT T5
      
      * move shift right to correct class
      
      * cleaner code
      
      * replace -100 values with pad token id
      
      * add assert statement
      
      * remove unnecessary for loop
      
      * make style
      75ec6c9e
  7. 19 Mar, 2020 1 commit
    • Patrick von Platen's avatar
      Support T5 Generation (#3228) · bbf26c4e
      Patrick von Platen authored
      
      
      * fix conflicts
      
      * update bart max length test
      
      * correct spelling mistakes
      
      * implemented model specific encode function
      
      * fix merge conflicts
      
      * better naming
      
      * save intermediate state -> need to rethink strucuture a bit
      
      * leave tf problem as it is for now
      
      * current version
      
      * add layers.pop
      
      * remove ipdb
      
      * make style
      
      * clean return cut decoding
      
      * remove ipdbs
      
      * Fix restoring layers in the decoders that doesnt exists.
      
      * push good intermediate solution for now
      
      * fix conflicts
      
      * always good to refuse to merge conflicts when rebasing
      
      * fix small bug
      
      * improve function calls
      
      * remove unused file
      
      * add correct scope behavior for t5_generate
      Co-authored-by: default avatarMorgan Funtowicz <funtowiczmo@gmail.com>
      bbf26c4e
  8. 26 Feb, 2020 1 commit
  9. 20 Feb, 2020 1 commit
    • Sam Shleifer's avatar
      New BartModel (#2745) · 53ce3854
      Sam Shleifer authored
      * Results same as fairseq
      * Wrote a ton of tests
      * Struggled with api signatures
      * added some docs
      
      53ce3854
  10. 06 Jan, 2020 2 commits
  11. 22 Dec, 2019 7 commits
  12. 21 Dec, 2019 2 commits
    • Aymeric Augustin's avatar
      Reformat source code with black. · fa84ae26
      Aymeric Augustin authored
      This is the result of:
      
          $ black --line-length 119 examples templates transformers utils hubconf.py setup.py
      
      There's a lot of fairly long lines in the project. As a consequence, I'm
      picking the longest widely accepted line length, 119 characters.
      
      This is also Thomas' preference, because it allows for explicit variable
      names, to make the code easier to understand.
      fa84ae26
    • Aymeric Augustin's avatar
      Take advantage of the cache when running tests. · b670c266
      Aymeric Augustin authored
      Caching models across test cases and across runs of the test suite makes
      slow tests somewhat more bearable.
      
      Use gettempdir() instead of /tmp in tests. This makes it easier to
      change the location of the cache with semi-standard TMPDIR/TEMP/TMP
      environment variables.
      
      Fix #2222.
      b670c266
  13. 16 Dec, 2019 1 commit
  14. 10 Dec, 2019 1 commit
  15. 08 Nov, 2019 1 commit
  16. 07 Nov, 2019 1 commit
  17. 06 Nov, 2019 1 commit