"...lm-evaluation-harness.git" did not exist on "c0f0f7e1a696a16756f1835eaf8366d2b041c73b"
  1. 18 Apr, 2020 1 commit
    • Thomas Wolf's avatar
      Cleanup fast tokenizers integration (#3706) · 827d6d6e
      Thomas Wolf authored
      
      
      * First pass on utility classes and python tokenizers
      
      * finishing cleanup pass
      
      * style and quality
      
      * Fix tests
      
      * Updating following @mfuntowicz comment
      
      * style and quality
      
      * Fix Roberta
      
      * fix batch_size/seq_length inBatchEncoding
      
      * add alignement methods + tests
      
      * Fix OpenAI and Transfo-XL tokenizers
      
      * adding trim_offsets=True default for GPT2 et RoBERTa
      
      * style and quality
      
      * fix tests
      
      * add_prefix_space in roberta
      
      * bump up tokenizers to rc7
      
      * style
      
      * unfortunately tensorfow does like these - removing shape/seq_len for now
      
      * Update src/transformers/tokenization_utils.py
      Co-Authored-By: default avatarStefan Schweter <stefan@schweter.it>
      
      * Adding doc and docstrings
      
      * making flake8 happy
      Co-authored-by: default avatarStefan Schweter <stefan@schweter.it>
      827d6d6e
  2. 16 Apr, 2020 1 commit
  3. 08 Apr, 2020 1 commit
  4. 04 Apr, 2020 1 commit
  5. 24 Mar, 2020 1 commit
  6. 02 Mar, 2020 1 commit
  7. 07 Feb, 2020 1 commit
  8. 29 Jan, 2020 4 commits
  9. 15 Jan, 2020 1 commit
  10. 13 Jan, 2020 1 commit
  11. 07 Jan, 2020 1 commit
  12. 06 Jan, 2020 2 commits
  13. 05 Jan, 2020 1 commit
  14. 28 Dec, 2019 1 commit
  15. 23 Dec, 2019 1 commit
  16. 22 Dec, 2019 20 commits
  17. 21 Dec, 2019 1 commit
    • Aymeric Augustin's avatar
      Reformat source code with black. · fa84ae26
      Aymeric Augustin authored
      This is the result of:
      
          $ black --line-length 119 examples templates transformers utils hubconf.py setup.py
      
      There's a lot of fairly long lines in the project. As a consequence, I'm
      picking the longest widely accepted line length, 119 characters.
      
      This is also Thomas' preference, because it allows for explicit variable
      names, to make the code easier to understand.
      fa84ae26