"vscode:/vscode.git/clone" did not exist on "ba3543b3fc9001231a9c010c737d49f7ad67811c"
  1. 11 Aug, 2020 10 commits
    • Abed khooli's avatar
      Create README.md (#6397) · 824e651e
      Abed khooli authored
      
      
      * Create README.md
      
      * Update model_cards/akhooli/gpt2-small-arabic-poetry/README.md
      
      * Update model_cards/akhooli/gpt2-small-arabic-poetry/README.md
      
      * Update model_cards/akhooli/gpt2-small-arabic-poetry/README.md
      
      * Update model_cards/akhooli/gpt2-small-arabic-poetry/README.md
      Co-authored-by: default avatarJulien Chaumond <chaumond@gmail.com>
      824e651e
    • guillaume-be's avatar
      [Performance improvement] "Bad tokens ids" optimization (#6064) · 40478291
      guillaume-be authored
      * Optimized banned token masking
      
      * Avoid duplicate EOS masking if in bad_words_id
      
      * Updated mask generation to handle empty banned token list
      
      * Addition of unit tests for the updated bad_words_ids masking
      
      * Updated timeout handling in `test_postprocess_next_token_scores_large_bad_words_list` unit test
      
      * Updated timeout handling in `test_postprocess_next_token_scores_large_bad_words_list` unit test (timeout does not work on Windows)
      
      * Moving Marian import to the test context to allow TF only environments to run
      
      * Moving imports to torch_available test
      
      * Updated operations device and test
      
      * Updated operations device and test
      
      * Added docstring and comment for in-place scores modification
      
      * Moving test to own test_generation_utils, use of lighter models for testing
      
      * removed unneded imports in test_modeling_common
      
      * revert formatting change for ModelTesterMixin
      
      * Updated caching, simplified eos token id test, removed unnecessary @require_torch
      
      * formatting compliance
      40478291
    • David LaPalomento's avatar
      Warn if debug requested without TPU fixes (#6308) (#6390) · 87e124c2
      David LaPalomento authored
      
      
      * Warn if debug requested without TPU fixes (#6308)
      Check whether a PyTorch compatible TPU is available before attempting to print TPU metrics after training has completed. This way, users who apply `--debug` without reading the documentation aren't suprised by a stacktrace.
      
      * Style
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      87e124c2
    • Junyuan Zheng's avatar
      Fix tokenizer saving and loading error (#6026) · cdf1f7ed
      Junyuan Zheng authored
      
      
      * fix tokenizer saving and loading bugs when adding AddedToken to additional special tokens
      
      * Add tokenizer test
      
      * Style
      
      * Style 2
      Co-authored-by: default avatarLysandre <lysandre.debut@reseau.eseo.fr>
      cdf1f7ed
    • Stas Bekman's avatar
      testing utils: capturing std streams context manager (#6231) · 83984a61
      Stas Bekman authored
      * testing utils: capturing std streams context manager
      
      * style
      
      * missing import
      
      * add the origin of this code
      83984a61
    • Stas Bekman's avatar
      add pl_glue example test (#6034) · f6c0680d
      Stas Bekman authored
      * add pl_glue example test
      
      * for now just test that it runs, next validate results of eval or predict?
      
      * complete the run_pl_glue test to validate the actual outcome
      
      * worked on my machine, CI gets less accuracy - trying higher epochs
      
      * match run_pl.sh hparms
      
      * more epochs?
      
      * trying higher lr
      
      * for now just test that the script runs to a completion
      
      * correct the comment
      
      * if cuda is available, add --fp16 --gpus=1 to cover more bases
      
      * style
      f6c0680d
    • Pradhy729's avatar
      Feed forward chunking (#6024) · b25cec13
      Pradhy729 authored
      
      
      * Chunked feed forward for Bert
      
      This is an initial implementation to test applying feed forward chunking for BERT.
      Will need additional modifications based on output and benchmark results.
      
      * Black and cleanup
      
      * Feed forward chunking in BertLayer class.
      
      * Isort
      
      * add chunking for all models
      
      * fix docs
      
      * Fix typo
      Co-authored-by: default avatarpatrickvonplaten <patrick.v.platen@gmail.com>
      b25cec13
    • Lysandre's avatar
      Add TPU testing once again · 8a3db6b3
      Lysandre authored
      8a3db6b3
    • zcain117's avatar
      Add missing docker arg for TPU CI. (#6393) · f65ac1fa
      zcain117 authored
      f65ac1fa
    • Sam Shleifer's avatar
      b9ecd92e
  2. 10 Aug, 2020 12 commits
  3. 09 Aug, 2020 2 commits
  4. 08 Aug, 2020 5 commits
  5. 07 Aug, 2020 11 commits