1. 07 Dec, 2020 1 commit
  2. 05 Dec, 2020 1 commit
    • Ethan Perez's avatar
      Don't pass in token_type_ids to BART for GLUE (#8929) · 8dfc8c72
      Ethan Perez authored
      Without this fix, training a `BARTForSequenceClassification` model with `run_pl_glue.py` gives `TypeError: forward() got an unexpected keyword argument 'token_type_ids'`, because BART does not have token_type_ids. I've solved this issue in the same way as it's solved for the "distilbert" model, and I can train BART models on SNLI without errors now.
      8dfc8c72
  3. 04 Dec, 2020 2 commits
  4. 01 Dec, 2020 1 commit
  5. 30 Nov, 2020 3 commits
  6. 26 Nov, 2020 4 commits
  7. 24 Nov, 2020 3 commits
  8. 23 Nov, 2020 2 commits
  9. 22 Nov, 2020 1 commit
  10. 20 Nov, 2020 1 commit
    • Quentin Lhoest's avatar
      Fix rag finetuning + add finetuning test (#8585) · 8062fa63
      Quentin Lhoest authored
      * replace init_ddp_connection for index init
      
      * style
      
      * add finetune test
      
      * add test data
      
      * move generate tensors to device
      
      * add test on EM metric
      
      * style
      
      * allow multi process test
      
      * keep gloo process group for retrieval
      
      * add multi-gpu test
      
      * use custom accelerator
      
      * clean test finetune
      
      * minor
      
      * style
      
      * style
      
      * typo
      
      * use python call instead of imported main fumction
      
      * return_dict fix in modeling_rag
      
      * use float32 in retrieval
      
      * store as float32 as well in the custom knowledge dataset example
      
      * style
      
      * rename to finetune_rag
      
      * style
      
      * update readme
      
      * rename utils and callbacks to utils_rag and callbacks_rag
      
      * fix test
      
      * patrick's comments
      
      * generate dummy data in the finetue test script
      
      * remove dummy data files
      
      * style
      8062fa63
  11. 19 Nov, 2020 6 commits
  12. 18 Nov, 2020 5 commits
  13. 17 Nov, 2020 4 commits
  14. 16 Nov, 2020 1 commit
    • Sylvain Gugger's avatar
      Switch `return_dict` to `True` by default. (#8530) · 1073a2bd
      Sylvain Gugger authored
      * Use the CI to identify failing tests
      
      * Remove from all examples and tests
      
      * More default switch
      
      * Fixes
      
      * More test fixes
      
      * More fixes
      
      * Last fixes hopefully
      
      * Use the CI to identify failing tests
      
      * Remove from all examples and tests
      
      * More default switch
      
      * Fixes
      
      * More test fixes
      
      * More fixes
      
      * Last fixes hopefully
      
      * Run on the real suite
      
      * Fix slow tests
      1073a2bd
  15. 15 Nov, 2020 1 commit
    • Thomas Wolf's avatar
      [breaking|pipelines|tokenizers] Adding slow-fast tokenizers equivalence tests... · f4e04cd2
      Thomas Wolf authored
      
      [breaking|pipelines|tokenizers] Adding slow-fast tokenizers equivalence tests pipelines - Removing sentencepiece as a required dependency (#8073)
      
      * Fixing roberta for slow-fast tests
      
      * WIP getting equivalence on pipelines
      
      * slow-to-fast equivalence - working on question-answering pipeline
      
      * optional FAISS tests
      
      * Pipeline Q&A
      
      * Move pipeline tests to their own test job again
      
      * update tokenizer to add sequence id methods
      
      * update to tokenizers 0.9.4
      
      * set sentencepiecce as optional
      
      * clean up squad
      
      * clean up pipelines to use sequence_ids
      
      * style/quality
      
      * wording
      
      * Switch to use_fast = True by default
      
      * update tests for use_fast at True by default
      
      * fix rag tokenizer test
      
      * removing protobuf from required dependencies
      
      * fix NER test for use_fast = True by default
      
      * fixing example tests (Q&A examples use slow tokenizers for now)
      
      * protobuf in main deps extras["sentencepiece"] and example deps
      
      * fix protobug install test
      
      * try to fix seq2seq by switching to slow tokenizers for now
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      Co-authored-by: default avatarLysandre Debut <lysandre@huggingface.co>
      f4e04cd2
  16. 12 Nov, 2020 2 commits
  17. 11 Nov, 2020 2 commits