1. 09 Jun, 2020 1 commit
    • Bharat Raghunathan's avatar
      [All models] Extend config.output_attentions with output_attentions function arguments (#4538) · 6e603cb7
      Bharat Raghunathan authored
      
      
      * DOC: Replace instances of ``config.output_attentions`` with function argument ``output_attentions``
      
      * DOC: Apply Black Formatting
      
      * Fix errors where output_attentions was undefined
      
      * Remove output_attentions in classes per review
      
      * Fix regressions on tests having `output_attention`
      
      * Fix further regressions in tests relating to `output_attentions`
      
      Ensure proper propagation of `output_attentions` as a function parameter
      to all model subclasses
      
      * Fix more regressions in `test_output_attentions`
      
      * Fix issues with BertEncoder
      
      * Rename related variables to `output_attentions`
      
      * fix pytorch tests
      
      * fix bert and gpt2 tf
      
      * Fix most TF tests for `test_output_attentions`
      
      * Fix linter errors and more TF tests
      
      * fix conflicts
      
      * DOC: Apply Black Formatting
      
      * Fix errors where output_attentions was undefined
      
      * Remove output_attentions in classes per review
      
      * Fix regressions on tests having `output_attention`
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * fix conflicts
      
      * fix pytorch tests
      
      * fix conflicts
      
      * fix conflicts
      
      * Fix linter errors and more TF tests
      
      * fix tf tests
      
      * make style
      
      * fix isort
      
      * improve output_attentions
      
      * improve tensorflow
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      6e603cb7
  2. 05 Jun, 2020 1 commit
  3. 29 May, 2020 2 commits
  4. 28 May, 2020 1 commit
  5. 27 May, 2020 1 commit
  6. 25 May, 2020 1 commit
    • Suraj Patil's avatar
      Longformer for question answering (#4500) · 03d8527d
      Suraj Patil authored
      * added LongformerForQuestionAnswering
      
      * add LongformerForQuestionAnswering
      
      * fix import for LongformerForMaskedLM
      
      * add LongformerForQuestionAnswering
      
      * hardcoded sep_token_id
      
      * compute attention_mask if not provided
      
      * combine global_attention_mask with attention_mask when provided
      
      * update example in  docstring
      
      * add assert error messages, better attention combine
      
      * add test for longformerForQuestionAnswering
      
      * typo
      
      * cast gloabl_attention_mask to long
      
      * make style
      
      * Update src/transformers/configuration_longformer.py
      
      * Update src/transformers/configuration_longformer.py
      
      * fix the code quality
      
      * Merge branch 'longformer-for-question-answering' of https://github.com/patil-suraj/transformers
      
       into longformer-for-question-answering
      Co-authored-by: default avatarPatrick von Platen <patrick.v.platen@gmail.com>
      03d8527d
  7. 19 May, 2020 2 commits
    • Patrick von Platen's avatar
      [Tests, GPU, SLOW] fix a bunch of GPU hardcoded tests in Pytorch (#4468) · aa925a52
      Patrick von Platen authored
      * fix gpu slow tests in pytorch
      
      * change model to device syntax
      aa925a52
    • Iz Beltagy's avatar
      Longformer (#4352) · 8f1d0471
      Iz Beltagy authored
      * first commit
      
      * bug fixes
      
      * better examples
      
      * undo padding
      
      * remove wrong VOCAB_FILES_NAMES
      
      * License
      
      * make style
      
      * make isort happy
      
      * unit tests
      
      * integration test
      
      * make `black` happy by undoing `isort` changes!!
      
      * lint
      
      * no need for the padding value
      
      * batch_size not bsz
      
      * remove unused type casting
      
      * seqlen not seq_len
      
      * staticmethod
      
      * `bert` selfattention instead of `n2`
      
      * uint8 instead of bool + lints
      
      * pad inputs_embeds using embeddings not a constant
      
      * black
      
      * unit test with padding
      
      * fix unit tests
      
      * remove redundant unit test
      
      * upload model weights
      
      * resolve todo
      
      * simpler _mask_invalid_locations without lru_cache + backward compatible masked_fill_
      
      * increase unittest coverage
      8f1d0471