1. 18 Aug, 2023 2 commits
    • Arthur's avatar
      [`split_special_tokens`] Add support for `split_special_tokens` argument to encode (#25081) · 30b3c46f
      Arthur authored
      * draft changes
      
      * update and add tests
      
      * styling for no
      
      * move test
      
      * path to usable model
      
      * update test
      
      * small update
      
      * update bertbased tokenizers
      
      * don'tuse kwargs for _tokenize
      
      * don'tuse kwargs for _tokenize
      
      * fix copies
      
      * update
      
      * update test for special tokenizers
      
      * fixup
      
      * skip two tests
      
      * remove pdb breakpiont()
      
      * wowo
      
      * rewrite custom tests
      
      * nits
      
      * revert chang in target keys
      
      * fix markup lm
      
      * update documentation of the argument
      30b3c46f
    • Alex McKinney's avatar
      Replaces calls to `.cuda` with `.to(torch_device)` in tests (#25571) · 9d7afd25
      Alex McKinney authored
      
      
      * Replaces calls to `.cuda` with `.to(torch_device)` in tests
      `torch.Tensor.cuda()` is a pre-0.4 solution to changing a tensor's device. It is recommended to prefer `.to(...)` for greater flexibility and error handling. Furthermore, this makes it more consistent with other tests (that tend to use `.to(torch_device)`) and ensures the correct device backend is used (if `torch_device` is neither `cpu` or `cuda`).
      
      * addressing review comments
      
      * more formatting changes in Bloom test
      
      * `make style`
      
      * Update tests/models/bloom/test_modeling_bloom.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * fixes style failures
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      9d7afd25
  2. 17 Aug, 2023 5 commits
  3. 16 Aug, 2023 4 commits
  4. 14 Aug, 2023 1 commit
  5. 11 Aug, 2023 3 commits
  6. 09 Aug, 2023 1 commit
  7. 08 Aug, 2023 2 commits
  8. 07 Aug, 2023 2 commits
    • Pedro Lira's avatar
      Add mask2former fp16 support (#25093) · 080a9711
      Pedro Lira authored
      * Add mask2former fp16 support
      
      * Clear consistency/quality issues
      
      * Fix consistency/quality (2)
      
      * Add integration test for mask2former (fp16 case)
      
      * Fix code quality
      
      * Add integration test for maskformer (fp16 case)
      
      * Add integration test for oneformer (fp16 case)
      
      * Remove slow decorator from fp16 tests
      
      * Fix lint
      
      * Remove usage of full inference and value checks for fp16
      
      * Temporarily comment slow for {mask, mask2, one}former
      
      * Add fp16 support to oneformer
      
      * Revert "Temporarily comment slow for {mask, mask2, one}former"
      
      This reverts commit e5371edabd301cf56079def0421a0a87df307cb0.
      
      * Remove dtype conversion noop
      080a9711
    • Yih-Dar's avatar
      Fix more offload edge cases (#25342) · c177606f
      Yih-Dar authored
      
      
      * fix
      
      * fix
      
      * fix
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      c177606f
  9. 04 Aug, 2023 1 commit
  10. 03 Aug, 2023 2 commits
  11. 02 Aug, 2023 4 commits
  12. 01 Aug, 2023 1 commit
  13. 31 Jul, 2023 2 commits
  14. 28 Jul, 2023 4 commits
  15. 27 Jul, 2023 3 commits
  16. 26 Jul, 2023 2 commits
  17. 25 Jul, 2023 1 commit
    • Sebastian Husch Lee's avatar
      [`T5`, `MT5`, `UMT5`] Add [T5, MT5, UMT5]ForSequenceClassification (#24726) · 8f36ab3e
      Sebastian Husch Lee authored
      * Initial addition of t5forsequenceclassification
      
      * Adding imports and adding tests
      
      * Formatting
      
      * Running make fix-copies
      
      * Adding mt5forseq
      
      * Formatting
      
      * run make fix-copies
      
      * Adding to docs
      
      * Add model_parallel
      
      * Fix bug
      
      * Fix
      
      * Remove TODO
      
      * Fixing tests for T5ForSequenceClassification
      
      * Undo changes to dependency_versions_table.py
      
      * Change classification head to work with T5Config directly
      
      * Change seq length to let tests pass
      
      * PR comments for formatting
      
      * Formatting
      
      * Initial addition of UMT5ForSequenceClassification
      
      * Adding to inits and formatting
      
      * run make fix-copies
      
      * Add doc for UMT5ForSeqClass
      
      * Update UMT5 config
      
      * Fix docs
      
      * Skip torch fx test for SequenceClassification
      
      * Formatting
      
      * Add skip to UMT5 tests as well
      
      * Fix umt5 tests
      
      * Running make fix-copies
      
      * PR comments
      
      * Fix for change to sentence_representation
      
      * Rename seq_len to hidden_size since that's what it is
      
      * Use base_model to follow format of the rest of the library
      
      * Update docs
      
      * Extract the decoder_input_ids changes and make one liner
      
      * Make one-liner
      8f36ab3e