1. 24 Aug, 2023 1 commit
  2. 18 Aug, 2023 1 commit
  3. 17 Aug, 2023 1 commit
  4. 10 Aug, 2023 1 commit
  5. 07 Aug, 2023 1 commit
  6. 31 Jul, 2023 1 commit
  7. 13 Jul, 2023 1 commit
  8. 11 Jul, 2023 1 commit
  9. 17 May, 2023 1 commit
  10. 12 May, 2023 2 commits
  11. 11 May, 2023 1 commit
  12. 10 May, 2023 1 commit
  13. 27 Apr, 2023 1 commit
  14. 24 Mar, 2023 1 commit
  15. 17 Mar, 2023 2 commits
  16. 13 Mar, 2023 1 commit
  17. 02 Mar, 2023 1 commit
    • amyeroberts's avatar
      Use PyAV instead of Decord in examples (#21572) · 3412f597
      amyeroberts authored
      * Use PyAV instead of Decord
      
      * Get frame indices
      
      * Fix number of frames
      
      * Update src/transformers/models/videomae/image_processing_videomae.py
      
      * Fix up
      
      * Fix copies
      
      * Update timesformer doctests
      
      * Update docstrings
      3412f597
  18. 02 Feb, 2023 1 commit
  19. 12 Dec, 2022 1 commit
  20. 07 Dec, 2022 1 commit
  21. 06 Dec, 2022 1 commit
  22. 05 Dec, 2022 1 commit
  23. 30 Nov, 2022 1 commit
  24. 18 Nov, 2022 1 commit
  25. 14 Nov, 2022 1 commit
  26. 26 Oct, 2022 1 commit
    • raghavanone's avatar
      Change the import of kenlm from github to pypi (#19770) · 7829c890
      raghavanone authored
      * Change the import of kenlm from github to pypi
      
      * Change the import of kenlm from github to pypi in circleci config
      
      * Fix code quality issues
      
      * Fix isort issue, add kenlm in extras for audio
      
      * Add kenlm to deps
      
      * Add kenlm to deps
      
      * Commit 'make fixup' changes
      
      * Remove version from kenlm deps
      
      * commit make fixup changes
      
      * Remove manual installation of kenlm
      
      * Remove manual installation of kenlm
      
      * Remove manual installation of kenlm
      7829c890
  27. 20 Oct, 2022 1 commit
  28. 11 Oct, 2022 1 commit
  29. 12 Sep, 2022 1 commit
  30. 10 Sep, 2022 2 commits
  31. 07 Sep, 2022 1 commit
  32. 23 Aug, 2022 1 commit
  33. 19 Aug, 2022 1 commit
  34. 16 Aug, 2022 2 commits
  35. 10 Aug, 2022 1 commit
    • Younes Belkada's avatar
      `bitsandbytes` - `Linear8bitLt` integration into `transformers` models (#17901) · 4a51075a
      Younes Belkada authored
      
      
      * first commit
      
      * correct replace function
      
      * add final changes
      
      - works like charm!
      - cannot implement tests yet
      - tested
      
      * clean up a bit
      
      * add bitsandbytes dependencies
      
      * working version
      
      - added import function
      - added bitsandbytes utils file
      
      * small fix
      
      * small fix
      
      - fix import issue
      
      * fix import issues
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * refactor a bit
      
      - move bitsandbytes utils to utils
      - change comments on functions
      
      * reformat docstring
      
      - reformat docstring on init_empty_weights_8bit
      
      * Update src/transformers/__init__.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * revert bad formatting
      
      * change to bitsandbytes
      
      * refactor a bit
      
      - remove init8bit since it is useless
      
      * more refactoring
      
      - fixed init empty weights issue
      - added threshold param
      
      * small hack to make it work
      
      * Update src/transformers/modeling_utils.py
      
      * Update src/transformers/modeling_utils.py
      
      * revmoe the small hack
      
      * modify utils file
      
      * make style + refactor a bit
      
      * create correctly device map
      
      * add correct dtype for device map creation
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * apply suggestions
      
      - remove with torch.grad
      - do not rely on Python bool magic!
      
      * add docstring
      
       - add docstring for new kwargs
      
      * add docstring
      
      - comment `replace_8bit_linear` function
      - fix weird formatting
      
      * - added more documentation
      - added new utility function for memory footprint tracking
      - colab demo to add
      
      * few modifs
      
      - typo doc
      - force cast into float16 when load_in_8bit is enabled
      
      * added colab link
      
      * add test architecture + docstring a bit
      
      * refactor a bit testing class
      
      * make style + refactor a bit
      
      * enhance checks
      
      - add more checks
      - start writing saving test
      
      * clean up a bit
      
      * male style
      
      * add more details on doc
      
      * add more tests
      
      - still needs to fix 2 tests
      
      * replace by "or"
      
      - could not fix it from GitHub GUI
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * refactor a bit testing code + add readme
      
      * make style
      
      * fix import issue
      
      * Update src/transformers/modeling_utils.py
      Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
      
      * add few comments
      
      * add more doctring + make style
      
      * more docstring
      
      * raise error when loaded in 8bit
      
      * make style
      
      * add warning if loaded on CPU
      
      * add small sanity check
      
      * fix small comment
      
      * add bitsandbytes on dockerfile
      
      * Improve documentation
      
      - improve documentation from comments
      
      * add few comments
      
      * slow tests pass on the VM but not on the CI VM
      
      * Fix merge conflict
      
      * make style
      
      * another test should pass on a multi gpu setup
      
      * fix bad import in testing file
      
      * Fix slow tests
      
      - remove dummy batches
      - no more CUDA illegal memory errors
      
      * odify dockerfile
      
      * Update docs/source/en/main_classes/model.mdx
      
      * Update Dockerfile
      
      * Update model.mdx
      
      * Update Dockerfile
      
      * Apply suggestions from code review
      
      * few modifications
      
      - lm head can stay on disk/cpu
      - change model name so that test pass
      
      * change test value
      
      - change test value to the correct output
      - torch bmm changed to baddmm in bloom modeling when merging
      
      * modify installation guidelines
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * replace `n`by `name`
      
      * merge `load_in_8bit` and `low_cpu_mem_usage`
      
      * first try - keep the lm head in full precision
      
      * better check
      
      - check the attribute `base_model_prefix` instead of computing the number of parameters
      
      * added more tests
      
      * Update src/transformers/utils/bitsandbytes.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Merge branch 'integration-8bit' of https://github.com/younesbelkada/transformers
      
       into integration-8bit
      
      * improve documentation
      
      - fix typos for installation
      - change title in the documentation
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
      4a51075a
  36. 08 Aug, 2022 1 commit