"examples/research_projects/luke/README.md" did not exist on "0094eba36381da08bf4d199386e25db32f37753b"
  • Younes Belkada's avatar
    `bitsandbytes` - `Linear8bitLt` integration into `transformers` models (#17901) · 4a51075a
    Younes Belkada authored
    
    
    * first commit
    
    * correct replace function
    
    * add final changes
    
    - works like charm!
    - cannot implement tests yet
    - tested
    
    * clean up a bit
    
    * add bitsandbytes dependencies
    
    * working version
    
    - added import function
    - added bitsandbytes utils file
    
    * small fix
    
    * small fix
    
    - fix import issue
    
    * fix import issues
    
    * Apply suggestions from code review
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * refactor a bit
    
    - move bitsandbytes utils to utils
    - change comments on functions
    
    * reformat docstring
    
    - reformat docstring on init_empty_weights_8bit
    
    * Update src/transformers/__init__.py
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * revert bad formatting
    
    * change to bitsandbytes
    
    * refactor a bit
    
    - remove init8bit since it is useless
    
    * more refactoring
    
    - fixed init empty weights issue
    - added threshold param
    
    * small hack to make it work
    
    * Update src/transformers/modeling_utils.py
    
    * Update src/transformers/modeling_utils.py
    
    * revmoe the small hack
    
    * modify utils file
    
    * make style + refactor a bit
    
    * create correctly device map
    
    * add correct dtype for device map creation
    
    * Apply suggestions from code review
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * apply suggestions
    
    - remove with torch.grad
    - do not rely on Python bool magic!
    
    * add docstring
    
     - add docstring for new kwargs
    
    * add docstring
    
    - comment `replace_8bit_linear` function
    - fix weird formatting
    
    * - added more documentation
    - added new utility function for memory footprint tracking
    - colab demo to add
    
    * few modifs
    
    - typo doc
    - force cast into float16 when load_in_8bit is enabled
    
    * added colab link
    
    * add test architecture + docstring a bit
    
    * refactor a bit testing class
    
    * make style + refactor a bit
    
    * enhance checks
    
    - add more checks
    - start writing saving test
    
    * clean up a bit
    
    * male style
    
    * add more details on doc
    
    * add more tests
    
    - still needs to fix 2 tests
    
    * replace by "or"
    
    - could not fix it from GitHub GUI
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * refactor a bit testing code + add readme
    
    * make style
    
    * fix import issue
    
    * Update src/transformers/modeling_utils.py
    Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
    
    * add few comments
    
    * add more doctring + make style
    
    * more docstring
    
    * raise error when loaded in 8bit
    
    * make style
    
    * add warning if loaded on CPU
    
    * add small sanity check
    
    * fix small comment
    
    * add bitsandbytes on dockerfile
    
    * Improve documentation
    
    - improve documentation from comments
    
    * add few comments
    
    * slow tests pass on the VM but not on the CI VM
    
    * Fix merge conflict
    
    * make style
    
    * another test should pass on a multi gpu setup
    
    * fix bad import in testing file
    
    * Fix slow tests
    
    - remove dummy batches
    - no more CUDA illegal memory errors
    
    * odify dockerfile
    
    * Update docs/source/en/main_classes/model.mdx
    
    * Update Dockerfile
    
    * Update model.mdx
    
    * Update Dockerfile
    
    * Apply suggestions from code review
    
    * few modifications
    
    - lm head can stay on disk/cpu
    - change model name so that test pass
    
    * change test value
    
    - change test value to the correct output
    - torch bmm changed to baddmm in bloom modeling when merging
    
    * modify installation guidelines
    
    * Apply suggestions from code review
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * Apply suggestions from code review
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * Apply suggestions from code review
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * replace `n`by `name`
    
    * merge `load_in_8bit` and `low_cpu_mem_usage`
    
    * first try - keep the lm head in full precision
    
    * better check
    
    - check the attribute `base_model_prefix` instead of computing the number of parameters
    
    * added more tests
    
    * Update src/transformers/utils/bitsandbytes.py
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    
    * Merge branch 'integration-8bit' of https://github.com/younesbelkada/transformers
    
     into integration-8bit
    
    * improve documentation
    
    - fix typos for installation
    - change title in the documentation
    Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
    Co-authored-by: default avatarMichael Benayoun <mickbenayoun@gmail.com>
    4a51075a
Dockerfile 3.05 KB