1. 26 Jun, 2024 1 commit
  2. 20 Jun, 2024 1 commit
    • Younes Belkada's avatar
      GGUF: Fix llama 3 GGUF (#31358) · 6d430616
      Younes Belkada authored
      * Create push-important-models.yml
      
      * llama3 support for GGUF
      
      * fixup
      
      * Update src/transformers/integrations/ggml.py
      
      * fix pre-tokenizer
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * address final comment
      
      * handle special tokens + add tests
      6d430616
  3. 17 Jun, 2024 2 commits
    • Ella Charlaix's avatar
      馃毃 Remove dataset with restrictive license (#31452) · 02300273
      Ella Charlaix authored
      remove dataset with restrictive license
      02300273
    • Albert Villanova del Moral's avatar
      Pass datasets trust_remote_code (#31406) · a14b055b
      Albert Villanova del Moral authored
      * Pass datasets trust_remote_code
      
      * Pass trust_remote_code in more tests
      
      * Add trust_remote_dataset_code arg to some tests
      
      * Revert "Temporarily pin datasets upper version to fix CI"
      
      This reverts commit b7672826.
      
      * Pass trust_remote_code in librispeech_asr_dummy docstrings
      
      * Revert "Pin datasets<2.20.0 for examples"
      
      This reverts commit 833fc17a.
      
      * Pass trust_remote_code to all examples
      
      * Revert "Add trust_remote_dataset_code arg to some tests" to research_projects
      
      * Pass trust_remote_code to tests
      
      * Pass trust_remote_code to docstrings
      
      * Fix flax examples tests requirements
      
      * Pass trust_remote_dataset_code arg to tests
      
      * Replace trust_remote_dataset_code with trust_remote_code in one example
      
      * Fix duplicate trust_remote_code
      
      * Replace args.trust_remote_dataset_code with args.trust_remote_code
      
      * Replace trust_remote_dataset_code with trust_remote_code in parser
      
      * Replace trust_remote_dataset_code with trust_remote_code in dataclasses
      
      * Replace trust_remote_dataset_code with trust_remote_code arg
      a14b055b
  4. 03 Jun, 2024 1 commit
    • Isotr0py's avatar
      Add Qwen2 GGUF loading support (#31175) · e4628434
      Isotr0py authored
      * add qwen2 gguf support
      
      * Update docs
      
      * fix qwen2 tokenizer
      
      * add qwen2 gguf test
      
      * fix typo in qwen2 gguf test
      
      * format code
      
      * Remove mistral, clarify the error message
      
      * format code
      
      * add typing and update docstring
      e4628434
  5. 31 May, 2024 1 commit
  6. 30 May, 2024 1 commit
  7. 27 May, 2024 1 commit
  8. 24 May, 2024 2 commits
  9. 23 May, 2024 1 commit
  10. 15 May, 2024 2 commits
  11. 13 May, 2024 1 commit
  12. 06 May, 2024 1 commit
  13. 02 May, 2024 1 commit
    • mobicham's avatar
      Add HQQ quantization support (#29637) · 59952994
      mobicham authored
      
      
      * update HQQ transformers integration
      
      * push import_utils.py
      
      * add force_hooks check in modeling_utils.py
      
      * fix | with Optional
      
      * force bias as param
      
      * check bias is Tensor
      
      * force forward for multi-gpu
      
      * review fixes pass
      
      * remove torch grad()
      
      * if any key in linear_tags fix
      
      * add cpu/disk check
      
      * isinstance return
      
      * add multigpu test + refactor tests
      
      * clean hqq_utils imports in hqq.py
      
      * clean hqq_utils imports in quantizer_hqq.py
      
      * delete hqq_utils.py
      
      * Delete src/transformers/utils/hqq_utils.py
      
      * ruff init
      
      * remove torch.float16 from __init__ in test
      
      * refactor test
      
      * isinstance -> type in quantizer_hqq.py
      
      * cpu/disk device_map check in quantizer_hqq.py
      
      * remove type(module) nn.linear check in quantizer_hqq.py
      
      * add BaseQuantizeConfig import inside HqqConfig init
      
      * remove hqq import in hqq.py
      
      * remove accelerate import from test_hqq.py
      
      * quant config.py doc update
      
      * add hqqconfig to main_classes doc
      
      * make style
      
      * __init__ fix
      
      * ruff __init__
      
      * skip_modules list
      
      * hqqconfig format fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * test_hqq.py remove mistral comment
      
      * remove self.using_multi_gpu is False
      
      * torch_dtype default val set and logger.info
      
      * hqq.py isinstance fix
      
      * remove torch=None
      
      * torch_device test_hqq
      
      * rename test_hqq
      
      * MODEL_ID in test_hqq
      
      * quantizer_hqq setattr fix
      
      * quantizer_hqq typo fix
      
      * imports quantizer_hqq.py
      
      * isinstance quantizer_hqq
      
      * hqq_layer.bias reformat quantizer_hqq
      
      * Step 2 as comment in quantizer_hqq
      
      * prepare_for_hqq_linear() comment
      
      * keep_in_fp32_modules fix
      
      * HqqHfQuantizer reformat
      
      * quantization.md hqqconfig
      
      * quantization.md model example reformat
      
      * quantization.md # space
      
      * quantization.md space   })
      
      * quantization.md space   })
      
      * quantization_config fix doc
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * axis value check in quantization_config
      
      * format
      
      * dynamic config explanation
      
      * quant config method in quantization.md
      
      * remove shard-level progress
      
      * .cuda fix modeling_utils
      
      * test_hqq fixes
      
      * make fix-copies
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      59952994
  14. 30 Apr, 2024 1 commit
  15. 22 Apr, 2024 1 commit
  16. 17 Apr, 2024 1 commit
  17. 09 Apr, 2024 1 commit
    • Marc Sun's avatar
      Fix quantization tests (#29914) · 58a939c6
      Marc Sun authored
      * revert back to torch 2.1.1
      
      * run test
      
      * switch to torch 2.2.1
      
      * udapte dockerfile
      
      * fix awq tests
      
      * fix test
      
      * run quanto tests
      
      * update tests
      
      * split quantization tests
      
      * fix
      
      * fix again
      
      * final fix
      
      * fix report artifact
      
      * build docker again
      
      * Revert "build docker again"
      
      This reverts commit 399a5f9d9308da071d79034f238c719de0f3532e.
      
      * debug
      
      * revert
      
      * style
      
      * new notification system
      
      * testing notfication
      
      * rebuild docker
      
      * fix_prev_ci_results
      
      * typo
      
      * remove warning
      
      * fix typo
      
      * fix artifact name
      
      * debug
      
      * issue fixed
      
      * debug again
      
      * fix
      
      * fix time
      
      * test notif with faling test
      
      * typo
      
      * issues again
      
      * final fix ?
      
      * run all quantization tests again
      
      * remove name to clear space
      
      * revert modfiication done on workflow
      
      * fix
      
      * build docker
      
      * build only quant docker
      
      * fix quantization ci
      
      * fix
      
      * fix report
      
      * better quantization_matrix
      
      * add print
      
      * revert to the basic one
      58a939c6
  18. 05 Apr, 2024 1 commit
  19. 15 Mar, 2024 1 commit
  20. 05 Mar, 2024 1 commit
  21. 27 Feb, 2024 1 commit
  22. 20 Feb, 2024 1 commit
  23. 19 Feb, 2024 2 commits
  24. 16 Feb, 2024 1 commit
  25. 14 Feb, 2024 1 commit
  26. 12 Feb, 2024 1 commit
  27. 30 Jan, 2024 2 commits
  28. 15 Jan, 2024 1 commit
  29. 12 Jan, 2024 2 commits
  30. 25 Dec, 2023 1 commit
  31. 21 Dec, 2023 1 commit
  32. 05 Dec, 2023 1 commit
  33. 13 Nov, 2023 1 commit
  34. 10 Nov, 2023 1 commit