1. 29 Jul, 2024 1 commit
  2. 23 Jul, 2024 1 commit
    • RhuiDih's avatar
      Enhancing SFT Training Efficiency Using Packing and FlashAttention2 with Position IDs (#31629) · 9cf4f2aa
      RhuiDih authored
      * add DataCollatorBatchFlattening
      
      * Update data_collator.py
      
      * change name
      
      * new FA2 flow if position_ids is provided
      
      * add comments
      
      * minor fix
      
      * minor fix data collator
      
      * add test cases for models
      
      * add test case for data collator
      
      * remove extra code
      
      * formating for ruff check and check_repo.py
      
      * ruff format
      
      ruff format tests src utils
      
      * custom_init_isort.py
      9cf4f2aa
  3. 22 Jul, 2024 1 commit
    • Marc Sun's avatar
      Add new quant method (#32047) · 96a074fa
      Marc Sun authored
      * Add new quant method
      
      * update
      
      * fix multi-device
      
      * add test
      
      * add offload
      
      * style
      
      * style
      
      * add simple example
      
      * initial doc
      
      * docstring
      
      * style again
      
      * works ?
      
      * better docs
      
      * switch to non persistant
      
      * remove print
      
      * fix init
      
      * code review
      96a074fa
  4. 17 Jul, 2024 1 commit
  5. 16 Jul, 2024 1 commit
  6. 05 Jul, 2024 2 commits
  7. 11 Jun, 2024 1 commit
    • amyeroberts's avatar
      Fast image processor (#28847) · f53fe35b
      amyeroberts authored
      
      
      * Draft fast image processors
      
      * Draft working fast version
      
      * py3.8 compatible cache
      
      * Enable loading fast image processors through auto
      
      * Tidy up; rescale behaviour based on input type
      
      * Enable tests for fast image processors
      
      * Smarter rescaling
      
      * Don't default to Fast
      
      * Safer imports
      
      * Add necessary Pillow requirement
      
      * Woops
      
      * Add AutoImageProcessor test
      
      * Fix up
      
      * Fix test for imagegpt
      
      * Fix test
      
      * Review comments
      
      * Add warning for TF and JAX input types
      
      * Rearrange
      
      * Return transforms
      
      * NumpyToTensor transformation
      
      * Rebase - include changes from upstream in ImageProcessingMixin
      
      * Safe typing
      
      * Fix up
      
      * convert mean/std to tesnor to rescale
      
      * Don't store transforms in state
      
      * Fix up
      
      * Update src/transformers/image_processing_utils_fast.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Warn if fast image processor available
      
      * Update src/transformers/models/vit/image_processing_vit_fast.py
      
      * Transpose incoming numpy images to be in CHW format
      
      * Update mapping names based on packages, auto set fast to None
      
      * Fix up
      
      * Fix
      
      * Add AutoImageProcessor.from_pretrained(checkpoint, use_fast=True) test
      
      * Update src/transformers/models/vit/image_processing_vit_fast.py
      Co-authored-by: default avatarPavel Iakubovskii <qubvel@gmail.com>
      
      * Add equivalence and speed tests
      
      * Fix up
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarPavel Iakubovskii <qubvel@gmail.com>
      f53fe35b
  8. 07 Jun, 2024 1 commit
    • Matt's avatar
      Remove ConversationalPipeline and Conversation object (#31165) · 065729a6
      Matt authored
      * Remove ConversationalPipeline and Conversation object, as they have been deprecated for some time and are due for removal
      
      * Update not-doctested.txt
      
      * Fix JA and ZH docs
      
      * Fix JA and ZH docs some more
      
      * Fix JA and ZH docs some more
      065729a6
  9. 14 May, 2024 1 commit
  10. 07 May, 2024 1 commit
  11. 02 May, 2024 1 commit
    • mobicham's avatar
      Add HQQ quantization support (#29637) · 59952994
      mobicham authored
      
      
      * update HQQ transformers integration
      
      * push import_utils.py
      
      * add force_hooks check in modeling_utils.py
      
      * fix | with Optional
      
      * force bias as param
      
      * check bias is Tensor
      
      * force forward for multi-gpu
      
      * review fixes pass
      
      * remove torch grad()
      
      * if any key in linear_tags fix
      
      * add cpu/disk check
      
      * isinstance return
      
      * add multigpu test + refactor tests
      
      * clean hqq_utils imports in hqq.py
      
      * clean hqq_utils imports in quantizer_hqq.py
      
      * delete hqq_utils.py
      
      * Delete src/transformers/utils/hqq_utils.py
      
      * ruff init
      
      * remove torch.float16 from __init__ in test
      
      * refactor test
      
      * isinstance -> type in quantizer_hqq.py
      
      * cpu/disk device_map check in quantizer_hqq.py
      
      * remove type(module) nn.linear check in quantizer_hqq.py
      
      * add BaseQuantizeConfig import inside HqqConfig init
      
      * remove hqq import in hqq.py
      
      * remove accelerate import from test_hqq.py
      
      * quant config.py doc update
      
      * add hqqconfig to main_classes doc
      
      * make style
      
      * __init__ fix
      
      * ruff __init__
      
      * skip_modules list
      
      * hqqconfig format fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * hqqconfig doc fix
      
      * test_hqq.py remove mistral comment
      
      * remove self.using_multi_gpu is False
      
      * torch_dtype default val set and logger.info
      
      * hqq.py isinstance fix
      
      * remove torch=None
      
      * torch_device test_hqq
      
      * rename test_hqq
      
      * MODEL_ID in test_hqq
      
      * quantizer_hqq setattr fix
      
      * quantizer_hqq typo fix
      
      * imports quantizer_hqq.py
      
      * isinstance quantizer_hqq
      
      * hqq_layer.bias reformat quantizer_hqq
      
      * Step 2 as comment in quantizer_hqq
      
      * prepare_for_hqq_linear() comment
      
      * keep_in_fp32_modules fix
      
      * HqqHfQuantizer reformat
      
      * quantization.md hqqconfig
      
      * quantization.md model example reformat
      
      * quantization.md # space
      
      * quantization.md space   })
      
      * quantization.md space   })
      
      * quantization_config fix doc
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * axis value check in quantization_config
      
      * format
      
      * dynamic config explanation
      
      * quant config method in quantization.md
      
      * remove shard-level progress
      
      * .cuda fix modeling_utils
      
      * test_hqq fixes
      
      * make fix-copies
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      59952994
  12. 25 Apr, 2024 1 commit
  13. 22 Apr, 2024 1 commit
  14. 02 Apr, 2024 1 commit
  15. 15 Mar, 2024 1 commit
  16. 12 Mar, 2024 1 commit
  17. 06 Mar, 2024 1 commit
  18. 05 Mar, 2024 1 commit
  19. 16 Feb, 2024 1 commit
  20. 14 Feb, 2024 2 commits
  21. 05 Feb, 2024 1 commit
  22. 02 Feb, 2024 1 commit
  23. 01 Feb, 2024 1 commit
  24. 25 Jan, 2024 1 commit
  25. 24 Jan, 2024 1 commit
    • Steven Liu's avatar
      [docs] DeepSpeed (#28542) · 738ec75c
      Steven Liu authored
      * config
      
      * optim
      
      * pre deploy
      
      * deploy
      
      * save weights, memory, troubleshoot, non-Trainer
      
      * done
      738ec75c
  26. 12 Jan, 2024 1 commit
  27. 02 Jan, 2024 1 commit
  28. 20 Dec, 2023 1 commit
  29. 18 Dec, 2023 1 commit
  30. 15 Dec, 2023 2 commits
  31. 11 Dec, 2023 1 commit
  32. 28 Nov, 2023 1 commit
  33. 27 Nov, 2023 1 commit
  34. 24 Nov, 2023 2 commits
  35. 20 Nov, 2023 1 commit
  36. 13 Nov, 2023 1 commit