1. 15 Feb, 2024 4 commits
    • amyeroberts's avatar
      Fix copies between DETR and DETA (#29037) · 8a0ed0a9
      amyeroberts authored
      8a0ed0a9
    • Donggeun Yu's avatar
      DeformableDetrModel support fp16 (#29013) · 5b6fa230
      Donggeun Yu authored
      
      
      * Update ms_deform_attn_cuda.cu
      
      * Update ms_deform_attn_cuda.cuh
      
      * Update modeling_deformable_detr.py
      
      * Update src/transformers/models/deformable_detr/modeling_deformable_detr.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update modeling_deformable_detr.py
      
      * python utils/check_copies.py --fix_and_overwrite
      
      * Fix dtype missmatch error
      
      * Update test_modeling_deformable_detr.py
      
      * Update test_modeling_deformable_detr.py
      
      * Update modeling_deformable_detr.py
      
      * Update modeling_deformable_detr.py
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      5b6fa230
    • Sangbum Daniel Choi's avatar
      Add cuda_custom_kernel in DETA (#28989) · 83e96dc0
      Sangbum Daniel Choi authored
      * enable graident checkpointing in DetaObjectDetection
      
      * fix missing part in original DETA
      
      * make style
      
      * make fix-copies
      
      * Revert "make fix-copies"
      
      This reverts commit 4041c86c29248f1673e8173b677c20b5a4511358.
      
      * remove fix-copies of DetaDecoder
      
      * enable swin gradient checkpointing
      
      * fix gradient checkpointing in donut_swin
      
      * add tests for deta/swin/donut
      
      * Revert "fix gradient checkpointing in donut_swin"
      
      This reverts commit 1cf345e34d3cc0e09eb800d9895805b1dd9b474d.
      
      * change supports_gradient_checkpointing pipeline to PreTrainedModel
      
      * Revert "add tests for deta/swin/donut"
      
      This reverts commit 6056ffbb1eddc3cb3a99e4ebb231ae3edf295f5b.
      
      * Revert "Revert "fix gradient checkpointing in donut_swin""
      
      This reverts commit 24e25d0a14891241de58a0d86f817d0b5d2a341f.
      
      * Simple revert
      
      * enable deformable detr gradient checkpointing
      
      * add gradient in encoder
      
      * add cuda_custom_kernel function in MSDA
      
      * make style and fix input of DetaMSDA
      
      * make fix-copies
      
      * remove n_levels in input of DetaMSDA
      
      * minor changes
      
      * refactor custom_cuda_kernel like yoso format
      https://github.com/huggingface/transformers/blob/0507e69d34f8902422eb4977ec066dd6bef179a0/src/transformers/models/yoso/modeling_yoso.py#L53
      83e96dc0
    • Arthur's avatar
      Fix static generation when compiling! (#28937) · f3788b09
      Arthur authored
      
      
      * wow I was scared!
      
      * fix everything
      
      * nits
      
      * make it BC?
      
      * add todo
      
      * nits
      
      * is_tracing should still be used to pass tracing tests
      
      * nits
      
      * some nits to make sure genration works with static cache uncompiled
      
      * fix sdpa
      
      * fix FA2 for both static and dynamic in a better way?
      
      * style
      
      * fix-copies
      
      * fix fix copies
      
      * fix sequential beam searcg
      
      * style
      
      * use `keys_to_ignore`
      
      * nit
      
      * correct dtype inference when init
      
      * :( the fix for FA2 is still not optimal to investigate!
      
      * styling
      
      * nits
      
      * nit
      
      * this might work better
      
      * add comment
      
      * Update src/transformers/models/llama/modeling_llama.py
      
      * "position_ids" -> "cache_position"
      
      * style
      
      * nit
      
      * Remove changes that should no be propagatted just yet
      
      * Apply suggestions from code review
      
      * Styling
      
      * make sure we raise an errir for static cache with FA2 enabled
      
      * move  to the bottom of the signature
      
      * style
      
      * Update src/transformers/models/llama/modeling_llama.py
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      
      * Update src/transformers/models/llama/modeling_llama.py
      
      * nit in the name
      
      ---------
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      f3788b09
  2. 14 Feb, 2024 15 commits
  3. 13 Feb, 2024 5 commits
  4. 12 Feb, 2024 11 commits
  5. 09 Feb, 2024 4 commits
  6. 08 Feb, 2024 1 commit