1. 17 Jul, 2024 1 commit
  2. 11 Jul, 2024 2 commits
    • Naman Garg's avatar
      Adding hiera (#30356) · c1e139c2
      Naman Garg authored
      
      
      * initialized Structure
      
      * Updated variable names
      
      * Added Config class, basic HF setup, convert_to_hf
      
      * Fixed Convert function, added hiera to HF files, Initilized test files
      
      * better naming for x in forward pass
      
      * Moved utils to hiera
      
      * Change hiera -> hiera_model
      
      * Fixed integration into tranformers
      
      * Fix: Convert Checkpoint
      
      * added documentation for hiera
      
      * added documentation for hiera
      
      * added Docstings to models, Transformers based changes
      
      * make style and quality
      
      * make style and quality
      
      * Integration & Block tests running
      
      * Fixed bugs
      
      * initialized Structure
      
      * Updated variable names
      
      * Added Config class, basic HF setup, convert_to_hf
      
      * Fixed Convert function, added hiera to HF files, Initilized test files
      
      * better naming for x in forward pass
      
      * Moved utils to hiera
      
      * Change hiera -> hiera_model
      
      * Fixed integration into tranformers
      
      * Fix: Convert Checkpoint
      
      * added documentation for hiera
      
      * added documentation for hiera
      
      * added Docstings to models, Transformers based changes
      
      * make style and quality
      
      * make style and quality
      
      * Integration & Block tests running
      
      * Fixed bugs
      
      * Removed tim dependency
      
      * added HieraBlock
      
      * fixed: Model name
      
      * added tests for HieraModel, HieraBlock
      
      * fixed imports
      
      * fixed quality & copies
      
      * Fixes
      
      * Update docs/source/en/model_doc/hiera.md
      
      Fix name
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/hiera.md
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update docs/source/en/model_doc/hiera.md
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/configuration_hiera.py
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/configuration_hiera.py
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/modeling_hiera.py
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/modeling_hiera.py
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      
      * Fixed formatting
      
      * Code quality & Import differences
      
      * quality and repo-consistency fix
      
      * fixed no torch error
      
      * Docstring fix
      
      * Docstring fix
      
      * doc string fix
      
      * fixed example usage
      
      * Resolved issues in modeling_hiera
      
      * Removed Hiera MAE
      
      * Added test and resolved bug
      
      * fixed doc string
      
      * First commit
      
      * Finished conversion script and model forward working
      
      * Resolved all issues
      
      * nits
      
      * Improving tests
      
      * Nits
      
      * More nits
      
      * Improving HieraForMaskedImageModeling
      
      * More improvements and nits
      
      * Fixed docstrings of outputs
      
      * More fixes
      
      * More imrpovments
      
      * Updated conversion script
      
      * Fixed docstrings
      
      * Improved tests
      
      * Fixed attentou outputs test
      
      * All tests green
      
      * Removed unnecessary file
      
      * contribution attribution
      
      * Resolved a few issues
      
      * Resolved Comments
      
      * Updated model repo id and fixed bugs
      
      * Removed loss print
      
      * Make tests green
      
      * Updated docstrings
      
      * Fix style
      
      * Fixed num_heads in config
      
      * Removed unnecessary video checkpoint related code in the conversion script
      
      * Fix style
      
      * Changed atol in conversion script
      
      * HieraConfig
      
      * Fix copies
      
      * Fixed typo
      
      * Resolved few issues
      
      * make
      
      * converted conv_nd -> nn.Module
      
      * Removed video complexities
      
      * Removed video complexities
      
      * fix style
      
      * Addressing comments
      
      * Update src/transformers/models/hiera/modeling_hiera.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/modeling_hiera.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/models/hiera/modeling_hiera.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Fix style
      
      * Fixed tests
      
      * Fixed typo
      
      * Fixed interpolate test
      
      * Made torch fx compatible
      
      * Made sure imageprocesor is correct
      
      * Addressed comments
      
      * Noise directly as torch
      
      * Remove unnecesary attr
      
      * Added return_dit
      
      * Update src/transformers/models/hiera/__init__.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Updated checkpoints
      
      * [run_slow] hiera
      
      * Fixed device mismatch
      
      * [run_slow] hiera
      
      * Fixed GPU tests
      
      * [run_slow] hiera
      
      ---------
      Co-authored-by: default avatarUbuntu <ubuntu@ip-172-31-29-50.us-east-2.compute.internal>
      Co-authored-by: default avatarSteven Liu <59462357+stevhliu@users.noreply.github.com>
      Co-authored-by: default avatarEduardo Pacheco <eduardo.pach@hotmail.com>
      Co-authored-by: default avatarEduardo Pacheco <69953243+EduardoPach@users.noreply.github.com>
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      c1e139c2
    • Arthur's avatar
      Refactor flash attention implementation in transformers (#31446) · e3143952
      Arthur authored
      
      
      * dumb commit
      
      * nit
      
      * update
      
      * something like this
      
      * unpack in modeling utils
      
      * safe import
      
      * oups
      
      * update
      
      * nits
      
      * diff convert gemma
      
      * update
      
      * start propagating
      
      * udpate other modeling code as well
      
      * update for sliding window models
      
      * nits
      
      * more init cleanups
      
      * styling
      
      * fixup
      
      * noice
      
      * pass fixup
      
      * typo typing_extension -> typing_extensions
      
      * torch.nn.functionnal -> torch.nn.functional
      
      * add to import structure
      
      * unpack
      
      * simplify a bit more for this first version
      
      * nut
      
      * update
      
      * update
      
      * nit
      
      * ease the import of `Unpack`
      
      * remove useless `use_sliding_window`
      
      * no qua please
      
      * protect import?
      
      * style
      
      * [run-slow]
      
      * [run slow] llama,gemma,mistral,mixtral
      
      * remove extra kwargs
      
      * fix llama
      
      * address review comments
      
      * apply diff_model_converter to modeling_gemma.py
      
      * remove cache_position 1
      
      * remove cache_position 2
      
      * some cleaning
      
      * refactor gemma2 as well
      
      * apply review comments
      
      * rename file to modeling_flash_attention_utils.py
      
      * siglip refactor
      
      * remove dead code
      
      * is the hub down?
      
      * still down?
      
      * fix siglip
      
      * fix gemma2
      
      * fatal: Could not read from remote repository.
      
      * fix typo in softcap implem
      
      * flacky
      
      * Failed: Timeout >120.0s
      
      ---------
      Co-authored-by: default avatarfxmarty <9808326+fxmarty@users.noreply.github.com>
      e3143952
  3. 08 Jul, 2024 1 commit
    • omahs's avatar
      Fix typos (#31819) · e5ca9b05
      omahs authored
      * fix typo
      
      * fix typo
      
      * fix typos
      
      * fix typo
      
      * fix typos
      e5ca9b05
  4. 27 Jun, 2024 1 commit
    • Arthur's avatar
      Add gemma 2 (#31659) · 0cf60f13
      Arthur authored
      
      
      * inital commit
      
      * Add doc
      
      * protect?
      
      * fixup stuffs
      
      * update tests
      
      * fix build documentation
      
      * mmmmmmm config attributes
      
      * style
      
      * nit
      
      * uodate
      
      * nit
      
      * Fix docs
      
      * protect some stuff
      
      ---------
      Co-authored-by: default avatarLysandre <lysandre@huggingface.co>
      0cf60f13
  5. 26 Jun, 2024 1 commit
  6. 25 Jun, 2024 2 commits
  7. 20 Jun, 2024 1 commit
  8. 17 Jun, 2024 1 commit
    • Albert Villanova del Moral's avatar
      Pass datasets trust_remote_code (#31406) · a14b055b
      Albert Villanova del Moral authored
      * Pass datasets trust_remote_code
      
      * Pass trust_remote_code in more tests
      
      * Add trust_remote_dataset_code arg to some tests
      
      * Revert "Temporarily pin datasets upper version to fix CI"
      
      This reverts commit b7672826.
      
      * Pass trust_remote_code in librispeech_asr_dummy docstrings
      
      * Revert "Pin datasets<2.20.0 for examples"
      
      This reverts commit 833fc17a.
      
      * Pass trust_remote_code to all examples
      
      * Revert "Add trust_remote_dataset_code arg to some tests" to research_projects
      
      * Pass trust_remote_code to tests
      
      * Pass trust_remote_code to docstrings
      
      * Fix flax examples tests requirements
      
      * Pass trust_remote_dataset_code arg to tests
      
      * Replace trust_remote_dataset_code with trust_remote_code in one example
      
      * Fix duplicate trust_remote_code
      
      * Replace args.trust_remote_dataset_code with args.trust_remote_code
      
      * Replace trust_remote_dataset_code with trust_remote_code in parser
      
      * Replace trust_remote_dataset_code with trust_remote_code in dataclasses
      
      * Replace trust_remote_dataset_code with trust_remote_code arg
      a14b055b
  9. 10 Jun, 2024 1 commit
  10. 07 Jun, 2024 1 commit
    • Matt's avatar
      Remove ConversationalPipeline and Conversation object (#31165) · 065729a6
      Matt authored
      * Remove ConversationalPipeline and Conversation object, as they have been deprecated for some time and are due for removal
      
      * Update not-doctested.txt
      
      * Fix JA and ZH docs
      
      * Fix JA and ZH docs some more
      
      * Fix JA and ZH docs some more
      065729a6
  11. 04 Jun, 2024 1 commit
  12. 03 Jun, 2024 2 commits
  13. 31 May, 2024 2 commits
    • Arthur's avatar
      Diff converter v2 (#30868) · 96eb0628
      Arthur authored
      * current working example!
      
      * commit regex and result file
      
      * update
      
      * nit
      
      * push the conversion file
      
      * oups
      
      * roadmap and nits
      
      * attempt diffs for 3 files
      
      * persimmon
      
      * nit
      
      * add diff file that is the same as the modeling_llama.py
      
      * fix rope nits
      
      * updates
      
      * updates with converted versions
      
      * give some breathing space to the code
      
      * delete
      
      * update
      
      * update
      
      * push the actual result
      
      * update regex patterns
      
      * update regex patterns
      
      * fix some issues
      
      * fix some issues
      
      * fix some issues
      
      * updates
      
      * updates
      
      * updates
      
      * updates
      
      * updates
      
      * revert changes done to llama
      
      * updates
      
      * update gemma
      
      * updates
      
      * oups
      
      * current state
      
      * current state
      
      * update
      
      * ouiiii
      
      * nit
      
      * clear diffs
      
      * nit
      
      * fixup
      
      * update
      
      * doc 馃殌
      
      * 馃敟
      
      * for now use gemma
      
      * deal with comments
      
      * style
      
      * handle funtions
      
      * deal with assigns
      
      * todos
      
      * process inheritage
      
      * keep decorators?
      
      * 馃
      
      * deal with duplicates
      
      * fixup
      
      * correctly remove duplicate code
      
      * run ruff post script
      
      * ruff deals pretty well with imports, let's leave it to him
      
      * ah maybe not lol
      
      * for now remove all imports from child.
      
      * nit
      
      * conversion of llama
      
      * okay
      
      * convert starcoder2
      
      * synch with main
      
      * update llama diff
      
      * updates
      
      * https://docs.astral.sh/ruff/rules/redefined-while-unused/
      
       fixes the imports, bit needs later version of ruff
      
      * updates
      
      * okay actual state
      
      * non zero exit
      
      * update!
      
      * revert unrelated
      
      * remove other diff files
      
      * updates
      
      * cleanup
      
      * update
      
      * less diff!
      
      * stash
      
      * current updates
      
      * updates
      
      * No need for call
      
      * finished fining deps
      
      * update
      
      * current changes
      
      * current state
      
      * current state
      
      * new status
      
      * nit
      
      * finally
      
      * fixes
      
      * nits
      
      * order is now expected
      
      * use logger info instead of prints
      
      * fixup
      
      * up
      
      * nit
      
      * update
      
      * nits
      
      * update
      
      * correct merge
      
      * update
      
      * update
      
      * update
      
      * add warning
      
      * update caution message
      
      * update
      
      * better merging strategy
      
      * copy class statements :wink
      
      * fixups
      
      * nits
      
      * update
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * nits
      
      * smaller header
      
      * do cleanup some stuff
      
      * even simpler header?
      
      * fixup
      
      * updates
      
      * ruff
      
      * update examples
      
      * nit
      
      * TODO
      
      * state
      
      * OUUUUUUF
      
      * current state
      
      * nits
      
      * final state
      
      * add a readme
      
      * fixup
      
      * remove diff llama
      
      * fix
      
      * nit
      
      * dummy noy funny
      
      * ruff format tests src utils --check
      
      * everless diffs
      
      * less diffs and fix test
      
      * fixes
      
      * naming nit?
      
      * update converter and add supper example
      
      * nits
      
      * updated for function signatures
      
      * update
      
      * update
      
      * add converted dummies
      
      * autoformat
      
      * single target assign fix
      
      * fixup
      
      * fix some imports
      
      * fixes
      
      * don't push them
      
      * `# noqa: F841`
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      96eb0628
    • Arthur's avatar
      helper (#31152) · f3f640dc
      Arthur authored
      
      
      * helper
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * updates
      
      * more doc
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      f3f640dc
  14. 28 May, 2024 1 commit
    • amyeroberts's avatar
      Deprecate low use models (#30781) · a564d10a
      amyeroberts authored
      * Deprecate models
      - graphormer
      - time_series_transformer
      - xlm_prophetnet
      - qdqbert
      - nat
      - ernie_m
      - tvlt
      - nezha
      - mega
      - jukebox
      - vit_hybrid
      - x_clip
      - deta
      - speech_to_text_2
      - efficientformer
      - realm
      - gptsan_japanese
      
      * Fix up
      
      * Fix speech2text2 imports
      
      * Make sure message isn't indented
      
      * Fix docstrings
      
      * Correctly map for deprecated models from model_type
      
      * Uncomment out
      
      * Add back time series transformer and x-clip
      
      * Import fix and fix-up
      
      * Fix up with updated ruff
      a564d10a
  15. 24 May, 2024 1 commit
  16. 22 May, 2024 2 commits
  17. 17 May, 2024 1 commit
  18. 16 May, 2024 1 commit
    • hyenal's avatar
      add sdpa to ViT [follow up of #29325] (#30555) · 1c21f48a
      hyenal authored
      
      
      remove blank line (+1 squashed commit)
      Squashed commits:
      [24ccd2061] [run-slow]vit_msn,vision_encoder_decoder (+24 squashed commits)
      Squashed commits:
      [08bd27e7a] [run-slow]vit_msn,vision_encoder_decoder
      [ec96a8db3] [run-slow]vit_msn
      [ead817eca] fix vit msn multi gpu
      [d12cdc8fd] [run-slow]audio_spectrogram_transformer,deit,vision_encoder_decoder,vision_text_dual_encoder,vit,vit_hybrid,vit_mae,vit_msn,videomae,yolos
      [3fdbfa88f] doc
      [a3ff33e4a] finish implementation
      [e20b7b7fb] Update test_modeling_common.py
      [e290c5810] Update test_modeling_flax_common.py
      [d3af86f46] comment
      [ff7dd32d8] more comments
      [59b137889] suggestion
      [7e2ba6d67] attn_implementation as attribute of the class
      [fe66ab71f] minor
      [38642b568] Apply suggestions from code review
      
      Accept comments
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [22cde7d52] Update tests/test_modeling_common.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [48e137cc6] Update tests/test_modeling_common.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [99f4c679f] Update tests/test_modeling_common.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [96cf20a6d] Update src/transformers/models/vit_msn/modeling_vit_msn.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [c59377d23] Update src/transformers/models/vit_mae/modeling_vit_mae.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [b70a47259] Update tests/models/vision_text_dual_encoder/test_modeling_vision_text_dual_encoder.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      [00c84d216] [run-slow]audio_spectrogram_transformer,deit,vision_encoder_decoder,vision_text_dual_encoder,vit,vit_hybrid,vit_mae,vit_msn,videomae,yolos
      [61f00ebb0] all tests are passing locally
      [e9e0b82b7] vision encoder/decoder
      [4d5076b56] test-vision (+20 squashed commits)
      Squashed commits:
      [d1add8db9] yolo
      [9fde65716] fix flax
      [986566c28] minor
      [ca2f21d1f] vit
      [3333efd7a] easy models change
      [ebfc21402] [run-slow]audio_spectrogram_transformer,deit,vision_encoder_decoder,vision_text_dual_encoder,vit,vit_hybrid,vit_mae,vit_msn,videomae,yolos
      [b8b8603ed] [run-slow]vision_encoder_decoder,vision_text_dual_encoder,yolos
      [48ecc7e26] all tests are passing locally
      [bff7fc366] minor
      [62f88306f] fix yolo and text_encoder tests
      [121507555] [run-slow]audio_spectrogram_transformer,deit,vit,vit_hybrid,vit_mae,vit_msn,videomae
      [1064cae0a] [run-slow]vision_encoder_decoder,vision_text_dual_encoder,yolos
      [b7f52ff3a] [run-slow]audio_spectrogram_transformer,deit,vit,vit_hybrid,vit_mae,vit_msn,videomae
      [cffaa10dd] fix-copies
      [ef6c511c4] test vit hybrid
      [7d4ba8644] vit hybrid
      [66f919033] [run-slow]audio_spectrogram_transformer,deit,vit,vit_hybrid,vit_mae,vit_msn,videomae
      [1fcc0a031] fixes
      [cfde6eb21] fixup
      [e77df1ed3] all except yolo end encoder decoder (+17 squashed commits)
      Squashed commits:
      [602913e22] vit + vit_mae are working
      [547f6c4cc] RUN_SLOW=1 pytest tests/models/audio_spectrogram_transformer/ tests/models/deit/ tests/models/videomae/  passes
      [61a97dfa9] it s the complete opposite...
      [aefab37d4] fix more tests
      [71802a1b9] fix all torch tests
      [40b12eb58] encoder - decoder tests
      [941552b69] slow decorator where appropriate
      [14d055d80] has_attentions to yolo and msn
      [3381fa19f] add correct name
      [e261316a7] repo consistency
      [31c6d0c08] fixup
      [9d214276c] minor fix
      [11ed2e1b7] chore
      [eca6644c4] add sdpa to vit-based models
      [cffbf390b] make fix-copies result
      [6468319b0] fix style
      [d324cd02a] add sdpa for vit
      Co-authored-by: default avatarLiubov Yaronskaya <luba.yaronskaya@gmail.com>
      1c21f48a
  19. 15 May, 2024 2 commits
  20. 13 May, 2024 2 commits
  21. 09 May, 2024 1 commit
  22. 07 May, 2024 3 commits
  23. 03 May, 2024 1 commit
  24. 30 Apr, 2024 1 commit
  25. 25 Apr, 2024 1 commit
  26. 24 Apr, 2024 6 commits