1. 30 Apr, 2024 1 commit
  2. 29 Apr, 2024 1 commit
  3. 22 Apr, 2024 1 commit
  4. 18 Apr, 2024 2 commits
  5. 17 Apr, 2024 2 commits
    • fxmarty's avatar
      Fix quality Olmo + SDPA (#30302) · ec92f983
      fxmarty authored
      fix olmo
      ec92f983
    • Shane A's avatar
      Add OLMo model family (#29890) · e4ea19b9
      Shane A authored
      * Add OLMo using add-new-model-like with Llama
      
      * Fix incorrect tokenizer for OLMo
      
      * Copy-paste relevant OLMo methods and their imports
      
      * Add OLMo config
      
      * Modify OLMo config to follow HF conventions
      
      * Remove unneeded Llama code from OLMo model
      
      * Add ability for OLMo model to output attentions
      
      * Add OLMoPreTrainedModel and OLMoModel
      
      * Add OLMoForCausalLM
      
      * Minor fixes to OLMo model for style and missing functions
      
      * Implement OLMo tokenizer
      
      * Implement OLMo to HF conversion script
      
      * Add tests for OLMo model
      
      * Add tests for OLMo fast tokenizer
      
      * Add auto-generated dummy objects
      
      * Remove unimplemented OLMo classes from auto and init classes and re-format
      
      * Add README and associated auto-generated files
      
      * Use OLMo names for common properties
      
      * Run make fixup
      
      * Remove `|` from OLMo typing
      
      * Remove unneeded tokenization_olmo.py
      
      * Revert model, config and converter to add-new-model-like Llama
      
      * Move logic for adding bos/eos token into GPTNeoxTokenizerFast
      
      * Change OLMoConfig defaults to match OLMo-7B
      
      * Use GPTNeoXToknizerFast in OLMo tokenizer tests
      
      * Modify auto-generated OLMoModelTests to work for OLMo
      
      * Add non-parametric layer norm OLMoLayerNorm
      
      * Update weight conversion script for OLMo
      
      * Fix __init__ and auto structure for OLMo
      
      * Fix errors from make fixup
      
      * Remove OLMoTokenizerFast from documentation
      
      * Add missing 'Copied from' for OLMoModel._update_causal_mask
      
      * Run make fix-copies
      
      * Rearrange string replacements in OLMoForCausalLM Copied from
      
      * Move OLMo and Llama CausalLM.forward example into global constants
      
      * Fix OLMO_GENERATION_EXAMPLE doc string typo
      
      * Add option for qkv clipping to OLMo
      
      * Rearrange OLMoConfig kwargs in convert_olmo_weights_to_hf
      
      * Add clip_qkv to OLMoConfig in convert_olmo_weights_to_hf
      
      * Fix OLMo tokenization bug using conversion script
      
      * Keep model in full precision after conversion
      
      * Do not add eos token automatically
      
      * Update references to OLMo model in HF Hub
      
      * Do not add eos token during encoding by default
      
      * Fix Llama generation example
      
      * Run make fixup
      
      * OLMo 7B integration test fix
      
      * Remove unneeded special case for OLMoConfig
      
      * OLMo 7B Twin 2T integration test fix
      
      * Fix test_model_7b_greedy_generation
      
      * Remove test_compile_static_cache
      
      * Fix OLMo and Llama generation example
      
      * Run make fixup
      
      * Revert "OLMo 7B integration test fix"
      
      This reverts commit 4df56a4b150681bfa559846f40e9b7b7f97d7908.
      
      * Revert "OLMo 7B Twin 2T integration test fix"
      
      This reverts commit 9ff65a4a294ace89ab047b793ca55e623a9ceefc.
      
      * Ungate 7B integration tests and fix greedy generation test
      
      * Add retries for flaky test_eager_matches_sdpa_generate
      
      * Fix output of doc example for OLMoForCausalLM.forward
      
      * Downsize OLMo doc test for OLMoForCausalLM.forward to 1B model
      
      * Try fix incorrect characters in OLMoForCausalLM.forward doct test
      
      * Try fix incorrect characters in OLMoForCausalLM.forward doc test using end quotes
      
      * Remove pretraining_tp from OLMo config and model
      
      * Add missing 'Copied from' instances
      
      * Remove unneeded causal_mask from OLMoModel
      
      * Revert Llama changes
      
      * Ignore copy for OLMoForCausalLM.forward
      
      * Change 'OLMo' to 'Olmo' in classes
      
      * Move minimal OLMo tokenization tests to model tests
      
      * Add missed 'Copied from' for repeat_kv
      e4ea19b9
  6. 05 Apr, 2024 1 commit
  7. 04 Apr, 2024 1 commit
  8. 30 Mar, 2024 1 commit
  9. 28 Mar, 2024 1 commit
  10. 22 Mar, 2024 1 commit
  11. 21 Mar, 2024 1 commit
  12. 20 Mar, 2024 1 commit
    • Arthur's avatar
      [`BC 4.37 -> 4.38`] for Llama family, memory and speed (#29753) · ff841900
      Arthur authored
      * attempt to fix
      
      * the actual fix that works with compilation!
      
      * this?
      
      * temporary update
      
      * nit?
      
      * dispatcg to memory efficient?
      
      * update both models that have static cache support
      
      * fix copies fix compile
      
      * make sure fix
      
      * fix cohere and gemma
      
      * fix beams?
      
      * nit
      
      * slipped through the cracks
      
      * nit
      
      * nits
      
      * update
      
      * fix-copies
      
      * skip failing tests
      
      * nits
      ff841900
  13. 19 Mar, 2024 1 commit
  14. 15 Mar, 2024 1 commit
    • Saurabh Dash's avatar
      Cohere Model Release (#29622) · 0e4a1c34
      Saurabh Dash authored
      
      
      * Cohere Model Release (#1)
      
      Cohere Model Release
      
      * Remove unnecessary files and code (#2)
      
      Some cleanup
      
      * Delete cohere-model directory (#3)
      
      * Make Fix (#5)
      
      * Pr fixes (#6)
      
      * fixes for pr
      
      * pr fixes for the format
      
      * pr fixes for the format
      
      * src/transformers/models/auto/tokenization_auto.py
      
      * Tokenizer test (#8)
      
      * tokenizer test
      
      * format fix
      
      * Adding Docs and other minor changes (#7)
      
      * Add modeling tests (#9)
      
      * Smol Fix (#11)
      
      * tokenization tests are fixed
      
      * format fixes
      
      * fix pr doc tests
      
      * fix pr doc tests
      
      * fix pr doc tests
      
      * fix pr style check
      
      * small changes in cohere.md
      
      * FIX: Address final comments for transformers integration (#13)
      
      * fix modeling final nits and add proper test file
      
      * for now leave empty tests
      
      * add integration test
      
      * push new test
      
      * fix modeling cohere (#14)
      
      * Update chat templates to use the new API (#15)
      
      ---------
      Co-authored-by: default avatarahmetustun <ahmetustun89@gmail.com>
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      0e4a1c34
  15. 14 Mar, 2024 1 commit
  16. 13 Mar, 2024 1 commit
  17. 08 Mar, 2024 1 commit
  18. 06 Mar, 2024 2 commits
  19. 01 Mar, 2024 1 commit
  20. 28 Feb, 2024 3 commits
  21. 26 Feb, 2024 1 commit
  22. 23 Feb, 2024 1 commit
  23. 22 Feb, 2024 2 commits
  24. 21 Feb, 2024 4 commits
  25. 20 Feb, 2024 1 commit
  26. 15 Feb, 2024 1 commit
    • Arthur's avatar
      Fix static generation when compiling! (#28937) · f3788b09
      Arthur authored
      
      
      * wow I was scared!
      
      * fix everything
      
      * nits
      
      * make it BC?
      
      * add todo
      
      * nits
      
      * is_tracing should still be used to pass tracing tests
      
      * nits
      
      * some nits to make sure genration works with static cache uncompiled
      
      * fix sdpa
      
      * fix FA2 for both static and dynamic in a better way?
      
      * style
      
      * fix-copies
      
      * fix fix copies
      
      * fix sequential beam searcg
      
      * style
      
      * use `keys_to_ignore`
      
      * nit
      
      * correct dtype inference when init
      
      * :( the fix for FA2 is still not optimal to investigate!
      
      * styling
      
      * nits
      
      * nit
      
      * this might work better
      
      * add comment
      
      * Update src/transformers/models/llama/modeling_llama.py
      
      * "position_ids" -> "cache_position"
      
      * style
      
      * nit
      
      * Remove changes that should no be propagatted just yet
      
      * Apply suggestions from code review
      
      * Styling
      
      * make sure we raise an errir for static cache with FA2 enabled
      
      * move  to the bottom of the signature
      
      * style
      
      * Update src/transformers/models/llama/modeling_llama.py
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      
      * Update src/transformers/models/llama/modeling_llama.py
      
      * nit in the name
      
      ---------
      Co-authored-by: default avatarYounes Belkada <49240599+younesbelkada@users.noreply.github.com>
      f3788b09
  27. 08 Feb, 2024 2 commits
  28. 06 Feb, 2024 1 commit
  29. 31 Jan, 2024 1 commit
  30. 15 Jan, 2024 1 commit