1. 26 Jun, 2024 3 commits
    • Pavel Iakubovskii's avatar
      Fix RT-DETR inference with float16 and bfloat16 (#31639) · b1ec7454
      Pavel Iakubovskii authored
      
      
      * [run_slow] rt_detr
      
      * Fix positional embeddings and anchors dtypes
      
      * [run slow] rt_detr
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Fixup
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      b1ec7454
    • Younes Belkada's avatar
      Llama et al. / FSDP : Fix breaking change in 4.40 for FSDP (#31161) · 3f93fd06
      Younes Belkada authored
      
      
      * fix llama fsdp
      
      * fixup
      
      * adding FSDP tests for CPU offloading
      
      * fixes
      
      * fix tests
      
      * fix tests
      
      * add it for mixtral
      
      * propagate the changes on other models
      
      * Update src/transformers/models/phi/modeling_phi.py
      
      * Delete utils/testing_scripts/fsdp_cpu_offloading.py
      
      Remove script - FSDP + CPU offloading it tested in the test suite
      
      * Delete utils/testing_scripts/dummy_fsdp_config.yml
      
      * Update + add cache_positions docstring
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      3f93fd06
    • Anton Vlasjuk's avatar
      [`GPT-NeoX`] Add SDPA support (#31031) · b07770c5
      Anton Vlasjuk authored
      * starting support for sdpa in `gptneox` models
      
      * small comment on tests
      
      * fix dropout
      
      * documentation and style
      
      * clarify concrete paths for reference
      
      * generalise attn projections and rope application
      
      added head mask check to sdpa mask creation
      
      handle sdpa memory backend bug via own version flag
      
      * update docs and style
      
      * move dtype casting outside of general attn_projection_and_rope function
      
      fix flash_attn_2 stuff
      
      * more generic attn warning if output_attns or head_mask
      
      * simplify head mask check by moving head mask creation to a later point
      
      * remove copied llama artifact
      
      * remove padding_mask from attention function signature
      
      * removing unnecessary comments, only "save" attn implementation once
      
      * [run_slow] gpt_neox
      b07770c5
  2. 25 Jun, 2024 4 commits
  3. 24 Jun, 2024 3 commits
  4. 21 Jun, 2024 2 commits
  5. 20 Jun, 2024 3 commits
  6. 19 Jun, 2024 4 commits
  7. 18 Jun, 2024 2 commits
  8. 17 Jun, 2024 5 commits
  9. 14 Jun, 2024 3 commits
  10. 13 Jun, 2024 1 commit
    • Pablo Montalvo's avatar
      add initial design for uniform processors + align model (#31197) · c624d5ba
      Pablo Montalvo authored
      * add initial design for uniform processors + align model
      
      * fix mutable default 馃憖
      
      * add configuration test
      
      * handle structured kwargs w defaults + add test
      
      * protect torch-specific test
      
      * fix style
      
      * fix
      
      * fix assertEqual
      
      * move kwargs merging to processing common
      
      * rework kwargs for type hinting
      
      * just get Unpack from extensions
      
      * run-slow[align]
      
      * handle kwargs passed as nested dict
      
      * add from_pretrained test for nested kwargs handling
      
      * [run-slow]align
      
      * update documentation + imports
      
      * update audio inputs
      
      * protect audio types, silly
      
      * try removing imports
      
      * make things simpler
      
      * simplerer
      
      * move out kwargs test to common mixin
      
      * [run-slow]align
      
      * skip tests for old processors
      
      * [run-slow]align, clip
      
      * !$#@!! protect imports, darn it
      
      * [run-slow]align, clip
      
      * [run-slow]align, clip
      
      * update doc
      
      * improve documentation for default values
      
      * add model_max_length testing
      
      This parameter depends on tokenizers received.
      
      * Raise if kwargs are specified in two places
      
      * fix
      
      * expand VideoInput
      
      * fix
      
      * fix style
      
      * remove defaults values
      
      * add comment to indicate documentation on adding kwargs
      
      * protect imports
      
      * [run-slow]align
      
      * fix
      
      * remove set() that breaks ordering
      
      * test more
      
      * removed unused func
      
      * [run-slow]align
      c624d5ba
  11. 12 Jun, 2024 2 commits
  12. 11 Jun, 2024 2 commits
    • amyeroberts's avatar
      Fast image processor (#28847) · f53fe35b
      amyeroberts authored
      
      
      * Draft fast image processors
      
      * Draft working fast version
      
      * py3.8 compatible cache
      
      * Enable loading fast image processors through auto
      
      * Tidy up; rescale behaviour based on input type
      
      * Enable tests for fast image processors
      
      * Smarter rescaling
      
      * Don't default to Fast
      
      * Safer imports
      
      * Add necessary Pillow requirement
      
      * Woops
      
      * Add AutoImageProcessor test
      
      * Fix up
      
      * Fix test for imagegpt
      
      * Fix test
      
      * Review comments
      
      * Add warning for TF and JAX input types
      
      * Rearrange
      
      * Return transforms
      
      * NumpyToTensor transformation
      
      * Rebase - include changes from upstream in ImageProcessingMixin
      
      * Safe typing
      
      * Fix up
      
      * convert mean/std to tesnor to rescale
      
      * Don't store transforms in state
      
      * Fix up
      
      * Update src/transformers/image_processing_utils_fast.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/auto/image_processing_auto.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Warn if fast image processor available
      
      * Update src/transformers/models/vit/image_processing_vit_fast.py
      
      * Transpose incoming numpy images to be in CHW format
      
      * Update mapping names based on packages, auto set fast to None
      
      * Fix up
      
      * Fix
      
      * Add AutoImageProcessor.from_pretrained(checkpoint, use_fast=True) test
      
      * Update src/transformers/models/vit/image_processing_vit_fast.py
      Co-authored-by: default avatarPavel Iakubovskii <qubvel@gmail.com>
      
      * Add equivalence and speed tests
      
      * Fix up
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarPavel Iakubovskii <qubvel@gmail.com>
      f53fe35b
    • Matt's avatar
      Chat Template support for function calling and RAG (#30621) · edc1dffd
      Matt authored
      
      
      * First draft, still missing automatic function conversion
      
      * First draft of the automatic schema generator
      
      * Lots of small fixes
      
      * the walrus has betrayed me
      
      * please stop committing your debug breakpoints
      
      * Lots of cleanup and edge cases, looking better now
      
      * Comments and bugfixes for the type hint parser
      
      * More cleanup
      
      * Add tests, update schema generator
      
      * Update tests, proper handling of return values
      
      * Small docstring change
      
      * More doc updates
      
      * More doc updates
      
      * Add json_schema decorator
      
      * Clean up the TODOs and finish the docs
      
      * self.maxDiff = None to see the whole diff for the nested list test
      
      * add import for add_json_schema
      
      * Quick test fix
      
      * Fix something that was bugging me in the chat template docstring
      
      * Less "anyOf" when unnecessary
      
      * Support return types for the templates that need them
      
      * Proper return type tests
      
      * Switch to Google format docstrings
      
      * Update chat templating docs to match new format
      
      * Stop putting the return type in with the other parameters
      
      * Add Tuple support
      
      * No more decorator - we just do it implicitly!
      
      * Add enum support to get_json_schema
      
      * Update docstring
      
      * Add copyright header
      
      * Update src/transformers/tokenization_utils_base.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update docs/source/en/chat_templating.md
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/utils/chat_template_utils.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update src/transformers/utils/chat_template_utils.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Add copyright header
      
      * make fixup
      
      * Fix indentation
      
      * Reformat chat_template_utils
      
      * Correct return value
      
      * Make regexes module-level
      
      * Support more complex, multi-line arg docstrings
      
      * Update error message for ...
      
      * Update ruff
      
      * Add document type validation
      
      * Refactor docs
      
      * Refactor docs
      
      * Refactor docs
      
      * Clean up Tuple error
      
      * Add an extra test for very complex defs and docstrings and clean everything up for it
      
      * Document enum block
      
      * Quick test fixes
      
      * Stop supporting type hints in docstring to fix bugs and simplify the regex
      
      * Update docs for the regex change
      
      * Clean up enum regex
      
      * Wrap functions in {"type": "function", "function": ...}
      
      * Update src/transformers/utils/chat_template_utils.py
      Co-authored-by: default avatarPablo Montalvo <39954772+molbap@users.noreply.github.com>
      
      * Temporary tool calling commit
      
      * Add type hints to chat template utils, partially update docs (incomplete!)
      
      * Code cleanup based on @molbap's suggestion
      
      * Add comments to explain regexes
      
      * Fix up type parsing for unions and lists
      
      * Add custom exception types and adjust tests to look for them
      
      * Update docs with a demo!
      
      * Docs cleanup
      
      * Pass content as string
      
      * Update tool call formatting
      
      * Update docs with new function format
      
      * Update docs
      
      * Update docs with a second tool to show the model choosing correctly
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      Co-authored-by: default avatarPablo Montalvo <39954772+molbap@users.noreply.github.com>
      edc1dffd
  13. 10 Jun, 2024 4 commits
  14. 07 Jun, 2024 2 commits
    • amyeroberts's avatar
      Rename test_model_common_attributes -> test_model_get_set_embeddings (#31321) · 25245ec2
      amyeroberts authored
      * Rename to test_model_common_attributes
      The method name is misleading - it is testing being able to get and set embeddings, not common attributes to all models
      
      * Explicitly skip
      25245ec2
    • BHUVAN M's avatar
      interpolation added for TVP. (#30863) · 3b9174f2
      BHUVAN M authored
      * Update TVP model to interpolate pre-trained image pad prompter encodings
      
      * feat: Add 2D positional embeddings interpolation in TvpVisualInputEmbedding
      
      * added required comments
      
      * Update TVP model to interpolate pre-trained image pad prompter encodings
      
      * feat: Add 2D positional embeddings interpolation in TvpVisualInputEmbedding
      
      * added required comments
      
      * docstring and argument fix
      
      * doc fixes and test case fix suggested in review.
      
      * varibale typo fix
      
      * styling and name fixes for padding interpolation flag.
      3b9174f2