1. 19 May, 2023 11 commits
    • Younes Belkada's avatar
      [`Blip`] Remove redundant shift right (#23153) · 3cb93090
      Younes Belkada authored
      * remove redundant shit right
      
      * fix failing tests
      
      * this time fix tests
      3cb93090
    • Dennis Loevlie's avatar
      Fix: Change tensors to integers for torch.dynamo and torch.compile compatibility (#23475) · 847e5691
      Dennis Loevlie authored
      * Fix: Change tensors to integers in torch.split() for torch.dynamo and torch.compile compatibility
      
      * Applied the suggested fix to the utils/check_copies.py test
      
      * Applied the suggested fix by changing the original function that gets copied
      847e5691
    • joaoareis's avatar
      389bdba6
    • Zachary Mueller's avatar
      Fix parallel mode check (#23409) · b455ad0a
      Zachary Mueller authored
      * Fix sagemaker/distributed state
      
      * Fix correctly
      
      * Bring back -1
      
      * Bring back local rank for distributed check
      
      * better version
      
      * Cleanest option
      b455ad0a
    • Yih-Dar's avatar
      Fix `transformers`' DeepSpeed CI job (#23463) · db4d7652
      Yih-Dar authored
      
      
      * fix
      
      * fix
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      db4d7652
    • Sylvain Gugger's avatar
      Use config to set name and description if not present (#23473) · 2aa0cc2c
      Sylvain Gugger authored
      Use config to set name and descriptiob if not present
      2aa0cc2c
    • Younes Belkada's avatar
      [`RWKV`] Rwkv fix for 8bit inference (#23468) · 21bd3be1
      Younes Belkada authored
      * rwkv fix for 8bit inference
      
      * add comment
      21bd3be1
    • Matt's avatar
      TF port of the Segment Anything Model (SAM) (#22970) · 1c460a52
      Matt authored
      
      
      * First commit
      
      * Add auto-translation with GPT-4
      
      * make fixup
      
      * Add a functional layernorm for TF
      
      * Add all the auxiliary imports etc.
      
      * Add the extra processor and tests
      
      * rebase to main
      
      * Add all the needed fixes to the GPT code
      
      * make fixup
      
      * Make convolutions channels-last so they run on CPU
      
      * make fixup
      
      * Fix final issues
      
      * Fix other models affected by test change
      
      * Clarify comment on the sparse_prompt_embeddings check
      
      * Refactor functional_layernorm, use shape_list in place of .shape in some places
      
      * Remove deprecated torch-alike code
      
      * Update tests/models/sam/test_modeling_tf_sam.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Update tests/models/sam/test_modeling_tf_sam.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Refactor processor with common methods and separated private methods
      
      * make fixup
      
      * Quietly delete the file that didn't do anything (sorry Sylvain)
      
      * Refactor the processor tests into one file
      
      * make fixup
      
      * Clean up some unnecessary indirection
      
      * Fix TF mask postprocessing
      
      * Add more processor equivalence tests
      
      * Refactor generate_crop_boxes to use framework-neutral np code
      
      * Make the serving output correctly conditional
      
      * Fix error message line length
      
      * Use dict keys rather than indices internally in both TF and PT SAM call/forward
      
      * Return dicts internally in the call/forward methods
      
      * Revert changes to common tests and just override check_pt_tf_outputs
      
      * Revert changes to other model tests
      
      * Clarify comments for functional layernorm
      
      * Add missing transpose from PT code
      
      * Removed unused copied from in PT code
      
      * Remove overrides for tests that don't exist in TF
      
      * Fix transpose and update tests for PT and TF to check pred_masks
      
      * Add training flag
      
      * Update tests to use TF checkpoints
      
      * Update index.mdx
      
      * Add missing cross-test decorator
      
      * Remove optional extra asterisks
      
      * Revert return_dict changes in PT code
      
      * Update src/transformers/models/sam/modeling_tf_sam.py
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      
      * Remove None return annotations on init methods
      
      * Update tests/models/sam/test_processor_sam.py
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Fix input_boxes shapes
      
      * make fixup
      
      ---------
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      Co-authored-by: default avatarSylvain Gugger <35901082+sgugger@users.noreply.github.com>
      1c460a52
    • Jiewen Tan's avatar
      Remove .data usages in optimizations.py (#23417) · 8aa8513f
      Jiewen Tan authored
      Patched the optimizers
      8aa8513f
    • Julien Chaumond's avatar
      README: Fix affiliation for MEGA (#23394) · 3cf01b20
      Julien Chaumond authored
      
      
      * README: Fix affiliation for MEGA
      
      * Fix quality
      
      ---------
      Co-authored-by: default avatarLysandre <lysandre@huggingface.co>
      3cf01b20
    • Connor Henderson's avatar
      feat: Whisper prompting (#22496) · 2acedf47
      Connor Henderson authored
      * initial working additions
      
      * clean and rename, add cond stripping initial prompt to decode
      
      * cleanup, edit create_initial_prompt_ids, add tests
      
      * repo consistency, flip order of conditional
      
      * fix error, move the processor fn to the tokenizer
      
      * repo consistency, update test ids to corresponding tokenizer
      
      * use convert_tokens_to_ids not get_vocab...
      
      * use actual conditional in generate
      
      * make sytle
      
      * initial address comments
      
      * initial working add new params to pipeline
      
      * first draft of sequential generation for condition_on_previous_text
      
      * add/update tests, make compatible with timestamps
      
      * make compatible with diff. input kwargs and max length
      
      * add None check
      
      * add temperature check
      
      * flip temp check operand
      
      * refocusing to prev pr scope
      
      * remove the params too
      
      * make style
      
      * edits, move max length incorporating prompt to whisper
      
      * address comments
      
      * remove asr pipeline prompt decoding, fix indexing
      
      * address comments (more tests, validate prompt)
      
      * un-comment out tests (from debug)
      
      * remove old comment
      
      * address comments
      
      * fix typo
      
      * remove timestamp token from test
      
      * make style
      
      * cleanup
      
      * copy method to fast tokenizer, set max_new_tokens for test
      
      * prompt_ids type just pt
      
      * address Amy's comments
      
      * make style
      2acedf47
  2. 18 May, 2023 15 commits
  3. 17 May, 2023 14 commits