"docs/vscode:/vscode.git/clone" did not exist on "a6d178e23892a2b833523bb1ce4611b32e171f4f"
  1. 19 Apr, 2024 14 commits
    • Jo茫o David's avatar
      Add TF swiftformer (#23342) · d2cec09b
      Jo茫o David authored
      
      
      * Duplicate swiftformer
      
      * Convert SwiftFormerPatchEmbedding
      
      * Convert SwiftFormerEmbeddings
      
      * Convert TFSwiftFormerMlp
      
      * Convert TFSwiftFormerConvEncoder
      
      * Convert TFSwiftFormerLocalRepresentation
      
      * convert TFSwiftFormerEncoderBlock
      
      * Convert SwiftFormerStage
      
      * Convert SwiftFormerEncoder
      
      * Add TFSWiftFormerPreTrainedModel
      
      * Convert SwiftFormerForImageClassification
      
      * Add kwargs and start drop path
      
      * Fix syntax
      
      * Change Model class name
      
      * Add TFSwiftFormer to __init__
      
      * Duplicate test_modeling_swiftformer
      
      * First test conversions
      
      * Change require_torch to require_tf
      
      * Add exports to swiftformer __init__
      
      * Add TFSwiftFormerModel wrapper
      
      * Fix __init__ and run black
      
      * Remove docstring from MainLayer, fix padding
      
      * Use keras.layers.Activation on keras.Sequential
      
      * Fix swiftformer exports
      
      * Fix activation layer from config
      
      * Remove post_inits
      
      * Use tf.keras.layers.ZeroPadding2D
      
      * Convert torch normalize
      
      * Change tf test input shape
      
      * Fix softmax and reduce_sum
      
      * Convert expand_dims and repeat
      
      * Add missing reshape and tranpose
      
      * Simplify TFSwiftFormerEncoderBlock.call
      
      * Fix mismatch in patch embeddings
      
      * Fix expected output shape to match channels last
      
      * Fix swiftformer typo
      
      * Disable test_onnx
      
      * Fix TFSwiftFormerForImageClassification call
      
      * Add unpack inputs
      
      * Convert flatten(2).mean(-1)
      
      * Change vision dummy inputs (to be reviewed)
      
      * Change test_forward_signature to use .call
      
      * Fix @unpack_inputs
      
      * Set return_tensors="tf" and rename class
      
      * Rename wrongly named patch_embeddings layer
      
      * Add serving_output and change dummy_input shape
      
      * Make dimensions BCHW and transpose inside embedding layer
      
      * Change SwiftFormerEncoderBlock
      
      * Fix ruff problems
      
      * Add image size to swiftformer config
      
      * Change tranpose to MainLayer and use -1 for reshape
      
      * Remove serving_outputs and dummy_inputs
      
      * Remove test_initialization test from tf model
      
      * Make Sequential component a separate layer
      
      * Fix layers' names
      
      * Tranpose encoder outputs
      
      * Fix tests and check if hidden states is not None
      
      * Fix TFSwiftFormerForImageClassification
      
      * Run make fixup
      
      * Run make fix-copies
      
      * Update modeling_tf_auto
      
      * Update docs
      
      * Fix modeling auto mapping
      
      * Update modelint_tf_swiftformer docs
      
      * Fill image_size doc and type
      
      * Add reduction=None to loss computation
      
      * Update docs
      
      * make style
      
      * Debug: Delete the tip to see if that changes anything
      
      * Re-add tip
      
      * Remove add_code_sample_docstrings
      
      * Remove unused import
      
      * Get the debug to actually tell us the problem it has with the docs
      
      * Try a substitution to match the PyTorch file?
      
      * Add swiftformer to ignore list
      
      * Add build() methods
      
      * Update copyright year
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Remove FIXME comment
      
      * Remove from_pt
      
      * Update copyright year
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Rename one-letter variables
      
      * Remove FIXMEs related to momentum
      
      * Remove old TODO comment
      
      * Remove outstanding FIXME comments
      
      * Get dropout rate from config
      
      * Add specific dropout config for MLP
      
      * Add convencoder dropout to config
      
      * Pass config to SwiftFormerDropPath layer
      
      * Fix drop_path variable name and add Adapted from comment
      
      * Run ruff
      
      * Removed copied from comment
      
      * Run fix copies
      
      * Change drop_path to identity to match pt
      
      * Cleanup build() methods and move to new keras imports
      
      * Update docs/source/en/model_doc/swiftformer.md
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Raise error if drop_path_rate > 0.0
      
      * Apply suggestions from code review
      
      Replace (self.dim), with self.dim,
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      
      * Remove drop_path function
      
      * Add training to TFSwiftFormerEncoder
      
      * Set self.built = True last
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Should have been added to previous commit
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Apply suggestions from code review
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Change default_feature_extractor to default_image_processor
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      
      * Import Keras from modeling_tf_utils
      
      * Remove relative import
      
      * Run ruff --fix
      
      * Move import keras to tf_available
      
      * Add copied from comment to test_forward_signature
      
      * Reduce batch size and num_labels
      
      * Extract loss logic to hf_compute_loss
      
      * Run ruff format
      
      ---------
      Co-authored-by: default avatarMatt <rocketknight1@gmail.com>
      Co-authored-by: default avataramyeroberts <22614925+amyeroberts@users.noreply.github.com>
      Co-authored-by: default avatarMatt <Rocketknight1@users.noreply.github.com>
      d2cec09b
    • hoshi-hiyouga's avatar
      Fix config + attn_implementation in AutoModelForCausalLM.from_pretrained (#30299) · 21c912e7
      hoshi-hiyouga authored
      * Update modeling_utils.py
      
      * Update test_modeling_utils.py
      
      * Update test_modeling_utils.py
      
      * Update test_modeling_utils.py
      21c912e7
    • Raushan Turganbay's avatar
      Do not remove half seq length in generation tests (#30016) · b1cd4874
      Raushan Turganbay authored
      
      
      * remove seq length from generation tests
      
      * style and quality
      
      * [test_all] & PR suggestion
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      
      * Update tests/generation/test_utils.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * [test all] remove unused variables
      
      ---------
      Co-authored-by: default avatarJoao Gante <joaofranciscocardosogante@gmail.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      b1cd4874
    • Marc Sun's avatar
      Update unwrap from accelerate (#29933) · b4fd49b6
      Marc Sun authored
      
      
      * Use unwrap with the one in accelerate
      
      * oups
      
      * update unwrap
      
      * fix
      
      * wording
      
      * raise error instead
      
      * comment
      
      * doc
      
      * Update src/transformers/modeling_utils.py
      Co-authored-by: default avatarZach Mueller <muellerzr@gmail.com>
      
      * style
      
      * put else
      
      ---------
      Co-authored-by: default avatarZach Mueller <muellerzr@gmail.com>
      b4fd49b6
    • Yoach Lacombe's avatar
      Restore casting of masked_spec_embed (#30336) · fbd8c51f
      Yoach Lacombe authored
      * fix Parameter dtype in audio models
      
      * restore casting of masked_spec_embed
      
      * restore casting of masked_spec_embed
      fbd8c51f
    • Matt's avatar
      Deprecate default chat templates (#30346) · 0927bfd0
      Matt authored
      * initial commit, remove warnings on default chat templates
      
      * stash commit
      
      * Raise a much sterner warning for default chat templates, and prepare for depreciation
      
      * Update the docs
      0927bfd0
    • Lysandre Debut's avatar
      Transformers Metadata (#30344) · e67ccf06
      Lysandre Debut authored
      e67ccf06
    • Yih-Dar's avatar
      parallel job limit for doctest (#30342) · 32d4bef6
      Yih-Dar authored
      
      
      limit
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      32d4bef6
    • Sanchit Gandhi's avatar
      [Whisper] Fix slow tests (#30152) · 4ed0e51c
      Sanchit Gandhi authored
      
      
      * fix tests
      
      * style
      
      * more fixes
      
      * move model to device
      
      * move logits to cpu
      
      * update expected values
      
      * use ungated dataset
      
      * fix
      
      * fix
      
      * update
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      4ed0e51c
    • Raushan Turganbay's avatar
      Pipeline: fix `pad_token_id` again (#30338) · 91472cf5
      Raushan Turganbay authored
      fix again
      91472cf5
    • Sanchit Gandhi's avatar
      cd09a8df
    • Bharat Ramanathan's avatar
      feat: Upgrade Weights & Biases callback (#30135) · 4ab7a282
      Bharat Ramanathan authored
      * feat: upgrade wandb callback with new features
      
      * fix: ci issues with imports and run fixup
      4ab7a282
    • Jacky Lee's avatar
      Enable multi-device for some models (#30207) · 30b45320
      Jacky Lee authored
      
      
      * feat: multidevice for resnet
      
      * feat: yes! resnet
      
      * fix: compare all elements in tuple
      
      * feat: support for regnet
      
      * feat: support for convnextv2
      
      * feat: support for bit
      
      * feat: support for cvt
      
      * feat: add support for focalnet
      
      * feat: support for yolos
      
      * feat: support for glpn
      
      * feat: support for imagegpt
      
      * feat: support for levit
      
      * feat: support for mgp_str
      
      * feat: support for mobilnet_v1
      
      * feat: support for mobilnet_v2
      
      * feat: support for mobilevit
      
      * feat: support for mobilevitv2
      
      * feat: support for poolformer
      
      * fix: copies
      
      * fix: code quality check
      
      * update: upstream changes from main
      
      * fix: consistency check
      
      * feat: support for sam
      
      * feat: support for switchformer
      
      * feat: support for swin
      
      * feat: support for swinv2
      
      * feat: support for timesformer
      
      * feat: suport for trocr
      
      * feat: support for upernet
      
      * fix: check copies
      
      * update: rerun CI
      
      * update: rerun again, maybe
      
      * update: one more rerun
      
      ---------
      Co-authored-by: default avatarJacky Lee <jackylee328@gmail.com>
      30b45320
    • NielsRogge's avatar
      [UDOP] Add special tokens to tokenizer (#29594) · ecfe9be7
      NielsRogge authored
      * Add special tokens
      
      * Add special tokens
      
      * Use fmt
      
      * Uncomment code
      
      * Add test
      
      * Remove scripts
      
      * Address comments
      
      * Improve tests
      
      * Address comment
      
      * Remove flag
      ecfe9be7
  2. 18 Apr, 2024 19 commits
  3. 17 Apr, 2024 7 commits
    • fxmarty's avatar
      Fix quality Olmo + SDPA (#30302) · ec92f983
      fxmarty authored
      fix olmo
      ec92f983
    • fxmarty's avatar
      Re-enable SDPA's FA2 path (#30070) · 05bdef16
      fxmarty authored
      
      
      * tentatively re-enable FA2 + SDPA
      
      * better comment
      
      * _ignore_causal_mask_sdpa as staticmethod
      
      * type hints
      
      * use past_seen_tokens instead
      
      * enable copied from for sdpa
      
      * ruff
      
      * llama simplifications on review
      
      * remove unnecessary self.is_causal check
      
      * fix copies
      
      * cleaning
      
      * precise message
      
      * better doc
      
      * add test
      
      * simplify
      
      * Update src/transformers/models/llama/modeling_llama.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/llama/modeling_llama.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Update src/transformers/models/llama/modeling_llama.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * style
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      05bdef16
    • Shane A's avatar
      Add OLMo model family (#29890) · e4ea19b9
      Shane A authored
      * Add OLMo using add-new-model-like with Llama
      
      * Fix incorrect tokenizer for OLMo
      
      * Copy-paste relevant OLMo methods and their imports
      
      * Add OLMo config
      
      * Modify OLMo config to follow HF conventions
      
      * Remove unneeded Llama code from OLMo model
      
      * Add ability for OLMo model to output attentions
      
      * Add OLMoPreTrainedModel and OLMoModel
      
      * Add OLMoForCausalLM
      
      * Minor fixes to OLMo model for style and missing functions
      
      * Implement OLMo tokenizer
      
      * Implement OLMo to HF conversion script
      
      * Add tests for OLMo model
      
      * Add tests for OLMo fast tokenizer
      
      * Add auto-generated dummy objects
      
      * Remove unimplemented OLMo classes from auto and init classes and re-format
      
      * Add README and associated auto-generated files
      
      * Use OLMo names for common properties
      
      * Run make fixup
      
      * Remove `|` from OLMo typing
      
      * Remove unneeded tokenization_olmo.py
      
      * Revert model, config and converter to add-new-model-like Llama
      
      * Move logic for adding bos/eos token into GPTNeoxTokenizerFast
      
      * Change OLMoConfig defaults to match OLMo-7B
      
      * Use GPTNeoXToknizerFast in OLMo tokenizer tests
      
      * Modify auto-generated OLMoModelTests to work for OLMo
      
      * Add non-parametric layer norm OLMoLayerNorm
      
      * Update weight conversion script for OLMo
      
      * Fix __init__ and auto structure for OLMo
      
      * Fix errors from make fixup
      
      * Remove OLMoTokenizerFast from documentation
      
      * Add missing 'Copied from' for OLMoModel._update_causal_mask
      
      * Run make fix-copies
      
      * Rearrange string replacements in OLMoForCausalLM Copied from
      
      * Move OLMo and Llama CausalLM.forward example into global constants
      
      * Fix OLMO_GENERATION_EXAMPLE doc string typo
      
      * Add option for qkv clipping to OLMo
      
      * Rearrange OLMoConfig kwargs in convert_olmo_weights_to_hf
      
      * Add clip_qkv to OLMoConfig in convert_olmo_weights_to_hf
      
      * Fix OLMo tokenization bug using conversion script
      
      * Keep model in full precision after conversion
      
      * Do not add eos token automatically
      
      * Update references to OLMo model in HF Hub
      
      * Do not add eos token during encoding by default
      
      * Fix Llama generation example
      
      * Run make fixup
      
      * OLMo 7B integration test fix
      
      * Remove unneeded special case for OLMoConfig
      
      * OLMo 7B Twin 2T integration test fix
      
      * Fix test_model_7b_greedy_generation
      
      * Remove test_compile_static_cache
      
      * Fix OLMo and Llama generation example
      
      * Run make fixup
      
      * Revert "OLMo 7B integration test fix"
      
      This reverts commit 4df56a4b150681bfa559846f40e9b7b7f97d7908.
      
      * Revert "OLMo 7B Twin 2T integration test fix"
      
      This reverts commit 9ff65a4a294ace89ab047b793ca55e623a9ceefc.
      
      * Ungate 7B integration tests and fix greedy generation test
      
      * Add retries for flaky test_eager_matches_sdpa_generate
      
      * Fix output of doc example for OLMoForCausalLM.forward
      
      * Downsize OLMo doc test for OLMoForCausalLM.forward to 1B model
      
      * Try fix incorrect characters in OLMoForCausalLM.forward doct test
      
      * Try fix incorrect characters in OLMoForCausalLM.forward doc test using end quotes
      
      * Remove pretraining_tp from OLMo config and model
      
      * Add missing 'Copied from' instances
      
      * Remove unneeded causal_mask from OLMoModel
      
      * Revert Llama changes
      
      * Ignore copy for OLMoForCausalLM.forward
      
      * Change 'OLMo' to 'Olmo' in classes
      
      * Move minimal OLMo tokenization tests to model tests
      
      * Add missed 'Copied from' for repeat_kv
      e4ea19b9
    • Nicolas Patry's avatar
      Upgrading to tokenizers 0.19.0 (#30289) · 8e5f76f5
      Nicolas Patry authored
      * [DO NOT MERGE] Testing tokenizers 0.19.0rc0
      
      * Accounting for the breaking change.
      
      * Ruff.
      
      * Upgrading to tokenizers `0.19` (new release with preprend_scheme fixed
      and new surface for BPE tiktoken bug).
      8e5f76f5
    • Pavel Iakubovskii's avatar
      Add strategy to store results in evaluation loop (#30267) · c15aad09
      Pavel Iakubovskii authored
      * Add evaluation loop container for interm. results
      
      * Add tests for EvalLoopContainer
      
      * Formatting
      
      * Fix padding_index in test and typo
      
      * Move EvalLoopContainer to pr_utils to avoid additional imports
      
      * Fix `eval_do_concat_batches` arg description
      
      * Fix EvalLoopContainer import
      c15aad09
    • st81's avatar
      Add token type ids to CodeGenTokenizer (#29265) · 8d6b5096
      st81 authored
      * Add create token type ids to CodeGenTokenizer
      
      * Fix inconsistent length of token type ids
      
      * Format source codes
      
      * Fix inconsistent order of methods
      
      * Update docstring
      
      * add test_tokenizer_integration test
      
      * Format source codes
      
      * Add `copied from` comment to CodeGenTokenizerFast
      
      * Add doc of create_token_type_ids_from_sequences
      
      * Make return_token_type_ids False by default
      
      * Make test_tokenizer_integration as slow test
      
      * Add return_token_type_ids to tokenizer init arg
      
      * Add test for tokenizer's init return_token_type_ids
      
      * Format source codes
      8d6b5096
    • Younes Belkada's avatar
      FIX: Fix push important models CI (#30291) · 812a5de2
      Younes Belkada authored
      Update push-important-models.yml
      812a5de2