1. 10 Apr, 2024 2 commits
  2. 09 Apr, 2024 4 commits
    • Marc Sun's avatar
      Fix quantization tests (#29914) · 58a939c6
      Marc Sun authored
      * revert back to torch 2.1.1
      
      * run test
      
      * switch to torch 2.2.1
      
      * udapte dockerfile
      
      * fix awq tests
      
      * fix test
      
      * run quanto tests
      
      * update tests
      
      * split quantization tests
      
      * fix
      
      * fix again
      
      * final fix
      
      * fix report artifact
      
      * build docker again
      
      * Revert "build docker again"
      
      This reverts commit 399a5f9d9308da071d79034f238c719de0f3532e.
      
      * debug
      
      * revert
      
      * style
      
      * new notification system
      
      * testing notfication
      
      * rebuild docker
      
      * fix_prev_ci_results
      
      * typo
      
      * remove warning
      
      * fix typo
      
      * fix artifact name
      
      * debug
      
      * issue fixed
      
      * debug again
      
      * fix
      
      * fix time
      
      * test notif with faling test
      
      * typo
      
      * issues again
      
      * final fix ?
      
      * run all quantization tests again
      
      * remove name to clear space
      
      * revert modfiication done on workflow
      
      * fix
      
      * build docker
      
      * build only quant docker
      
      * fix quantization ci
      
      * fix
      
      * fix report
      
      * better quantization_matrix
      
      * add print
      
      * revert to the basic one
      58a939c6
    • Yih-Dar's avatar
      Fix slow tests for important models to be compatible with A10 runners (#29905) · 08a194fc
      Yih-Dar authored
      
      
      * fix mistral and mixtral
      
      * add pdb
      
      * fix mixtral tesst
      
      * fix
      
      * fix mistral ?
      
      * add fix gemma
      
      * fix mistral
      
      * fix
      
      * test
      
      * anoter test
      
      * fix
      
      * fix
      
      * fix mistral tests
      
      * fix them again
      
      * final fixes for mistral
      
      * fix padding right
      
      * fix whipser fa2
      
      * fix
      
      * fix
      
      * fix gemma
      
      * test
      
      * fix llama
      
      * fix
      
      * fix
      
      * fix llama gemma
      
      * add class attribute
      
      * fix CI
      
      * clarify whisper
      
      * compute_capability
      
      * rename names in some comments
      
      * Add   # fmt: skip
      
      * make style
      
      * Update tests/models/mistral/test_modeling_mistral.py
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * update
      
      * update
      
      ---------
      Co-authored-by: default avatarYounes Belkada <younesbelkada@gmail.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      08a194fc
    • Matt's avatar
      Revert workaround for TF safetensors loading (#30128) · ec59a421
      Matt authored
      * See if we can get tests to pass with the fixed weights
      
      * See if we can get tests to pass with the fixed weights
      
      * Replace the revisions now that we don't need them anymore
      ec59a421
    • Sourab Mangrulkar's avatar
      Fix failing DeepSpeed model zoo tests (#30112) · 4e3490f7
      Sourab Mangrulkar authored
      * fix sequence length errors
      
      * fix label column name error for vit
      
      * fix the lm_head embedding!=linear layer mismatches for Seq2Seq models
      4e3490f7
  3. 08 Apr, 2024 5 commits
  4. 05 Apr, 2024 5 commits
  5. 04 Apr, 2024 1 commit
    • byi8220's avatar
      [`ProcessingIdefics`] Attention mask bug with padding (#29449) · 75b76a5e
      byi8220 authored
      * Defaulted IdeficsProcessor padding to 'longest', removed manual padding
      
      * make fixup
      
      * Defaulted processor call to padding=False
      
      * Add padding to processor call in IdeficsModelIntegrationTest as well
      
      * Defaulted IdeficsProcessor padding to 'longest', removed manual padding
      
      * make fixup
      
      * Defaulted processor call to padding=False
      
      * Add padding to processor call in IdeficsModelIntegrationTest as well
      
      * redefaulted padding=longest again
      
      * fixup/doc
      75b76a5e
  6. 03 Apr, 2024 5 commits
  7. 02 Apr, 2024 6 commits
    • Nicolas Patry's avatar
      Hard error when ignoring tensors. (#27484) (#29906) · 9b0a8ea7
      Nicolas Patry authored
      
      
      * Hard error when ignoring tensors. (#27484)
      
      * [WIP] Hard error when ignoring tensors.
      
      * Better selection/error when saving a checkpoint.
      
      - Find all names we should normally drop (those are in the transformers
        config)
      - Find all disjoint tensors (for those we can safely trigger a copy to
        get rid of the sharing before saving)
      - Clone those disjoint tensors getting rid of the issue
      - Find all identical names (those should be declared in the config
        but we try to find them all anyway.)
      - For all identical names:
        - If they are in the config, just ignore them everything is fine
        - If they are not, warn about them.
      - For all remainder tensors which are shared yet neither identical NOR
        disjoint. raise a hard error.
      
      * Adding a failing test on `main` that passes here.
      
      * We don't need to keep the subfolder logic in this test.
      
      * Apply suggestions from code review
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Add small tests.
      
      * Dead variable.
      
      * Fixup.
      
      * Fixing tied_Weights_keys on generic models.
      
      * Fixup + T5 encoder/decoder tying (with different layers)
      
      * Code quality.
      
      * Dynamic member.
      
      * trigger
      
      * Fixing encoder name for other types of encoder/decoder combos.
      
      * Fix scoping.
      
      * Update .github/workflows/self-scheduled.yml
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      
      * Fixing the tied_weights after the call.
      
      ---------
      Co-authored-by: default avatarArthur <48595927+ArthurZucker@users.noreply.github.com>
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      9b0a8ea7
    • Minsub Lee (Matt)'s avatar
      Fix `skip_special_tokens` for `Wav2Vec2CTCTokenizer._decode` (#29311) · 15cd6871
      Minsub Lee (Matt) authored
      * Fix skip_special_tokens process for Wav2Vec2CTCTokenizer._decode
      
      * Fix skip_special_tokens for Wav2Vec2CTCTokenizer._decode
      
      * Exclude pad_token filtering since it is used as CTC-blank token
      
      * Add small test for skip_special_tokens
      
      * Update decoding test for added new token
      15cd6871
    • Yoach Lacombe's avatar
      Add Flash Attention 2 support to Musicgen and Musicgen Melody (#29939) · 0d04b1e2
      Yoach Lacombe authored
      * add FA2 to o.g Musicgen
      
      * make style
      
      * add FA2 support to Musicgen Melody
      
      * add generation FA2 tests to o.g Musicgen
      
      * make style and fix copies
      
      * add Musicgen to FA2 docs + deprecate list
      
      * add sdpa supports to Musicgen's
      
      * make style and fix copies
      
      * refactor attention implementation arguments
      
      * add Copied from to sdpa tests
      
      * add copied form in sdpa tests melody
      
      * add copied for FA2 generation tests
      
      * add FA2 inference copied from
      
      * make style
      0d04b1e2
    • théo gigant's avatar
      Adding FlaxNoRepeatNGramLogitsProcessor (#29677) · fed27ffc
      théo gigant authored
      * fix issue with logit processor in beam search in Flax
      
      * adding FlaxNoRepeatNGramLogitsProcessor class + unit test
      
      * style correction and code verification
      
      * add FlaxNoRepeatNGramLogitsProcessor to the test_processor_list and test_processor_list_jitted tests
      
      * fix an issue where ngrams are banned only if they appear ==1 time + update description of get_previous_ngrams
      
      * replace non-jit compatible masking of ngrams that are not yet generated with jittable version
      
      * Revert "fix issue with logit processor in beam search in Flax"
      
      This reverts commit 09b70d7e4dc32d0cc4db61af09a835a9cd238b50.
      
      * add FlaxNoRepeatNGramLogitsProcessor to _get_logits_processor
      
      * change the method of casting to boolean of banned tokens indices
      
      * fix code style
      
      * remove some useless operations + significantly faster computation of update indices using jax.lax.fori_loop
      
      * remove useless loop iterations
      
      * set some variables that were calculated and used multiple times
      
      * fix format
      fed27ffc
    • Hovnatan Karapetyan's avatar
      Fix 29807 sinusoidal positional encodings in Flaubert, Informer and XLM (#29904) · 416711c3
      Hovnatan Karapetyan authored
      * Fix sinusoidal_embeddings in FlaubertModel
      
      * Fix for Informer
      
      * Fix for XLM
      
      * Move sinusoidal emb for XLM
      
      * Move sinusoidal emb for Flaubert
      
      * Small cleanup
      
      * Add comments on tests code copied from
      
      * Add with Distilbert->
      416711c3
    • Arthur's avatar
      [`generate`] fix breaking change for patch (#29976) · 83b26dd7
      Arthur authored
      * fix bug and add tests
      
      * nit
      
      * otherway to get the cur len instead of attention mask
      
      * more places where this might have been broken
      
      * nit
      
      * oups
      
      * inputs_embeds vs input_embeds
      
      * test generated outptus
      
      * style
      
      * nit
      
      * fix
      
      * skip failing biogpt
      83b26dd7
  8. 01 Apr, 2024 4 commits
  9. 31 Mar, 2024 1 commit
  10. 29 Mar, 2024 1 commit
  11. 28 Mar, 2024 6 commits