"vscode:/vscode.git/clone" did not exist on "023f51fe16e34e0ca2b5598791ae508874d5b443"
  1. 02 Apr, 2024 8 commits
    • Minsub Lee (Matt)'s avatar
      Fix `skip_special_tokens` for `Wav2Vec2CTCTokenizer._decode` (#29311) · 15cd6871
      Minsub Lee (Matt) authored
      * Fix skip_special_tokens process for Wav2Vec2CTCTokenizer._decode
      
      * Fix skip_special_tokens for Wav2Vec2CTCTokenizer._decode
      
      * Exclude pad_token filtering since it is used as CTC-blank token
      
      * Add small test for skip_special_tokens
      
      * Update decoding test for added new token
      15cd6871
    • Michael's avatar
    • Yoach Lacombe's avatar
      Add Flash Attention 2 support to Musicgen and Musicgen Melody (#29939) · 0d04b1e2
      Yoach Lacombe authored
      * add FA2 to o.g Musicgen
      
      * make style
      
      * add FA2 support to Musicgen Melody
      
      * add generation FA2 tests to o.g Musicgen
      
      * make style and fix copies
      
      * add Musicgen to FA2 docs + deprecate list
      
      * add sdpa supports to Musicgen's
      
      * make style and fix copies
      
      * refactor attention implementation arguments
      
      * add Copied from to sdpa tests
      
      * add copied form in sdpa tests melody
      
      * add copied for FA2 generation tests
      
      * add FA2 inference copied from
      
      * make style
      0d04b1e2
    • th茅o gigant's avatar
      Adding FlaxNoRepeatNGramLogitsProcessor (#29677) · fed27ffc
      th茅o gigant authored
      * fix issue with logit processor in beam search in Flax
      
      * adding FlaxNoRepeatNGramLogitsProcessor class + unit test
      
      * style correction and code verification
      
      * add FlaxNoRepeatNGramLogitsProcessor to the test_processor_list and test_processor_list_jitted tests
      
      * fix an issue where ngrams are banned only if they appear ==1 time + update description of get_previous_ngrams
      
      * replace non-jit compatible masking of ngrams that are not yet generated with jittable version
      
      * Revert "fix issue with logit processor in beam search in Flax"
      
      This reverts commit 09b70d7e4dc32d0cc4db61af09a835a9cd238b50.
      
      * add FlaxNoRepeatNGramLogitsProcessor to _get_logits_processor
      
      * change the method of casting to boolean of banned tokens indices
      
      * fix code style
      
      * remove some useless operations + significantly faster computation of update indices using jax.lax.fori_loop
      
      * remove useless loop iterations
      
      * set some variables that were calculated and used multiple times
      
      * fix format
      fed27ffc
    • Marc Sun's avatar
      [bnb] Fix bug in `_replace_with_bnb_linear` (#29958) · 33288ff1
      Marc Sun authored
      fix bug
      33288ff1
    • Hovnatan Karapetyan's avatar
      Fix 29807 sinusoidal positional encodings in Flaubert, Informer and XLM (#29904) · 416711c3
      Hovnatan Karapetyan authored
      * Fix sinusoidal_embeddings in FlaubertModel
      
      * Fix for Informer
      
      * Fix for XLM
      
      * Move sinusoidal emb for XLM
      
      * Move sinusoidal emb for Flaubert
      
      * Small cleanup
      
      * Add comments on tests code copied from
      
      * Add with Distilbert->
      416711c3
    • Arthur's avatar
      [`generate`] fix breaking change for patch (#29976) · 83b26dd7
      Arthur authored
      * fix bug and add tests
      
      * nit
      
      * otherway to get the cur len instead of attention mask
      
      * more places where this might have been broken
      
      * nit
      
      * oups
      
      * inputs_embeds vs input_embeds
      
      * test generated outptus
      
      * style
      
      * nit
      
      * fix
      
      * skip failing biogpt
      83b26dd7
    • Steven Liu's avatar
      [docs] Big model loading (#29920) · 096f3046
      Steven Liu authored
      * update
      
      * feedback
      096f3046
  2. 01 Apr, 2024 4 commits
  3. 31 Mar, 2024 1 commit
  4. 30 Mar, 2024 6 commits
  5. 29 Mar, 2024 2 commits
  6. 28 Mar, 2024 18 commits
  7. 27 Mar, 2024 1 commit