1. 11 Oct, 2023 5 commits
    • Patrick von Platen's avatar
      [Assistant Generation] Improve Encoder Decoder (#26701) · da69de17
      Patrick von Platen authored
      * [Assistant Generation] Improve enc dec
      
      * save more
      
      * Fix logit processor checks
      
      * Clean
      
      * make style
      
      * fix deprecation
      
      * fix generation test
      
      * Apply suggestions from code review
      
      * fix biogpt
      
      * make style
      da69de17
    • Yih-Dar's avatar
      `Copied from` for test files (#26713) · 5334796d
      Yih-Dar authored
      
      
      * copied statement for test files
      
      ---------
      Co-authored-by: default avatarydshieh <ydshieh@users.noreply.github.com>
      5334796d
    • Ben Gubler's avatar
      Update docs to explain disabling callbacks using report_to (#26155) · 9f406392
      Ben Gubler authored
      * feat: update callback doc to explain disabling callbacks using report_to
      
      * docs: update report_to docstring
      9f406392
    • Billy Bradley's avatar
      In assisted decoding, pass model_kwargs to model's forward call (fix... · dcc49d8a
      Billy Bradley authored
      In assisted decoding, pass model_kwargs to model's forward call (fix prepare_input_for_generation in all models) (#25242)
      
      * In assisted decoding, pass model_kwargs to model's forward call
      
      Previously, assisted decoding would ignore any additional kwargs
      that it doesn't explicitly handle. This was inconsistent with other
      generation methods, which pass the model_kwargs through
      prepare_inputs_for_generation and forward the returned dict to the
      model's forward call.
      
      The prepare_inputs_for_generation method needs to be amended in all
      models, as previously it only kept the last input ID when a past_key_values
      was passed.
      
      * Improve variable names in _extend_attention_mask
      
      * Refactor extending token_type_ids into a function
      
      * Replace deepcopy with copy to optimize performance
      
      * Update new persimmon model with llama changes for assisted generation
      
      * Update new mistral model for assisted generation with prepare_inputs_for_generation
      
      * Update position_ids creation in falcon prepare_inputs_for_generation to support assisted generation
      dcc49d8a
    • Thien Tran's avatar
      Make Whisper Encoder's sinusoidal PE non-trainable by default (#26032) · 1e3c9dda
      Thien Tran authored
      
      
      * set encoder's PE as non-trainable
      
      * freeze flax
      
      * init sinusoids
      
      * add test for non-trainable embed positions
      
      * simplify TF encoder embed_pos
      
      * revert tf
      
      * clean up
      
      * add sinusoidal init for jax
      
      * make consistent sinusoidal function
      
      * fix dtype
      
      * add default dtype
      
      * use numpy for sinusoids. fix jax
      
      * add sinusoid init for TF
      
      * fix
      
      * use custom embedding
      
      * use specialized init for each impl
      
      * fix sinusoids init. add test for pytorch
      
      * fix TF dtype
      
      * simplify sinusoid init for flax and tf
      
      * add tests for TF
      
      * change default dtype to float32
      
      * add sinusoid test for flax
      
      * Update src/transformers/models/whisper/modeling_flax_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * Update src/transformers/models/whisper/modeling_tf_whisper.py
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      
      * move sinusoidal init to _init_weights
      
      ---------
      Co-authored-by: default avatarsanchit-gandhi <sanchit@huggingface.co>
      Co-authored-by: default avatarSanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
      1e3c9dda
  2. 10 Oct, 2023 6 commits
  3. 09 Oct, 2023 10 commits
  4. 06 Oct, 2023 11 commits
  5. 05 Oct, 2023 8 commits