1. 26 Jan, 2024 6 commits
  2. 25 Jan, 2024 3 commits
  3. 24 Jan, 2024 4 commits
    • nakranivaibhav's avatar
      Improved type hinting for all attention parameters (#28479) · 5d29530e
      nakranivaibhav authored
      * Changed type hinting for all attention inputs to 'Optional[Tuple[torch.FloatTensor,...]] = None'
      
      * Fixed the ruff formatting issue
      
      * fixed type hinting for all hidden_states to 'Optional[Tuple[torch.FloatTensor, ...]] = None'
      
      * Changed type hinting in these 12 scripts modeling_dpr.py,modeling_nat.py,idefics/vision.py,modeling_tf_dpr.py,modeling_luke.py,modeling_swin.py,modeling_tf_swin.py,modeling_blip.py,modeling_tf_blip.py,modeling_donut_swin.py,modeling_dinat.py,modeling_swinv2.py
      
      * test fail update
      
      * fixed type hinting for these 15 scripts modeling_xlnet.py,modeling_tf_xlnet.py,modeling_led.py,modeling_tf_led.py,modleing_rwkv.py,modeling_dpt.py,modeling_tf_cvt.py,modeling_clip.py,modeling_flax_clip.py,modeling_tf_clip.py,modeling_longformer.py,modeling_tf_longformer.py,modeling_siglip.py,modeling_clap.py,modeling_git.py
      
      * Changed type hinting in these 12 scripts modeling_dpr.py,modeling_nat.py,idefics/vision.py,modeling_tf_dpr.py,modeling_luke.py,modeling_swin.py,modeling_tf_swin.py,modeling_blip.py,modeling_tf_blip.py,modeling_donut_swin.py,modeling_dinat.py,modeling_swinv2.py
      
      * test fail update
      
      * Removed the myvenv file
      
      * Fixed type hinting for these 8 scripts modeling_tvlt.py,modeling_sam.py,modeling_tf_sam.py,modeling_tvp.py,modeling_rag.py,modeling_tf_rag.py,modeling_tf_xlm.py,modeling_xlm.py
      5d29530e
    • amyeroberts's avatar
      Add back in generation types (#28681) · bb6aa8bc
      amyeroberts authored
      bb6aa8bc
    • jeffhataws's avatar
      Use save_safetensor to disable safe serialization for XLA (#28669) · 0549000c
      jeffhataws authored
      * Use save_safetensor to disable safe serialization for XLA
      
      https://github.com/huggingface/transformers/issues/28438
      
      * Style fixup
      0549000c
    • Khai Mai's avatar
      Exclude the load balancing loss of padding tokens in Mixtral-8x7B (#28517) · c5c69096
      Khai Mai authored
      * fix the function load_balancing_loss_func in Mixtral_Moe to include attention_mask
      
      * format code using black and ruff
      
      * skip computing mask if attention_mask=None
      
      * add tests for load balancing loss Mixtral-Moe
      
      * fix assert loss is different in mixtral_test
      
      * fix pad_leng
      
      * use assertNotAlmostEqual and print to debug
      
      * remove print for debug
      
      * minor updates
      
      * reduce rtol and atol
      c5c69096
  4. 23 Jan, 2024 10 commits
  5. 22 Jan, 2024 5 commits
  6. 21 Jan, 2024 1 commit
  7. 19 Jan, 2024 11 commits