- 21 Nov, 2023 8 commits
-
-
amyeroberts authored
* Enable tracing with DINOv2 model * ABC * Add note to model doc
-
fxmarty authored
* fix various bugs with flash attention * bump * fix test * fix mistral * use skiptest instead of return that may be misleading * fix on review
-
fxmarty authored
* add scheduled ci on amdgpu * fix likely typo * more tests, avoid parallelism * precise comment * fix report channel * trigger docker build on this branch * fix * fix * run rocm scheduled ci * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Leo Tronchon authored
* fix image_attention gate in idefics modeling * update comment * cleaner gating * fix gate condition * create attention gate once * update comment * update doc of cross-attention forward * improve comment * bring back no_images * pass cross_attention_gate similarly to no_images gate * add information on gate shape * fix no_images placement * make tests for gate * take off no_images logic * update test based on comments * raise value error if cross_attention_gate is None * send cross_attention_gate to device * Revert "send cross_attention_gate to device" This reverts commit 054f84228405bfa2e75fecc502f6a96dc83cdc0b. * send cross_attention_gate to device * fix device in test + nit * fill hidden_states with zeros instead of multiplying with the gate * style * Update src/transformers/models/idefics/modeling_idefics.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/idefics/modeling_idefics.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Joao Gante authored
-
NielsRogge authored
* Improve convnext backbone * Fix convnext2
-
Younes Belkada authored
* add support for old GC method * add also disable * up * oops
-
Dave Berenbaum authored
* dvclive callback: warn instead of fail when logging non-scalars * tests: log lr as scalar
-
- 20 Nov, 2023 9 commits
-
-
amyeroberts authored
* Fix torch.fx import issue for torch 1.12 * Fix up * Python verion dependent import * Woops - fix * Fix
-
Yeonwoo Sung authored
Update Korean tutorial for using LLMs, and refactor the nested conditional statements in hr_argparser.py (#27489) docs: Update Korean LLM tutorial to use Mistral-7B, not Llama-v1
-
Dmitrii Mukhutdinov authored
* Enable large-v3 downloading and update language list * Fix type annotation * make fixup * Export Whisper feature extractor * Fix error after extractor loading * Do not use pre-computed mel filters * Save the full preprocessor properly * Update docs * Remove comment Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Add alignment heads consistent with each Whisper version * Remove alignment heads calculation * Save fast tokenizer format as well * Fix slow to fast conversion * Fix bos/eos/pad token IDs in the model config * Add decoder_start_token_id to config --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Said Taghadouini authored
* timm to pytorch conversion for vit model fix * remove unecessary print statments * Detect non-supported ViTs in transformers & better handle id2label mapping * detect non supported hybrid resnet-vit models in conversion script * remove check for overlap between cls token and pos embed
-
Younes Belkada authored
* add fa2 support for from_config * Update test_modeling_common.py
-
Mathias Nielsen authored
* Renamed variable extension to builder_name * If builder name is jsonl change to json to align with load_datasets * Apply suggestions from code review Co-authored-by:
Quentin Lhoest <42851186+lhoestq@users.noreply.github.com> --------- Co-authored-by:
Quentin Lhoest <42851186+lhoestq@users.noreply.github.com>
-
Peter Pan authored
Signed-off-by:Peter Pan <Peter.Pan@daocloud.io>
-
Xabier de Zuazo authored
Add `convert_hf_to_openai.py` script to Whisper documentation resources.
-
Joel Tang authored
* Load idx2sym from pretrained vocab file in Transformer XL When loading vocab file from a pretrained tokenizer for Transformer XL, although the pickled vocabulary file contains a idx2sym key, it isn't loaded, because it is discarded as the empty list already exists as an attribute. Solution is to explicitly take it into account, just like for sym2idx. * ran make style
-
- 19 Nov, 2023 1 commit
-
-
Rafael Padilla authored
Co-authored-by:Rafael Padilla <rafael.padilla@huggingface.co>
-
- 18 Nov, 2023 1 commit
-
-
Omar Sanseviero authored
-
- 17 Nov, 2023 7 commits
-
-
jiaqiw09 authored
* translate deepspeed.md * update
-
V.Prasanna kumar authored
fixed the broken links belogs to dataset library of transformers
-
V.Prasanna kumar authored
-
Joao Gante authored
-
Joao Gante authored
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 16 Nov, 2023 13 commits
-
-
jiaqiw09 authored
* translate * update * update
-
Nathaniel Egwu authored
* Updated albert.md doc for ALBERT model * Update docs/source/en/model_doc/albert.md Fixed Resources heading Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update the ALBERT model doc resources Fixed resource example for fine-tuning the ALBERT sentence-pair classification. Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/model_doc/albert.md Removed resource duplicate Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Updated albert.md doc with reviewed changes * Updated albert.md doc for ALBERT * Update docs/source/en/model_doc/albert.md Removed duplicates from updated docs/source/en/model_doc/albert.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/en/model_doc/albert.md --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Joao Gante authored
-
Arthur authored
* try to stylify using ruff * might need to remove these changes? * use ruf format andruff check * use isinstance instead of type comparision * use # fmt: skip * use # fmt: skip * nits * soem styling changes * update ci job * nits isinstance * more files update * nits * more nits * small nits * check and format * revert wrong changes * actually use formatter instead of checker * nits * well docbuilder is overwriting this commit * revert notebook changes * try to nuke docbuilder * style * fix feature exrtaction test * remve `indent-width = 4` * fixup * more nits * update the ruff version that we use * style * nuke docbuilder styling * leve the print for detected changes * nits * Remove file I/O Co-authored-by:
charliermarsh <charlie.r.marsh@gmail.com> * style * nits * revert notebook changes * Add # fmt skip when possible * Add # fmt skip when possible * Fix * More ` # fmt: skip` usage * More ` # fmt: skip` usage * More ` # fmt: skip` usage * NIts * more fixes * fix tapas * Another way to skip * Recommended way * Fix two more fiels * Remove asynch Remove asynch --------- Co-authored-by:
charliermarsh <charlie.r.marsh@gmail.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Marc Sun authored
add error msg
-
Lucain authored
* Set usedforsecurity=False in hashlib methods (FIPS compliance) * trigger ci * tokenizers version * deps * bump hfh version * let's try this
-
Patrick von Platen authored
* Revert "add attention_mask and position_ids in assisted model (#26892)" This reverts commit 184f60dc. * more debug
-
Matt authored
* Move the TF pin for 2.15 * make fixup
-
Phuc Van Phan authored
-
Arthur authored
add flash attn markers
-
Dean Wyatte authored
support onnx for causal lm sequence classification
-
Hz, Ji authored
* translate model.md to chinese * apply review suggestion Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
- 15 Nov, 2023 1 commit
-
-
Marc Sun authored
* fix * style * add test
-