- 28 Jun, 2024 6 commits
-
-
Jade Choghari authored
* Add French translation of run scripts tutorial * Update docs/source/fr/run_scripts_fr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update docs/source/fr/run_scripts_fr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update docs/source/fr/run_scripts_fr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update docs/source/fr/run_scripts_fr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update docs/source/fr/run_scripts_fr.md Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Jade Choghari <chogharijade@icloud.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Arthur authored
* softcapping * soft cap before the mask * style * ... * super nit
-
Sangbum Daniel Choi authored
* add gather_use_object arguments * fix name and pass the CI test for Seq2SeqTrainer * make style * make it to functools * fix typo * add accelerate version: * adding warning * Update src/transformers/trainer.py Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com> * make style * Update src/transformers/training_args.py * check function move to initial part * add test for eval_use_gather_object --------- Co-authored-by:
Marc Sun <57196510+SunMarc@users.noreply.github.com>
-
Jacky Lee authored
* fix: use return_dict parameter * fix: type checks * fix: unused imports * update: one-line if else * remove: recursive check
-
hoshi-hiyouga authored
Update modeling_gemma2.py Co-authored-by:Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Wing Lian authored
* don't zero out the attention_mask when using sliding window with flash attention * chore: lint
-
- 27 Jun, 2024 12 commits
-
-
Sanchit Gandhi authored
* fix gemma2 * handle in generate
-
Steven Liu authored
quick usage to top
-
Billy Cao authored
-
Arthur authored
* nit * toctree issue * protect gemma2 tests as well * sdpa supported
-
Lysandre authored
-
Arthur authored
* inital commit * Add doc * protect? * fixup stuffs * update tests * fix build documentation * mmmmmmm config attributes * style * nit * uodate * nit * Fix docs * protect some stuff --------- Co-authored-by:Lysandre <lysandre@huggingface.co>
-
Raushan Turganbay authored
remove
-
Sangbum Daniel Choi authored
* change anchor_image_size None for compatibility * make fix-copies
-
Billy Cao authored
* Allow dtype str for torch_dtype in from_pretrained * Update docstring * Add tests for str torch_dtype
-
Arthur authored
* fix and simplify the script! * add co-author --------- Co-authored-by:crackalamoo <crackalamoo@users.noreply.github.com>
-
Merve Noyan authored
* fixed models * format with bumped ruff version on my local * fix copies * add tracing checks * format * Update src/transformers/utils/generic.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * format * style fix * Update modeling_mobilevit.py * add docstring and change name * Update __init__.py * Update __init__.py --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Raushan Turganbay authored
* fix * better
-
- 26 Jun, 2024 12 commits
-
-
amyeroberts authored
* Skip tests properly * [test_all] * Add 'reason' as kwarg for skipTest * [test_all] Fix up * [test_all]
-
Billy Cao authored
* Fix dtype casting in modeling_swin2sr to allow non-FP32 inference * Fix formattting * Fix for swinv2 too * Update src/transformers/models/swin2sr/modeling_swin2sr.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/models/swinv2/modeling_swinv2.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Add FP16 tests for swin2sr and swinv2 * [run_slow] swin2sr, swinv2 * [run_slow] swin2sr, swinv2 --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Joao Gante authored
-
Pablo Montalvo authored
* fix extended attention mask * add slow test for detection instance * [run-slow]paligemma
-
Raushan Turganbay authored
* squash into single commit * run diff once more * docstring * tests * minor chnages and ready to go * Update src/transformers/models/llava_next_video/processing_llava_next_video.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update tests/models/vipllava/test_modeling_vipllava.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * [run-slow] llava-next-video * [run-slow] llava-next-video * [run-slow] llava_next_video * fix two tests * fix slow tests * remove logit checks due to numeric errors * run test once more * [run-slow] llava_next_video * final try to pass the test * [run-slow] llava_next_video * [run-slow] llava_next_video * [run-slow] llava_next_video * style * fix * style --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Pavel Iakubovskii authored
* [run_slow] rt_detr * Fix positional embeddings and anchors dtypes * [run slow] rt_detr * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Fixup --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Younes Belkada authored
* fix llama fsdp * fixup * adding FSDP tests for CPU offloading * fixes * fix tests * fix tests * add it for mixtral * propagate the changes on other models * Update src/transformers/models/phi/modeling_phi.py * Delete utils/testing_scripts/fsdp_cpu_offloading.py Remove script - FSDP + CPU offloading it tested in the test suite * Delete utils/testing_scripts/dummy_fsdp_config.yml * Update + add cache_positions docstring --------- Co-authored-by:amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Pavel Iakubovskii authored
Update code snippet
-
Marc Sun authored
-
Anton Vlasjuk authored
* starting support for sdpa in `gptneox` models * small comment on tests * fix dropout * documentation and style * clarify concrete paths for reference * generalise attn projections and rope application added head mask check to sdpa mask creation handle sdpa memory backend bug via own version flag * update docs and style * move dtype casting outside of general attn_projection_and_rope function fix flash_attn_2 stuff * more generic attn warning if output_attns or head_mask * simplify head mask check by moving head mask creation to a later point * remove copied llama artifact * remove padding_mask from attention function signature * removing unnecessary comments, only "save" attn implementation once * [run_slow] gpt_neox
-
Vladimir Iashin authored
removes unnecessary second projection call
-
Saurav Maheshkar authored
docs: move translations to i18n
-
- 25 Jun, 2024 6 commits
-
-
amyeroberts authored
* Add ViTImageProcessor to tests * Correct data format * Review comments
-
Pablo Montalvo authored
improve error message for mismatched code blocks
-
Locke authored
preprocessing_num_workers option to speedup preprocess
-
Raushan Turganbay authored
* squash in single commit * add docs * dummy obj * more changes in diff converter * tiny fix * make docs happy * skip test * repo consistency tests * update docstring * style * fix tests * change diff imports * [run-slow] instructblipvideo * [run-slow] instructblipvideo * fix tests and remove logit check * [run-slow] instructblipvideo
-
jiqing-feng authored
* fix output data type of image classification * add tests for low-precision pipeline * add bf16 pipeline tests * fix bf16 tests * Update tests/pipelines/test_pipelines_image_classification.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * fix import * fix import torch * fix style --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Raushan Turganbay authored
* device-map siglip * move split modules to PretrainedSigLip
-
- 24 Jun, 2024 4 commits
-
-
Ren茅 Gentzen authored
* Contrained numpy to <2.0 * Updated dependency_versions_table --------- Co-authored-by:Ren茅 Gentzen <rene.gentzen@mittelstand.ai>
-
amyeroberts authored
-
Quentin Gallou茅dec authored
-
Matt authored
* Add Jinja as a requirement with the right version cutoff * Correct package name!
-