- 02 Oct, 2023 7 commits
-
-
Younes Belkada authored
* fix bnb test with code revision * fix test * Apply suggestions from code review * Update src/transformers/models/auto/auto_factory.py * Update src/transformers/models/auto/auto_factory.py * Update src/transformers/models/auto/auto_factory.py
-
Younes Belkada authored
* try * nit * nits
-
HelgeS authored
-
marcmk6 authored
* fix issue of canine forward requires input_ids anyway The `forward` requires `input_ids` for deriving other variables in all cases. Change this to use the given one between `input_ids` and `inputs_embeds` * fix canine forward The current `forward` requires (the shape of) `input_ids` for deriving other variables whenever `input_ids` or `inputs_embeds` is provided. Change this to use the given one instead of `input_ids` all the time. * fix format * fix format
-
Jan Philipp Harries authored
fix requests connection error Co-authored-by:Jan Philipp Harries <jphme@users.noreply.github.com>
-
Florian Seiler authored
* Fix num_heads in _upad_input The variable num_key_value_heads has falsely been named num_heads, which led to reshaping the query_layer using the wrong attention head count. (It would have been enough to use the correct variable self.num_heads instead of num_heads, but I renamed num_heads to num_key_value_heads for clarity) * fixed copies using make fix-copies and ran make fixup --------- Co-authored-by:fseiler <f.seiler@jerocom.de>
-
Lysandre Debut authored
* Revert "Falcon: fix revision propagation (#26006)" This reverts commit 118c676ef3124423e5d062b665f05cde55bc9a90. * Revert "Put Falcon back (#25960)" This reverts commit 22a69f1d.
-
- 29 Sep, 2023 6 commits
-
-
Sanchit Gandhi authored
* improve docs/errors * why whisper * Update docs/source/en/pipeline_tutorial.md Co-authored-by:
Lysandre Debut <hi@lysand.re> * specify pt only --------- Co-authored-by:
Lysandre Debut <hi@lysand.re>
-
Sanchit Gandhi authored
* from seq2seq speech * [Flax] Example script for speech seq2seq * tests and fixes * make style * fix: label padding tokens * fix: label padding tokens over list * update ln names for Whisper * try datasets iter loader * create readme and append results * style * make style * adjust lr * use pt dataloader * make fast * pin gen max len * finish * add pt to requirements for test * fix pt -> torch * add accelerate
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
skip Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Maria Khalusova authored
* navigation improvement between text generation pipelines and text generation docs * make style
-
Steven Liu authored
update
-
- 28 Sep, 2023 10 commits
-
-
Sanchit Gandhi authored
make decoding faster
-
Amelie Schreiber authored
* Fixed in-place operation error in EsmEmbeddings * Fixed in-place operation error in EsmEmbeddings again --------- Co-authored-by:Schreiber-Finance <amelie.schreiber.finance@gmail.com>
-
Marc Sun authored
* fix_mbart_tied_weights * add test
-
fleance authored
Do not warn about unexpected decoder weights when loading T5EncoderModel and LongT5EncoderModel (#26211) Ignore decoder weights when using T5EncoderModel and LongT5EncoderModel Both T5EncoderModel and LongT5EncoderModel do not have any decoder layers, so loading a pretrained model checkpoint such as t5-small will give warnings about keys found in the model checkpoint that are not in the model itself. To prevent this log warning, r"decoder" has been added to _keys_to_ignore_on_load_unexpected for both T5EncoderModel and LongT5EncoderModel
-
Younes Belkada authored
[`PEFT`] introducing `adapter_kwargs` for loading adapters from different Hub location (`subfolder`, `revision`) than the base model (#26270) * make use of adapter_revision * v1 adapter kwargs * fix CI * fix CI * fix CI * fixup * add BC * Update src/transformers/integrations/peft.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * fixup * change it to error * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py * fixup * change * Update src/transformers/integrations/peft.py --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Fakhir Ali authored
* [VITS] Fix speaker_embed device mismatch - pass device arg to speaker_id tensor * [VITS] put speaker_embed on device when int * [VITS] device=self.device instead of self.embed_speaker.weight.device * [VITS] make tensor directly on device using torch.full()
-
Tanishq Abraham authored
* change mention of decoder_input_ids to input_ids and same with decoder_input_embeds * Style --------- Co-authored-by:Lysandre <lysandre@huggingface.co>
-
Phuc Van Phan authored
* docs: change assert to raise and some small docs * docs: add rule and some document * fix: fix bug * fix: fix bug * chorse: revert logging * chorse: revert
-
Yih-Dar authored
* fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Norm Inui authored
* optimize layoutv2, v3 for VRAM saving * reformat codes --------- Co-authored-by:NormXU <xunuo@datagrand.com>
-
- 27 Sep, 2023 12 commits
-
-
Wonhyeong Seo authored
* dos: ko: perf_train_gpu_many.mdx * feat: chatgpt draft * fix: manual edits * fix: resolve suggestions Change description Follow the glossary Fix discrepancies Co-Authored-By:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> Co-Authored-By:
이서정 <97655267+sjlee-wise@users.noreply.github.com> Co-Authored-By:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by:
Hyunho <105839613+hyunhp@users.noreply.github.com> Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> Co-authored-by:
이서정 <97655267+sjlee-wise@users.noreply.github.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Wonhyeong Seo authored
* docs:ko:Debugging.md * feat: chatgpt draft * fix: resolve suggestions Co-Authored-By:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-Authored-By:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by:
Jang KyuJin <106062329+kj021@users.noreply.github.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Florian Zimmermeister authored
* initial * toctree * add tf model * run scripts * peft * llm and agents * Update docs/source/de/peft.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/peft.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/peft.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/run_scripts.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/run_scripts.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/transformers_agents.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * Update docs/source/de/transformers_agents.md Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Yih-Dar authored
* update * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Lysandre Debut authored
* Fix doctest * Adding modeling also for now
-
Chris Bamford authored
* [Mistral] Mistral-7B-v0.1 support * fixing names * slightly longer test * fixups * not_doctested * wrongly formatted references * make fixuped --------- Co-authored-by:
Timothee Lacroix <t@eugen.ai> Co-authored-by:
timlacroix <t@mistral.ai>
-
Younes Belkada authored
* fix PEFT multi adapters support * refactor a bit * save pretrained + BC + added tests * Update src/transformers/integrations/peft.py Co-authored-by:
Benjamin Bossan <BenjaminBossan@users.noreply.github.com> * add more tests * add suggestion * final changes * adapt a bit * fixup * Update src/transformers/integrations/peft.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * adapt from suggestions --------- Co-authored-by:
Benjamin Bossan <BenjaminBossan@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
statelesshz authored
Co-authored-by:statelesshz <jihuazhong1@huawei.com>
-
Younes Belkada authored
* add use_cache tests for FA * fixup
-
Uri Alon authored
* Fixing tokenizer when tokenizers is not installed * Adding __repr__ function and repr=True in dataclass * Revert "Adding __repr__ function and repr=True in dataclass" This reverts commit 18839505d1cada3170ed623744d3e75008a18bdc.
-
Nour Eddine ZEKAOUI authored
-
Shauray Singh authored
* fix * fixup * tests * fixup
-
- 26 Sep, 2023 5 commits
-
-
Nathan Lambert authored
add rmsprop
-
Matt authored
* Add config.bias to LLaMA to allow InternLM models to be ported as LLaMA checkpoints * Rename bias -> attention_bias and add docstring
-
Hugo Laurençon authored
Fix deepspeed issue with Idefics
-
sanjeevk-os authored
-
titi authored
-