- 04 Oct, 2023 3 commits
-
-
Sylvain Gugger authored
* Fix number of minimal calls to the Hub with peft integration * Alternate design * And this way? * Revert * Nits to fix * Add util * Print when changes are made * Add list to ignore * Add more rules * Manual fixes * deal with kwargs * deal with enum defaults * avoid many digits for floats * Manual fixes * Fix regex * Fix regex * Auto fix * Style * Apply script * Add ignored list * Add check that templates are filled * Adding to CI checks * Add back semi-fix * Ignore more objects * More auto-fixes * Ignore missing objects * Remove temp semi-fix * Fixes * Update src/transformers/models/pvt/configuration_pvt.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update utils/check_docstrings.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/utils/quantization_config.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Deal with float defaults * Fix small defaults * Address review comment * Treat * Post-rebase cleanup * Address review comment * Update src/transformers/models/deprecated/mctct/configuration_mctct.py Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr> * Address review comment --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr>
-
Bharat Ramanathan authored
-
statelesshz authored
-
- 03 Oct, 2023 9 commits
-
-
Sanchit Gandhi authored
* [Whisper] Allow basic text normalization * up * style copies
-
Lysandre authored
-
Arthur authored
* remove unprotected import to PIL * cleanup --------- Co-authored-by:Lysandre <lysandre@huggingface.co>
-
Younes Belkada authored
* fix issues with PEFT * logger warning futurewarning issues * fixup * adapt from suggestions * oops * rm test
-
Younes Belkada authored
* add FA-2 support for mistral * fixup * add sliding windows * fixing few nits * v1 slicing cache - logits do not match * add comment * fix bugs * more mem efficient * add warning once * add warning once * oops * fixup * more comments * copy * add safety checker * fixup * Update src/transformers/models/mistral/modeling_mistral.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * copied from * up * raise when padding side is right * fixup * add doc + few minor changes * fixup --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Arthur authored
* fix stripping * nits * fix another test * styling * fix? * update * revert bad merge * found the bug * YES SIR * is that change really required? * make fast even faster * re order functions
-
Srijan Sahay Srivastava authored
* [Doctest] Add configuration_encoder_decoder.py Added configuration_encoder_decoder.py to utils/documentation_tests.txt for doctest * Revert "[Doctest] Add configuration_encoder_decoder.py" This reverts commit bd653535a4356dc3c9f43e65883819079a2053b0. * [Doctest] Add configuration_encoder_decoder.py add configuration_encoder_decoder.py to utils/documentation_tests.txt * [Doctest] Add configuration_encoder_decoder.py add configuration_encoder_decoder.py to utils/documentation_tests.txt * [Doctest] Add configuration_encoder_decoder.py add configuration_encoder_decoder.py to utils/documentation_tests.txt * changed as per request * fixed line 46
-
Nathan Cahill authored
* add tokenizer kwarg inputs * Adding tokenizer_kwargs to _sanitize_parameters * Add truncation=True example to tests * Update test_pipelines_fill_mask.py * Update test_pipelines_fill_mask.py * make fix-copies and make style * Update fill_mask.py Replace single tick with double * make fix-copies * Style --------- Co-authored-by:Lysandre <lysandre@huggingface.co>
-
Patrick von Platen authored
[Logging] Change warning to info
-
- 02 Oct, 2023 11 commits
-
-
Arthur authored
* add build_inputs_with_special_tokens to LlamaFast * fixup * Update src/transformers/models/llama/tokenization_llama_fast.py
-
Arthur authored
* fix encoding when the fill token is None * add tests and edge cases * fiuxp * Update tests/models/code_llama/test_tokenization_code_llama.py
-
Adithya Hegde Kota authored
* [Doctest] Add configuration_roformer.py * [Doctest] Add configuration_roformer.py * [Doctest] Add configuration_roformer.py * [Doctest] Add configuration_roformer.py * Removed documentation_test.txt * Removed configuration_roformer.py * Update not_doctested.txt
-
Arthur authored
* fix stripping * remove some warnings and update some warnings * revert changes for other PR
-
Younes Belkada authored
Update modeling_utils.py
-
Arthur authored
* fix wav2vec2 * nit * stash * one more file to update * fix byt5 * vocab size is 256, don't change that! * use other revision * test persimon in smaller size * style * tests * nits * update add tokens from pretrained * test tokenization * nits * potential fnet fix? * more nits * nits * correct test * assert close * udpate * ouch * fix it * some more nits * FINALLU * use `adept` checkpoints * more adept checkpoints * that was invlved!
-
Younes Belkada authored
* try * nit * nits
-
marcmk6 authored
* fix issue of canine forward requires input_ids anyway The `forward` requires `input_ids` for deriving other variables in all cases. Change this to use the given one between `input_ids` and `inputs_embeds` * fix canine forward The current `forward` requires (the shape of) `input_ids` for deriving other variables whenever `input_ids` or `inputs_embeds` is provided. Change this to use the given one instead of `input_ids` all the time. * fix format * fix format
-
Jan Philipp Harries authored
fix requests connection error Co-authored-by:Jan Philipp Harries <jphme@users.noreply.github.com>
-
Florian Seiler authored
* Fix num_heads in _upad_input The variable num_key_value_heads has falsely been named num_heads, which led to reshaping the query_layer using the wrong attention head count. (It would have been enough to use the correct variable self.num_heads instead of num_heads, but I renamed num_heads to num_key_value_heads for clarity) * fixed copies using make fix-copies and ran make fixup --------- Co-authored-by:fseiler <f.seiler@jerocom.de>
-
Lysandre Debut authored
* Revert "Falcon: fix revision propagation (#26006)" This reverts commit 118c676ef3124423e5d062b665f05cde55bc9a90. * Revert "Put Falcon back (#25960)" This reverts commit 22a69f1d.
-
- 29 Sep, 2023 2 commits
-
-
Sanchit Gandhi authored
* improve docs/errors * why whisper * Update docs/source/en/pipeline_tutorial.md Co-authored-by:
Lysandre Debut <hi@lysand.re> * specify pt only --------- Co-authored-by:
Lysandre Debut <hi@lysand.re>
-
Maria Khalusova authored
* navigation improvement between text generation pipelines and text generation docs * make style
-
- 28 Sep, 2023 9 commits
-
-
Sanchit Gandhi authored
make decoding faster
-
Amelie Schreiber authored
* Fixed in-place operation error in EsmEmbeddings * Fixed in-place operation error in EsmEmbeddings again --------- Co-authored-by:Schreiber-Finance <amelie.schreiber.finance@gmail.com>
-
Marc Sun authored
* fix_mbart_tied_weights * add test
-
fleance authored
Do not warn about unexpected decoder weights when loading T5EncoderModel and LongT5EncoderModel (#26211) Ignore decoder weights when using T5EncoderModel and LongT5EncoderModel Both T5EncoderModel and LongT5EncoderModel do not have any decoder layers, so loading a pretrained model checkpoint such as t5-small will give warnings about keys found in the model checkpoint that are not in the model itself. To prevent this log warning, r"decoder" has been added to _keys_to_ignore_on_load_unexpected for both T5EncoderModel and LongT5EncoderModel
-
Younes Belkada authored
[`PEFT`]聽introducing `adapter_kwargs` for loading adapters from different Hub location (`subfolder`, `revision`) than the base model (#26270) * make use of adapter_revision * v1 adapter kwargs * fix CI * fix CI * fix CI * fixup * add BC * Update src/transformers/integrations/peft.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * fixup * change it to error * Update src/transformers/modeling_utils.py * Update src/transformers/modeling_utils.py * fixup * change * Update src/transformers/integrations/peft.py --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Fakhir Ali authored
* [VITS] Fix speaker_embed device mismatch - pass device arg to speaker_id tensor * [VITS] put speaker_embed on device when int * [VITS] device=self.device instead of self.embed_speaker.weight.device * [VITS] make tensor directly on device using torch.full()
-
Tanishq Abraham authored
* change mention of decoder_input_ids to input_ids and same with decoder_input_embeds * Style --------- Co-authored-by:Lysandre <lysandre@huggingface.co>
-
Yih-Dar authored
* fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Norm Inui authored
* optimize layoutv2, v3 for VRAM saving * reformat codes --------- Co-authored-by:NormXU <xunuo@datagrand.com>
-
- 27 Sep, 2023 5 commits
-
-
Chris Bamford authored
* [Mistral] Mistral-7B-v0.1 support * fixing names * slightly longer test * fixups * not_doctested * wrongly formatted references * make fixuped --------- Co-authored-by:
Timothee Lacroix <t@eugen.ai> Co-authored-by:
timlacroix <t@mistral.ai>
-
Younes Belkada authored
* fix PEFT multi adapters support * refactor a bit * save pretrained + BC + added tests * Update src/transformers/integrations/peft.py Co-authored-by:
Benjamin Bossan <BenjaminBossan@users.noreply.github.com> * add more tests * add suggestion * final changes * adapt a bit * fixup * Update src/transformers/integrations/peft.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * adapt from suggestions --------- Co-authored-by:
Benjamin Bossan <BenjaminBossan@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
statelesshz authored
Co-authored-by:statelesshz <jihuazhong1@huawei.com>
-
Uri Alon authored
* Fixing tokenizer when tokenizers is not installed * Adding __repr__ function and repr=True in dataclass * Revert "Adding __repr__ function and repr=True in dataclass" This reverts commit 18839505d1cada3170ed623744d3e75008a18bdc.
-
Shauray Singh authored
* fix * fixup * tests * fixup
-
- 26 Sep, 2023 1 commit
-
-
Nathan Lambert authored
add rmsprop
-