- 09 Oct, 2023 1 commit
-
-
NielsRogge authored
* Convert checkpoints * Update doc test * Address comment
-
- 06 Oct, 2023 11 commits
-
-
Jabasukuriputo Wang authored
-
Yih-Dar authored
example fix docstring Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Arthur authored
* make sure eos and bos are properly handled for fast tokenizer * fix code llama as well * nits * fix the conversion script as well * fix failing test
-
statelesshz authored
* remove SharedDDP as it was drepracated * apply review suggestion * make style * Oops,forgot to remove the compute_loss context manager in Seq2SeqTrainer. * remove the unnecessary conditional statement * keep the logic of IPEX * clean code * mix precision setup & make fixup --------- Co-authored-by:statelesshz <jihuazhong1@huawei.com>
-
Yih-Dar authored
* fix * fix * Fix * Fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
rui-ren authored
-
Matt authored
-
fxmarty authored
* remove unnecessary unsqueeze-squeeze in llama * correct other models * fix * revert gpt_neox_japanese * fix copie * fix test
-
Tianqi Liu authored
* Update tokenization_code_llama_fast.py * Update test_tokenization_code_llama.py * Update test_tokenization_code_llama.py
-
Towdo authored
-
Ramiro Leal-Cavazos authored
* Remove unnecessary `view` of `position_ids` in `modeling_llama` When `position_ids` is `None`, its value is generated using `torch.arange`, which creates a tensor of size `(seq_length + past_key_values_length) - past_key_values_length = seq_length`. The tensor is then unsqueezed, resulting in a tensor of shape `(1, seq_length)`. This means that the last `view` to a tensor of shape `(-1, seq_length)` is a no-op. This commit removes the unnecessary view. * Remove no-op `view` of `position_ids` in rest of transformer models
-
- 05 Oct, 2023 11 commits
-
-
Yih-Dar authored
Fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Maria Khalusova authored
* build the table in index.md with links to the model_doc * removed list generation on index.md * fixed missing models * make style
-
Yih-Dar authored
Fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
eajechiloae authored
don't close clearml task if it was created externally
-
Marvin Gabler authored
* feat: close #26566, changed model & config files to accept arbitary in and out channels * updated docstrings * fix: linter error * fix: update Copy docstrings * fix: linter update * fix: rename num_channels_in to num_channels to prevent breaking changes * fix: make num_channels_out None per default * Update src/transformers/models/swin2sr/configuration_swin2sr.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * fix: update tests to include num_channels_out * fix:linter * fix: remove normalization with precomputed rgb values when #input_channels!=#output_channels --------- Co-authored-by:
marvingabler <marvingabler@outlook.de> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Younes Belkada authored
* fix silent bug `keep_in_fp32` modules * final fix * added a common test. * Trigger CI * revert
-
Charles Bensimon authored
* Make `ModelOutput` serializable Original PR from diffusers : https://github.com/huggingface/diffusers/pull/5234 * Black
-
Yih-Dar authored
* fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yun Dai authored
* Set `presents=None` when `use_cache` is set to False for activation ckpt * Update modeling_falcon.py * fix black
-
Arthur authored
* Faster rotary embedding for GPTNeoX * there might be un-necessary moves from device * fixup * fix dtype issue * add copied from statements * fox copies * oupsy * add copied from Llama for scaled ones as well * fixup * fix * fix copies
-
Arthur authored
fix
-
- 04 Oct, 2023 14 commits
-
-
Yeyang authored
* translate installation to zh * fix translation typo
-
Sanchit Gandhi authored
* fix wav2vec2 doctest * suggestion * fix * final fix * revert since we need AddedTokens
-
Galland authored
-
Arthur authored
skip flaky
-
Soyoung Yoon authored
Fix bug in convert_t5x_checkpoint_to_pytorch.py
-
Matt authored
-
dg845 authored
Add # Copied from statements to audio feature extractors that use the floats_list function.
-
Sanchit Gandhi authored
* fix copies * fix missing docstring * make style * oops
-
Phuc Van Phan authored
* refactor: change default block_size * fix: return tf to origin * fix: change files to origin * rebase * rebase * rebase * rebase * rebase * rebase * rebase * rebase * refactor: add min block_size to files * reformat: add min block_size for run_clm tf
-
Matt authored
* Add add_generation_prompt argument to apply_chat_template * Add add_generation_prompt argument to apply_chat_template and update default templates * Fix typo * Add generation prompts section to chat templating guide * Add generation prompts section to chat templating guide * Minor style fix
-
Sylvain Gugger authored
* Fix number of minimal calls to the Hub with peft integration * Alternate design * And this way? * Revert * Nits to fix * Add util * Print when changes are made * Add list to ignore * Add more rules * Manual fixes * deal with kwargs * deal with enum defaults * avoid many digits for floats * Manual fixes * Fix regex * Fix regex * Auto fix * Style * Apply script * Add ignored list * Add check that templates are filled * Adding to CI checks * Add back semi-fix * Ignore more objects * More auto-fixes * Ignore missing objects * Remove temp semi-fix * Fixes * Update src/transformers/models/pvt/configuration_pvt.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update utils/check_docstrings.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/utils/quantization_config.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Deal with float defaults * Fix small defaults * Address review comment * Treat * Post-rebase cleanup * Address review comment * Update src/transformers/models/deprecated/mctct/configuration_mctct.py Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr> * Address review comment --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre.debut@reseau.eseo.fr>
-
Bharat Ramanathan authored
-
statelesshz authored
-
dependabot[bot] authored
Bump pillow in /examples/research_projects/decision_transformer Bumps [pillow](https://github.com/python-pillow/Pillow) from 9.3.0 to 10.0.1. - [Release notes](https://github.com/python-pillow/Pillow/releases) - [Changelog](https://github.com/python-pillow/Pillow/blob/main/CHANGES.rst) - [Commits](https://github.com/python-pillow/Pillow/compare/9.3.0...10.0.1 ) --- updated-dependencies: - dependency-name: pillow dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
- 03 Oct, 2023 3 commits
-
-
김준재_T3056 authored
-
Lysandre Debut authored
* Skip tests temporarily * style * Add additional test
-
Jungnerd authored
* docs: ko: sementic_segmentation.md * feat: manual draft * fix: manual edits * fix: resolve suggestions Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> * fix: resolve suggestions Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * fix: edit the title --------- Co-authored-by:
Wonhyeong Seo <wonhseo@kakao.com> Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-