- 24 Jul, 2023 6 commits
-
-
Iskren Ivov Chernev authored
* Better handling missing SYS in llama conversation tokenizer The existing code failed to add SYS if the conversation has history without SYS, but did modify the passed conversation as it did. Rearrange the code so modification to the conversation object are taken into account for token id generation. * Fix formatting with black * Avoid one-liners * Also fix fast tokenizer * Drop List decl
-
Lucain authored
* Support GatedRepoError + use raise from * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Use token instead of use_auth_token in error messages --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Maria Khalusova authored
* first pass at the single gpu doc * overview: improved clarity and navigation * WIP * updated intro and deepspeed sections * improved torch.compile section * more improvements * minor improvements * make style * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * feedback addressed * mdx -> md * link fix * feedback addressed --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Bharat Ramanathan authored
fix: store training args to wandb config without sanitization. Allows resuming runs by reusing the wandb config. Co-authored-by:Bharat Ramanathan <ramanathan.parameshwaran@gohuddl.com>
-
Arthur authored
set default logger
-
Stas Bekman authored
* [check_config_docstrings.py] improve diagnostics * style * rephrase * fix
-
- 21 Jul, 2023 16 commits
-
-
Wonhyeong Seo authored
fix: update ko/serialization.md * chatgpt draft
-
Sylvain Gugger authored
-
Ivan Sorokin authored
* improve from_pretrained for zero3 multi gpus mode * Add check if torch.distributed.is_initialized * Revert torch.distributed --------- Co-authored-by:Stas Bekman <stas@stason.org>
-
Arthur authored
remove persistent tensor
-
Younes Belkada authored
add simple check for bnb
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Avoid importing all models when instantiating a pipeline * Remove sums that don't work
-
Sylvain Gugger authored
-
Arthur authored
* pad token should be None by default * fix tests * nits
-
Joya Chen authored
* Update tokenization_llama.py * Update tokenization_llama_fast.py * Update src/transformers/models/llama/tokenization_llama_fast.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/llama/tokenization_llama.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/llama/tokenization_llama.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/llama/tokenization_llama_fast.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Sourab Mangrulkar authored
* fix fsdp prepare to remove the warnings and fix excess memory usage * Update training_args.py * parity for FSDP+XLA * Update trainer.py
-
Wonhyeong Seo authored
* fix: english/korean quicktour.md * fix: resolve suggestions Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Kihoon Son <75935546+kihoon71@users.noreply.github.com> * fix: follow glossary * 韺岇澑韸滊嫕 -> 氙胳劯臁办爼 --------- Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Kihoon Son <75935546+kihoon71@users.noreply.github.com>
-
Jim Allanson authored
* fix: cast input pixels to appropriate dtype for image_to_text tasks * fix: add casting to pixel inputs of additional models after running copy checks
-
Sourab Mangrulkar authored
* fix fsdp load * Update trainer.py * remove saving duplicate state_dict
-
- 20 Jul, 2023 15 commits
-
-
Apoorv Khandelwal authored
* [trainer] fallback for deepspeed param count * [trainer] more readable numel count
-
Benjamin Badger authored
Co-authored-by:Joao Gante <joaofranciscocardosogante@gmail.com>
-
Zach Mueller authored
Change logic
-
Younes Belkada authored
add GC support for RWKV
-
dependabot[bot] authored
Bump aiohttp in /examples/research_projects/decision_transformer Bumps [aiohttp](https://github.com/aio-libs/aiohttp) from 3.8.1 to 3.8.5. - [Release notes](https://github.com/aio-libs/aiohttp/releases) - [Changelog](https://github.com/aio-libs/aiohttp/blob/v3.8.5/CHANGES.rst) - [Commits](https://github.com/aio-libs/aiohttp/compare/v3.8.1...v3.8.5 ) --- updated-dependencies: - dependency-name: aiohttp dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Shauray Singh authored
* testing * example script * fix typehinting * some tests * make test * optional update * Union of arguments * does this fix the issue * remove reports * set default to False * documentation change * None support * does not need None * Fix typing annotations for FSDP and DeepSpeed in TrainingArguments (#24549) * Fix typing annotations for FSDP and DeepSpeed in TrainingArguments * Change dict to Dict * Revert "Fix typing annotations for FSDP and DeepSpeed in TrainingArguments" (#24574) Revert "Fix typing annotations for FSDP and DeepSpeed in TrainingArguments (#24549)" This reverts commit c5e29d43 . * Fix typing annotations for FSDP and DeepSpeed in TrainingArguments (#24549) * Fix typing annotations for FSDP and DeepSpeed in TrainingArguments * Change dict to Dict * merge * hacky fix * fixup --------- Co-authored-by:
Max Ryabinin <mryabinin0@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Shauray Singh authored
* make docs * fixup * resolved * remove debugs * Revert "fixup" This reverts commit 5e0f636aae0bf8707bc8bdaa6a9427fbf66834ed. * prev (ignore) * fixup broke some files * remove files * reverting modeling_reformer * lang fix
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Premtim Sa authored
Fixing small typo: kwrags -> kwargs
-
dependabot[bot] authored
Bump pygments in /examples/research_projects/decision_transformer Bumps [pygments](https://github.com/pygments/pygments) from 2.11.2 to 2.15.0. - [Release notes](https://github.com/pygments/pygments/releases) - [Changelog](https://github.com/pygments/pygments/blob/master/CHANGES) - [Commits](https://github.com/pygments/pygments/compare/2.11.2...2.15.0 ) --- updated-dependencies: - dependency-name: pygments dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Joao Gante authored
-
statelesshz authored
* replace no_cuda with use_cpu in test_pytorch_examples * remove codes that never be used * fix style
-
Tom Aarsen authored
* Resolve typo in check_repo.py * Specify encoding when opening modeling files * Deprecate the OpenLlama architecture * Add disclaimer pointing to Llama I'm open to different wordings here * Match the capitalisation of LLaMA
-
ranchlai authored
* Add text classification example * set the problem type and finetuning task * ruff reformated * fix bug for unseting label_to_id for regression * update README.md * fixed finetuning task * update comment * check if label exists in feature before removing * add useful logging
-
Jungnerd authored
* docs: ko: `document_question_answering.md` * fix: resolve suggestions Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> * fix: resolve suggestions Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com> Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> --------- Co-authored-by:
Sohyun Sim <96299403+sim-so@users.noreply.github.com> Co-authored-by:
Hyeonseo Yun <0525yhs@gmail.com>
-
- 19 Jul, 2023 3 commits
-
-
Stas Bekman authored
[doc] image_processing_vilt.py wrong default
-
Younes Belkada authored
* add possibility to disable TP * fixup * adapt from offline discussions
-
Travis Cline authored
Update llama2.md Fix typos in the llama2 model doc
-