- 13 Sep, 2023 11 commits
-
-
Joao Gante authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
-
김준재_T3056 authored
-
Younes Belkada authored
* Final fix RWMV 4bit * fixup * add a test * add more clarifications
-
Vaibhav Srivastav authored
update names mapping for spectrogram and waveform models
-
Juarez Bochi authored
Add missing space in utils.py Warning now reads as "... to control thegeneration length. We ..."
-
Younes Belkada authored
* fix 4bit `num_parameters` * stronger check
-
amyeroberts authored
Fix docstring typo
-
Sourab Mangrulkar authored
* fix the deepspeed tests * resolve comment
-
Sourab Mangrulkar authored
-
- 12 Sep, 2023 12 commits
-
-
Tanay Mehta authored
* add: check to remove metaspace from marian tokenizer * fix: metaspace character being removed from everywhere * fix: remove redundant check at top * add: test for marian tokenizer decode fix * fix: simplified the test
-
Joao Gante authored
-
Phuc Van Phan authored
-
Wang, Yi authored
* enable optuna multi-objectives feature Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * update hpo doc * update docstring Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * extend direction to List[str] type Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * Update src/transformers/integrations/integration_utils.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
MinJae Kang authored
* docs: ko-contributing.md * feat: chatGPT draft * feat: manual edits * feat: change linked document * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> * fix: resolve suggestion Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * fix: resolve suggestion Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * fix: resolve suggestion Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com> * fix: resolve suggestion * fix: resolve suggestion * feat: delete file to resolve error --------- Co-authored-by:
Haewon Kim <ehdvkf02@naver.com> Co-authored-by:
SeongWooChoi <46990061+nuatmochoi@users.noreply.github.com>
-
Maria Khalusova authored
* tts guide updates with a pipeline * Apply suggestions from code review Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> * Update docs/source/en/tasks/text-to-speech.md Co-authored-by:
Vaibhav Srivastav <vaibhavs10@gmail.com> --------- Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> Co-authored-by:
Vaibhav Srivastav <vaibhavs10@gmail.com>
-
MinJae Kang authored
* docs: ko-llama2.md * feat: chatGPT draft and manul edits * feat: added inline TOC * fix: inline TOC * fix: resolve suggestions Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * fix: resolve suggestion Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> * fix: resolve suggestion Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com> --------- Co-authored-by:
Jungnerd <46880056+jungnerd@users.noreply.github.com>
-
pokjay authored
* Fix issues in test_exponential_decay_length_penalty Fix tests which were broken and add validation of negative scores. Current test didn't take into account that ExponentialDecayLengthPenalty updates the score inplace, resulting in updates to base tested Tensor. In addition, the gt assert had empty Tensors due to indexing along the batch dimension. Test is currently expected to fail to show ExponentialDecayLengthPenalty issues with negative scores * Fix ExponentialDecayLengthPenalty negative logits issue In cases where the scores are negative, ExponentialDecayLengthPenalty decreases the score of eos_token_id instead of increasing it. To fix this issue we compute the penalty of the absolute value and add it to the original score. * Add examples for ExponentialDecayLengthPenalty * Fix styling issue in ExponentialDecayLengthPenalty doc * Apply suggestions from code review Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Style and quality fix * Fix example outputs --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
larekrow authored
-
Joao Gante authored
-
Younes Belkada authored
import tensorflow inside relevant methods in trainer_utils
-
Arthur authored
* intiial commit * updates * nits * update conversion script * update conversion script * use path to load * add tips etc * some modeling logic * modeling update * more nits * nits * normal layer norm * update config and doc * nits * update doc remove unused * update * fix inits and stuff * fixup * revert wrong changes * updates * more nits * add default config values to the configuration file * fixup happy * update * 2 tests left * update readmes * more nits * slow test and more documentation * update readme * fix licences * styling * use fast if possible when saving tokenizer * remove todo * remove tokenization tests * small last nits * Apply suggestions from code review Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com> * nits to skip the timout doctest * fix integration test * fix test * update eos token * update to allow fast tokenization * styling * fix codeLlama as well for the update post processor * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add more copied from statements * update * doc passes doctest * remove `# final layer norm?` * change docstring prompot * update * Update README.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * don't doctest the conversion script as it requires more packages * don't init a model in the config * oups * fix doctest --------- Co-authored-by:
Matt <Rocketknight1@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 11 Sep, 2023 4 commits
-
-
Phuc Van Phan authored
* docs: add space to docs * docs: remove reduntant space
-
Patrick von Platen authored
* improve import time * Update src/transformers/integrations/__init__.py * sort import
-
Phuc Van Phan authored
-
Hang authored
only main process should call _save when deepspeed zero3
-
- 09 Sep, 2023 1 commit
-
-
Arthur authored
* skip failing tests until #26054 is merged * fixup
-
- 08 Sep, 2023 5 commits
-
-
Arthur authored
* fix `set_infilling_processor` to properly reset * Add docstring! * fixups * more details in the docuemtation about the tokenization * styl;e
-
Harheem Kim authored
* docs: ko-llama.md * fix: chatgpt draft * feat: manual edits * fix: resolve suggestions
-
Angela Yi authored
* Ignore warning if tracing with dynamo * fix import error * separate to function * add test
-
Thien Tran authored
* add missing doc for activation dropout * fix doc for SEW-D dropout * deprecate hidden_dropout for SEW-D
-
Alexander Krauck authored
This commit corrects the dropout implementation in Graphormer, aligning it with the original implementation and improving performance. Specifically: 1. The `attention_dropout` variable, intended for use in GraphormerMultiheadAttention, was defined but not used. This has been corrected to use `attention_dropout` instead of the regular `dropout`. 2. The `activation_dropout` for the activations in the feed-forward layers was missing. Instead, the regular `dropout` was used. This commit adds `activation_dropout` to the feed-forward layers. These changes ensure the dropout implementation matches the original Graphormer and delivers empirically better performance.
-
- 07 Sep, 2023 7 commits
-
-
dumpmemory authored
* fix loss inconsistent after resume #25340 * fix typo * clean code * reformatted code * adjust code according to comments * adjust check_dataloader_randomsampler location * return sampler only * handle sampler is None * Update src/transformers/trainer_pt_utils.py thanks @amyeroberts Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
MyungHa Kwon authored
fix typo
-
raghavanone authored
* Fix vilt config init parameter to match the ones in documentation * Fix the documentation
-
Muskan Kumar authored
* Added HerBERT to README.md * Update README.md to contain HerBERT (#26016) * Resolved #26016: Updated READMEs and index.md to contain Herbert Updated READMEs and ran make fix-copies
-
Sanchit Gandhi authored
* fix tokenizer * make bs even * fix multi gpu test * style * model forward * fix torch import * revert tok pin
-
CokeDong authored
* Add tgs metrics * bugfix and black formatting * workaround for tokens counting * formating and bugfix * Fix * Add opt-in for tgs metrics * make style and fix error * Fix doc * fix docbuild * hf-doc-build * fix * test * Update src/transformers/training_args.py renaming Co-authored-by:
Zach Mueller <muellerzr@gmail.com> * Update src/transformers/training_args.py renaming Co-authored-by:
Zach Mueller <muellerzr@gmail.com> * Fix some symbol * test * Update src/transformers/trainer_utils.py match nameing patterns Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/trainer.py nice Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Fix reviews * Fix * Fix black --------- Co-authored-by:
Zach Mueller <muellerzr@gmail.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-