"docs/vscode:/vscode.git/clone" did not exist on "c73ee1333dc4dc63a71cb6180d0f35fdf4b44958"
- 18 Apr, 2024 1 commit
-
-
Lysandre authored
-
- 17 Apr, 2024 1 commit
-
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 15 Apr, 2024 1 commit
-
-
JINO ROHIT authored
-
- 10 Apr, 2024 1 commit
-
-
Pavel Iakubovskii authored
* Remove unused augmentation * Fix pad_if_smaller() and remove unused augmentation * Add indentation * Fix requirements * Update dataset use instructions * Replace transforms with albumentations * Replace identity transform with None * Fixing formatting * Fixed comment place
-
- 09 Apr, 2024 1 commit
-
-
NielsRogge authored
* Undo * Use tokenizer * Undo data collator
-
- 08 Apr, 2024 2 commits
-
-
JINO ROHIT authored
-
Haz Sameen Shahgir authored
updated examples/pytorch/language-modeling scripts and requirements.txt to require datasets>=2.14.0 (#30120) updated requirements.txt and require_version() calls in examples/pytorch/language-modeling to require datasets>=2.14.0
-
- 05 Apr, 2024 1 commit
-
-
NielsRogge authored
* Add image processor to trainer * Replace tokenizer=image_processor everywhere
-
- 02 Apr, 2024 1 commit
-
-
Mario 艩a拧ko authored
-
- 30 Mar, 2024 1 commit
-
-
Jacky Lee authored
* improve: error message for best model metric * update: raise warning instead of error
-
- 21 Mar, 2024 1 commit
-
-
Jacky Lee authored
feat: add support for torch_dtype Co-authored-by:Jacky Lee <jackylee328@gmail.com>
-
- 20 Mar, 2024 1 commit
-
-
Arthur Zucker authored
-
- 15 Mar, 2024 1 commit
-
-
Quentin Lhoest authored
* Update run_glue.py * Update run_glue.py * Update run_glue_no_trainer.py
-
- 12 Mar, 2024 2 commits
-
-
Joao Gante authored
check max_position_embeddings
-
Hilco van der Wilk authored
* Update legacy Repository usage in `examples/pytorch/text-classification/run_glue_no_trainer.py` Marked for deprecation here https://huggingface.co/docs/huggingface_hub/guides/upload#legacy-upload-files-with-git-lfs * Fix import order * Replace all example usage of deprecated Repository * Fix remaining repo call and rename args variable * Revert removing creation of gitignore files and don't change research examples
-
- 11 Mar, 2024 2 commits
-
-
Yitong Huang authored
* add USE_TORCH_XLA env * rename torch_tpu to torch_xla * better is_torch_xla_available; fix some fsdp and performance issues * fix format * fix bug when pjrt_device is cpu * fix bug * fix the deprecation handling --------- Co-authored-by:
anw90 <ang868@gmail.com> Co-authored-by:
wangang.wa <wangang.wa@alibaba-inc.com>
-
Tanay Mehta authored
* add: initial script to train clm fim * fix: if training model from scratch, new tokens will be added and embeddings resized * fix: fixed attention_mask errors when generating FIM data * fix: file formatted using black * add: run_fim_no_trainer.py and fixed some comments in run_fim.py * add: added fim examples to the README.md and ran code fixup * fix: little bug in both fim training scripts * fix: remove comment from notebook and added a note on fim related params * fix: minor typo in README * add: suggested minor changes to README and run_fim.py * add: gradient_accumulation_steps and gradient_checkpointing args * add: improved model embedding resizing * add: pad_to_multiple_of and attn_implementation params * add: requested minor changes * add: deepspeed zero compatibility * add: resize embeddings layer with zero3 support for fim model initialization
-
- 21 Feb, 2024 1 commit
-
-
Arthur Zucker authored
-
- 19 Feb, 2024 2 commits
-
-
Arthur authored
* change version * nuke * this doesn't make sense * update some requirements.py * revert + no main * nits * change cache number * more pin * revert --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Jay Zhou authored
-
- 16 Feb, 2024 1 commit
-
-
Lysandre Debut authored
* Script & Manual edition * Update
-
- 12 Feb, 2024 2 commits
-
-
Klaus Hipp authored
Add language identifiers to code blocks
-
Alexey Fadeev authored
Updated datasets requirements. Need a package version >= 2.14.0
-
- 07 Feb, 2024 1 commit
-
-
Yih-Dar authored
* fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 02 Feb, 2024 1 commit
-
-
Klaus Hipp authored
* Fix typos and grammar mistakes in docs and examples * Fix typos in docstrings and comments * Fix spelling of `tokenizer` in model tests * Remove erroneous spaces in decorators * Remove extra spaces in Markdown link texts
-
- 01 Feb, 2024 1 commit
-
-
zspo authored
Co-authored-by:p_spozzhang <p_spozzhang@tencent.com>
-
- 30 Jan, 2024 1 commit
-
-
Matt authored
* Pin torch to <2.2.0 * Pin torchvision and torchaudio as well * Playing around with versions to see if this helps * twiddle something to restart the CI * twiddle it back * Try changing the natten version * make fixup * Revert "Try changing the natten version" This reverts commit de0d6592c35dc39ae8b5a616c27285db28262d06. * make fixup * fix fix fix * fix fix fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 29 Jan, 2024 1 commit
-
-
Klaus Hipp authored
-
- 26 Jan, 2024 1 commit
-
-
Steven Liu authored
* change datasets * fix
-
- 22 Jan, 2024 2 commits
-
-
bofeng huang authored
* Fix lr_scheduler * Fix lr scheduler
-
jheitmann authored
-
- 19 Jan, 2024 1 commit
-
-
Amy Roberts authored
-
- 18 Jan, 2024 2 commits
-
-
Yoach Lacombe authored
* add w2v2bert compatibility * Update examples/pytorch/speech-recognition/run_speech_recognition_ctc.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Yoach Lacombe authored
* first commit * correct default value non causal * update config and modeling code * update converting checkpoint * clean modeling and fix tests * make style * add new config parameters to docstring * fix copied from statements * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * make position_embeddings_type docstrings clearer * clean converting script * remove function not used * clean modeling file * apply suggestion for test file + add convert script to not_doctested * modify tests according to review - cleaner logic and more tests * Apply nit suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add checker of valid position embeddings type * instantiate new layer norm layer with the right eps * fix freeze_feature_encoder since it can be None in some cases * add test same output in convert script * restore wav2vec2conformer and add new model * create processor and FE + clean * add new model code * fix convert script and set default config parameters * correct model id paths * make style * make fix-copies and cleaning files * fix copied from statements * complete .md and fixe copies * clean convert script argument defaults * fix config parameters docstrings * fix config docstring * add copied from and enrich FE tests * fix copied from and repo-consistency * add autotokenizer * make test input length shorter and change docstring code * fix docstrings and copied from * add add_adapter to ASR training example * make testing of adapters more robust * adapt to multi adapter layers * refactor input_values->input_features and remove w2v2-bert feature extractor * remove pretraining model * remove depreciated features and useless lines * add copied from and ignore statements to modeling tests * remove pretraining model #2 * change import in convert script * change default in convert script * update readme and remove useless line * Update tests/models/wav2vec2_bert/test_processor_wav2vec2_bert.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * refactor BERT to Bert for consistency * remove useless ignore copy statement * add persistent to buffer in rotary * add eps in LayerNorm init and remove copied from * add adapter activation parameters and add copied from statements * Fix copied statements and add unitest.skip reasons * add copied statement in test_processor * refactor processor * make style * replace numpy random by torch rand * remove expected output CTC * improve converting script with processor class * Apply suggestions from code review Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * remove gumbel class * remove tests related to previously deleted class * Update src/transformers/models/wav2vec2_bert/configuration_wav2vec2_bert.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * correct typos * remove uused parameters * update processor to takes both text and audio * update checkpoints * update expected output and add ctc expected output * add label_attention_mask * replace pt with np in processor tests * fix typo * revert to behaviour with labels_attention_mask --------- Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
- 16 Jan, 2024 1 commit
-
-
regisss authored
* Remove `task` arg in `load_dataset` in image-classification example * Manage case where "train" is not in dataset * Add new args to manage image and label column names * Similar to audio-classification example * Fix README * Update tests
-
- 15 Jan, 2024 2 commits
-
-
Timothy Cronin authored
* improve dev setup comments and hints * fix tests for new dev setup hints
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 12 Jan, 2024 1 commit
-
-
Joao Gante authored
-
- 11 Jan, 2024 1 commit
-
-
Alex Hedges authored
While using `run_clm.py`,[^1] I noticed that some files were being added to my global cache, not the local cache. I set the `cache_dir` parameter for the one call to `evaluate.load()`, which partially solved the problem. I figured that while I was fixing the one script upstream, I might as well fix the problem in all other example scripts that I could. There are still some files being added to my global cache, but this appears to be a bug in `evaluate` itself. This commit at least moves some of the files into the local cache, which is better than before. To create this PR, I made the following regex-based transformation: `evaluate\.load\((.*?)\)` -> `evaluate\.load\($1, cache_dir=model_args.cache_dir\)`. After using that, I manually fixed all modified files with `ruff` serving as useful guidance. During the process, I removed one existing usage of the `cache_dir` parameter in a script that did not have a corresponding `--cache-dir` argument declared. [^1]: I specifically used `pytorch/language-modeling/run_clm.py` from v4.34.1 of the library. For the original code, see the following URL: https://github.com/huggingface/transformers/tree/acc394c4f5e1283c19783581790b3dc3105a3697/examples/pytorch/language-modeling/run_clm.py.
-
- 13 Dec, 2023 1 commit
-
-
Lysandre authored
-