- 01 Feb, 2024 1 commit
-
-
zspo authored
Co-authored-by:p_spozzhang <p_spozzhang@tencent.com>
-
- 19 Jan, 2024 1 commit
-
-
Amy Roberts authored
-
- 11 Jan, 2024 1 commit
-
-
Alex Hedges authored
While using `run_clm.py`,[^1] I noticed that some files were being added to my global cache, not the local cache. I set the `cache_dir` parameter for the one call to `evaluate.load()`, which partially solved the problem. I figured that while I was fixing the one script upstream, I might as well fix the problem in all other example scripts that I could. There are still some files being added to my global cache, but this appears to be a bug in `evaluate` itself. This commit at least moves some of the files into the local cache, which is better than before. To create this PR, I made the following regex-based transformation: `evaluate\.load\((.*?)\)` -> `evaluate\.load\($1, cache_dir=model_args.cache_dir\)`. After using that, I manually fixed all modified files with `ruff` serving as useful guidance. During the process, I removed one existing usage of the `cache_dir` parameter in a script that did not have a corresponding `--cache-dir` argument declared. [^1]: I specifically used `pytorch/language-modeling/run_clm.py` from v4.34.1 of the library. For the original code, see the following URL: https://github.com/huggingface/transformers/tree/acc394c4f5e1283c19783581790b3dc3105a3697/examples/pytorch/language-modeling/run_clm.py.
-
- 13 Dec, 2023 1 commit
-
-
Lysandre authored
-
- 11 Dec, 2023 1 commit
-
-
Adam Louly authored
* fix no sequence length models error * block size check --------- Co-authored-by: Adam Louly <adamlouly@microsoft.com@orttrainingdev9.d32nl1ml4oruzj4qz3bqlggovf.px.internal.cloudapp.net>
-
- 17 Nov, 2023 1 commit
-
-
V.Prasanna kumar authored
fixed the broken links belogs to dataset library of transformers
-
- 15 Nov, 2023 1 commit
-
-
Adam Louly authored
fix max pos issue Co-authored-by: Adam Louly <adamlouly@microsoft.com@orttrainingdev9.d32nl1ml4oruzj4qz3bqlggovf.px.internal.cloudapp.net>
-
- 02 Nov, 2023 1 commit
-
-
Lysandre authored
-
- 31 Oct, 2023 1 commit
-
-
Dong-geon Lee authored
-
- 27 Oct, 2023 1 commit
-
-
Lucain authored
-
- 12 Oct, 2023 1 commit
-
-
Tom Aarsen authored
Add missing spaces in adjacent strings
-
- 04 Oct, 2023 1 commit
-
-
Phuc Van Phan authored
* refactor: change default block_size * fix: return tf to origin * fix: change files to origin * rebase * rebase * rebase * rebase * rebase * rebase * rebase * rebase * refactor: add min block_size to files * reformat: add min block_size for run_clm tf
-
- 03 Oct, 2023 1 commit
-
-
Lysandre authored
-
- 11 Sep, 2023 2 commits
-
-
Phuc Van Phan authored
* docs: add space to docs * docs: remove reduntant space
-
Phuc Van Phan authored
-
- 04 Sep, 2023 1 commit
-
-
Lysandre authored
-
- 21 Aug, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 07 Aug, 2023 1 commit
-
-
Jackmin801 authored
* pytorch examples * pytorch mim no trainer * cookiecutter * flax examples * missed line in pytorch run_glue * tensorflow examples * tensorflow run_clip * tensorflow run_mlm * tensorflow run_ner * tensorflow run_clm * pytorch example from_configs * pytorch no trainer examples * Revert "tensorflow run_clip" This reverts commit 261f86ac1f1c9e05dd3fd0291e1a1f8e573781d5. * fix: duplicated argument
-
- 02 Aug, 2023 1 commit
-
-
Yih-Dar authored
* fix * fix * fix * fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 28 Jul, 2023 1 commit
-
-
Yih-Dar authored
* pytorch examples * tensorflow examples * flax examples --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 20 Jul, 2023 1 commit
-
-
Zach Mueller authored
Change logic
-
- 17 Jul, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 07 Jun, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 18 May, 2023 1 commit
-
-
Boda Sadallah authored
* fix bug in group_texts function, that was inserting short batches * fully exclude short batches and return empty dict instead * fix style
-
- 09 May, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 13 Apr, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 22 Mar, 2023 1 commit
-
-
Wang, Yi authored
* add low_cpu_mem_usage option in run_clm.py example which will benefit LLM loading Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * update all the example and README under language-modeling Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> --------- Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com>
-
- 14 Mar, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 22 Feb, 2023 2 commits
-
-
Sylvain Gugger authored
* Respect documentation on passive log level * Fix test and set log level in examples * Add doc
-
Aaron Gokaslan authored
-
- 06 Feb, 2023 2 commits
-
-
Sylvain Gugger authored
* Result of black 23.1 * Update target to Python 3.7 * Switch flake8 to ruff * Configure isort * Configure isort * Apply isort with line limit * Put the right black version * adapt black in check copies * Fix copies
-
Stas Bekman authored
-
- 31 Jan, 2023 1 commit
-
-
Quentin Lhoest authored
* simplify column_names in run_clm * simplify column_names in run_mlm * minor
-
- 30 Jan, 2023 1 commit
-
-
Stas Bekman authored
* [run_clm example] add streaming dataset support * unrefactor kwargs * fix * fix * require datasets>=2.0.0 * port to mlm
-
- 23 Jan, 2023 1 commit
-
-
Sylvain Gugger authored
-
- 03 Jan, 2023 1 commit
-
-
Wang, Yi authored
* [run_clm example] add torch_dtype option for model load. for BLOOM 175B model. peak memory will reduce about 350G for inference. the weight of BLOOM in model hub is bfloat16 Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * add other type in option * fix style Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com>
-
- 01 Dec, 2022 1 commit
-
-
Sylvain Gugger authored
-
- 03 Nov, 2022 1 commit
-
-
Sylvain Gugger authored
* Only resize embeddings when necessary * Add comment
-
- 01 Nov, 2022 1 commit
-
-
Sylvain Gugger authored
-
- 10 Oct, 2022 1 commit
-
-
Lysandre authored
-