- 15 Jan, 2024 12 commits
-
-
Timothy Cronin authored
* improve dev setup comments and hints * fix tests for new dev setup hints
-
Boris Dayma authored
-
Joao Gante authored
-
Matt authored
* Add a use_safetensors arg to TFPreTrainedModel.from_pretrained() * One more catch! * One more one more catch
-
Rishit Ratna authored
-
Marc Sun authored
* fix test * reduce length * smaller model
-
thedamnedrhino authored
* added args to the pipeline * added test * more sensical tests * fixup * docs * typo ; * docs * made changes to support named args * fixed test * docs update * styles * docs * docs
-
yuanwu2017 authored
* Add the XPU check for pipeline mode When setting xpu device for pipeline, It needs to use is_torch_xpu_available to load ipex and determine whether the device is available. Signed-off-by:
yuanwu <yuan.wu@intel.com> * Don't move model to device when hf_device_map isn't None 1. Don't move model to device when hf_device_map is not None 2. The device string maybe includes the device index, so use 'in'instead of equal Signed-off-by:
yuanwu <yuan.wu@intel.com> * Raise the error when xpu is not available Signed-off-by:
yuanwu <yuan.wu@intel.com> * Update src/transformers/pipelines/base.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/pipelines/base.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Modify the error message Signed-off-by:
yuanwu <yuan.wu@intel.com> * Change message format. Signed-off-by:
yuanwu <yuan.wu@intel.com> --------- Signed-off-by:
yuanwu <yuan.wu@intel.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Younes Belkada authored
* v1 tags * remove unneeded conversion * v2 * rm unneeded warning * add more utility methods * Update src/transformers/utils/hub.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/utils/hub.py Co-authored-by:
Lucain <lucainp@gmail.com> * Update src/transformers/utils/hub.py Co-authored-by:
Lucain <lucainp@gmail.com> * more enhancements * oops * merge tags * clean up * revert unneeded change * add extensive docs * more docs * more kwargs * add test * oops * fix test * Update src/transformers/modeling_utils.py Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update src/transformers/utils/hub.py Co-authored-by:
Lucain <lucainp@gmail.com> * Update src/transformers/modeling_utils.py * Update src/transformers/trainer.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update src/transformers/modeling_utils.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add more conditions * more logic --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> Co-authored-by:
Lucain <lucainp@gmail.com> Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com>
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Tom Aarsen authored
Update warning, a word was missing
-
Francisco Kurucz authored
Fix URL to Ai Sweden Models reference and model loading
-
- 13 Jan, 2024 2 commits
-
-
Joao Gante authored
* fix candidate device * this line shouldn't have been in
-
Apoorv Saxena authored
* MVP * fix ci * more ci * remove redundant kwarg * added and wired up PromptLookupCandidateGenerator * rebased with main, working * removed print * style fixes * fix test * fixed tests * added test for prompt lookup decoding * fixed circleci * fixed test issue * Update src/transformers/generation/candidate_generator.py Co-authored-by:
Joao Gante <joaofranciscocardosogante@gmail.com> * Update src/transformers/generation/candidate_generator.py Co-authored-by:
Joao Gante <joaofranciscocardosogante@gmail.com> * Update src/transformers/generation/candidate_generator.py * Update src/transformers/generation/candidate_generator.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Joao Gante <joao@huggingface.co> Co-authored-by:
Joao Gante <joaofranciscocardosogante@gmail.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
- 12 Jan, 2024 12 commits
-
-
Siddartha Naidu authored
-
Matt authored
* Fix TF Regnet docstring * Fix TF Regnet docstring * Make a change to the PyTorch Regnet too to make sure the CI is checking it * Add skips for TFRegnet * Update error message for docstring checker
-
Joao Gante authored
-
Joao Gante authored
-
Joao Gante authored
-
Joao Gante authored
-
sungho-ham authored
Fix xlnet torch.ones usage Co-authored-by:sungho-ham <sungho.ham@linecorp.com>
-
dependabot[bot] authored
Bump jinja2 in /examples/research_projects/decision_transformer Bumps [jinja2](https://github.com/pallets/jinja) from 2.11.3 to 3.1.3. - [Release notes](https://github.com/pallets/jinja/releases) - [Changelog](https://github.com/pallets/jinja/blob/main/CHANGES.rst) - [Commits](https://github.com/pallets/jinja/compare/2.11.3...3.1.3 ) --- updated-dependencies: - dependency-name: jinja2 dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Younes Belkada authored
* add mixtral fused modules * add changes from modeling utils * add test * fix test + rope theta issue * Update src/transformers/modeling_utils.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * add tests --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
amyeroberts authored
* Update meatdata loading for oneformer * Enable loading from a model repo * Update docstrings * Fix tests * Update tests * Clarify repo_path behaviour
-
amyeroberts authored
* Mark two logger tests as flaky * Add description to is_flaky
-
Younes Belkada authored
* add llava + fused modules * Update src/transformers/models/llava/modeling_llava.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
- 11 Jan, 2024 14 commits
-
-
Hankyeol Kyung authored
* [docs] Fix broken link Signed-off-by:
Hankyeol Kyung <kghnkl0103@gmail.com> * [docs] Use shorter domain Signed-off-by:
Hankyeol Kyung <kghnkl0103@gmail.com> --------- Signed-off-by:
Hankyeol Kyung <kghnkl0103@gmail.com>
-
Matt authored
-
jiqing-feng authored
* update version for cpu training * update docs for cpu training * fix readme * fix readme
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
liangxuZhang authored
* Correct the implementation of auxiliary loss of mixtrtal * correct the implementation of auxiliary loss of mixtrtal * Implement a simpler calculation method --------- Co-authored-by:zhangliangxu3 <zhangliangxu3@jd.com>
-
Matt authored
* Correctly resolve trust_remote_code=None for AutoTokenizer * Second attempt at a proper resolution
-
Gustavo de Rosa authored
* chore(phi): Updates configuration_phi with missing keys. * chore(phi): Adds first draft of combined modeling_phi. * fix(phi): Fixes according to latest review. * fix(phi): Removes pad_vocab_size_multiple to prevent inconsistencies. * fix(phi): Fixes unit and integration tests. * fix(phi): Ensures that everything works with microsoft/phi-1 for first integration. * fix(phi): Fixes output of docstring generation. * fix(phi): Fixes according to latest review. * fix(phi): Fixes according to latest review. * fix(tests): Re-enables Phi-1.5 test. * fix(phi): Fixes attention overflow on PhiAttention (for Phi-2). * fix(phi): Improves how queries and keys are upcast. * fix(phi): Small updates on latest changes.
-
Harisankar Babu authored
* optionally preprocess segmentation maps for mobilevit * changed pretrained model name to that of segmentation model * removed voc-deeplabv3 from model archive list * added preprocess_image and preprocess_mask methods for processing images and segmentation masks respectively * added tests for segmentation masks based on segformer feature extractor * use crop_size instead of size * reverting to initial model
-
Alex Hedges authored
While using `run_clm.py`,[^1] I noticed that some files were being added to my global cache, not the local cache. I set the `cache_dir` parameter for the one call to `evaluate.load()`, which partially solved the problem. I figured that while I was fixing the one script upstream, I might as well fix the problem in all other example scripts that I could. There are still some files being added to my global cache, but this appears to be a bug in `evaluate` itself. This commit at least moves some of the files into the local cache, which is better than before. To create this PR, I made the following regex-based transformation: `evaluate\.load\((.*?)\)` -> `evaluate\.load\($1, cache_dir=model_args.cache_dir\)`. After using that, I manually fixed all modified files with `ruff` serving as useful guidance. During the process, I removed one existing usage of the `cache_dir` parameter in a script that did not have a corresponding `--cache-dir` argument declared. [^1]: I specifically used `pytorch/language-modeling/run_clm.py` from v4.34.1 of the library. For the original code, see the following URL: https://github.com/huggingface/transformers/tree/acc394c4f5e1283c19783581790b3dc3105a3697/examples/pytorch/language-modeling/run_clm.py.
-
Yih-Dar authored
fix docker file Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
update Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
ikkvix authored
* change truncate_sequences * Update tokenization_utils_base.py * change format * fix when ids_to_move=0 * fix * Update src/transformers/tokenization_utils_base.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> --------- Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
amyeroberts authored
Enable multi-label image classification
-
jiqing-feng authored
* Assitant model may on a different device * fix tensor device
-