- 03 Jan, 2023 14 commits
-
-
radcheb authored
* Update trainer.py * fix style Co-authored-by:Radhwane Chebaane <rchebaane.external@epo.org>
-
Alara Dirik authored
* add post_process_object_detection method * style changes
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* Avoid null CI * Avoid null CI * rename * more clear error message * Update .circleci/config.yml Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * clean up Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Anna Krogager authored
-
samuelpullely authored
* Enable `decoder_attention_mask` in `generate` function * Make style corrections * Run `make repo-consistency` * Add integration test
-
JeongYeon Nam authored
* fix: valid ratio has right value * chore: remove unnecessary line Co-authored-by:Jeongyeon Nam <jy.nam@navercorp.com>
-
Wang, Yi authored
* [run_clm example] add torch_dtype option for model load. for BLOOM 175B model. peak memory will reduce about 350G for inference. the weight of BLOOM in model hub is bfloat16 Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * add other type in option * fix style Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com>
-
Yih-Dar authored
Remove more unused attributes in config classes Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
NielsRogge authored
* First draft * Make model instantiation work * Fix copied from statement * More fixes * Add correct output head * Improve configuration * Add conversion script * Improve conversion script * Remove token_type_ids * Fix conversion of projection layers * Convert all weights * Use cats image * Make logits match * Generate caption on cats image * Add GITProcessor * Update conversion script * Add support for more checkpoints * Fix conversion script * Add initial tests * Remove cross-attention * More improvements * Remove is_decoder * Improve model tests * Improve tests * Improve model outputs * Fix model outputs equivalence * Fix more tests * Remove unused code * Use generate to generate text, no use of cache for now * Use generate more appropriately * Fix config tests * Fix style * Add support for use_cache Co-authored-by:
Joao Gante <joaofranciscocardosogante@gmail.com> * Fix style * Fix GIT vision encoder * Update README * Fix integration test * Set bos and eos token ids * Improve docs * Improve code * Add support for provided attention_mask * Add copied from statement * Fix gradient checkpointing test * Set model_input_names * Investigate model_input_names * Remove script * Fix model inputs * Fix docstring * Rename GIT to Git * Support more models * Add support for textvqa model * Add video support * Extend conversion script for video * Add support for large variant * Add support for more models * Fix config archive map * Update integration test * Fix README * Fix CLIP mean and std * Update processor * Fix use_cache for video, thanks @gante * Remove print statements * Remove assertion * Add processor tests * Fix model_input_names * Use Auto API for processor * Fix processor tests * Fix integration test * Fix pipeline test * Make tests faster * Update conversion script * Update conversion script * Convert more checkpoints * Update conversion script * Fix typo * Update docstrings * Improve code snippets * Fix doc tests * Add more code examples茅 * Fix doc tests * Add integration tests * Fix unused variable * revert * Add GIT to Japanese README Co-authored-by:
Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local> Co-authored-by:
Joao Gante <joaofranciscocardosogante@gmail.com> Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com>
-
Alara Dirik authored
fix post_process_object_detection descriptions
-
Konstantin Kotik authored
* feat: add min new length logit processor * test: add min new length logit processor * docs: add MinNewTokensLengthLogitsProcessor * feat: import MinNewTokensLengthLogitsProcessor * fix: update pytorch dummy objects * refactor & fix: rename attributes and var and get rid of dynamic attribute * tests: align test with new interface * docs: fix typo * docs: minor clarification * Empty-Commit * empty commit * run automated quality edits Co-authored-by:Joao Gante <joao@huggingface.co>
-
Joao Gante authored
-
ivanllt authored
Fix start_docstring for deparallelize method
-
- 02 Jan, 2023 1 commit
-
-
Joao Gante authored
* beam sample in beam search * rag now works with the updated beam search * delete legacy (non-XLA) generation code related to beam sample
-
- 31 Dec, 2022 4 commits
-
-
Hao Wang authored
* update pyknp to rhoknp * fix linter * fix linter * fix linter * fix linter * fix linter * support rhoknp==1.1.0, fix testcase
-
bofeng huang authored
* Add generate kwargs to AutomaticSpeechRecognitionPipeline * Add test for generation kwargs
-
bofeng huang authored
* Add generate kwargs to AutomaticSpeechRecognitionPipeline * Add test for generation kwargs
-
Stas Bekman authored
[trainer: distributed_concat] ensure all_gather's input are contiguous
-
- 30 Dec, 2022 3 commits
-
-
Samuel Xu authored
Fixing error message
-
bofeng huang authored
* Fix error message * Fix code quality
-
Matthew McDermott authored
-
- 29 Dec, 2022 4 commits
-
-
ivanllt authored
-
bofeng huang authored
* add torch_dtype attribute to Pipeline * Use torch_dtype to cast input tensor type in AutomaticSpeechRecognitionPipeline * Fix code quality * Add TextGenerationPipeline fp16 test * Fix code quality * Remove useless require in tests Co-authored-by:
Nicolas Patry <patry.nicolas@protonmail.com> Co-authored-by:
Nicolas Patry <patry.nicolas@protonmail.com>
-
Harsh Trivedi authored
load the state dict on cpu.
-
Alex Hedges authored
* Remove non-breaking space in comment It was likely added unintionally. * Remove remaining non-breaking spaces
-
- 28 Dec, 2022 2 commits
-
-
Joao Gante authored
correctly detect default max length
-
Akshaya Annavajhala authored
* Avoid collisions in writing metrics via 2 APIs - azureml + mlflow MLflow tracking API is enabled by default in AzureML and HF MLflow integration is more fully featured. I'd remove the AzureML integration but leaving the current behavior for backwards compatibility (though it should really be removed) * Trigger CI
-
- 27 Dec, 2022 3 commits
-
-
Yih-Dar authored
* torch.jit._state * Fix past CI * Fix for perceiver * Fix REALM * Fix for Bloom * Fix for SwinMode * Fix for TrajectoryTransformerModel * Fix for test_wav2vec2_with_lm * make style Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Eli Simhayev authored
fix Jupyter typos
-
Kamal Raj Kanakarajan authored
Update flan-t5.mdx
-
- 26 Dec, 2022 2 commits
-
-
Younes Belkada authored
* fix fp16 loading issue * add backward compatibility * better refactor * better readability - remove `force_upcast_dtype` as it is used once - use `inspect` - add `TODO`
-
Nathan Barry authored
-
- 24 Dec, 2022 1 commit
-
-
M谩rton Makrai authored
-
- 23 Dec, 2022 6 commits
-
-
Arthur authored
get correct checkpoint
-
Younes Belkada authored
* add japanese translation of template * fix japanese translation - fix special cases - fix typos - manually translate special cases Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com>
-
Jasmijn Bastings authored
* Add script to convert T5X T5 (v1.0 and v1.1) checkpoints to PyTorch * Remove unnecessary check and update docstring * Format docstring * Fix whitespace in docstring
-
Nicolas Patry authored
* Supporting `fp16` for asr pipeline * Adding test. * Style. * Oops. * Flake8 update ? * Fixing flake8 ? * Revert "Flake8 update ?" This reverts commit 0b917fcb520e5f34d1933d9d37d8f32b64553048. * Style (acctidentally deleted flake8 F401.) * Move to a bigger test (no small whisper model, and s2t doesn't seem to accept torch_dtype=fp16). Also we need to use a GPU to actually compute on fp16. * Using BatchFeature capability.
-
Syed Abdul Gaffar Shakhadri authored
poolformer onnx Co-authored-by:syed <syed.abdul@sandlogic.com>
-
Sourab Mangrulkar authored
-