- 06 Jun, 2024 16 commits
-
-
Alex Gorodnitskiy authored
Fix DonutSwinLayer attention mask device
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/bertabs Bumps [transformers](https://github.com/huggingface/transformers) from 3.5.1 to 4.38.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Vu Huy Nguyen authored
* Add list check for image and question * Handle passing two lists and update docstring * Add tests * Add support for dataset * Add test for dataset as input * fixup * fix unprotected import * fix unprotected import * fix import again * fix param type
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/codeparrot Bumps [transformers](https://github.com/huggingface/transformers) from 4.19.0 to 4.38.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v4.19.0...v4.38.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
amyeroberts authored
* Mark MobileNetV1ModelTest::test_batching_equivalence as flaky * Add link to issue * woops
-
Omar Salman authored
* Initial attempt * Updates: PR suggestions * Interpolate the relative position bias when interpolate_pos_encoding is True * Add slow tag for the added tests * Add in DATA2VEC_VISION_INPUTS_DOCSTRING
-
Marc Sun authored
* fix accelerate tests for roberta xl * style
-
Baole Ai authored
* Fix _save_tpu: use _maybe_convert_to_cpu instead of to cpu. * fix lint
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/bertology Bumps [transformers](https://github.com/huggingface/transformers) from 3.5.1 to 4.38.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Huazhong Ji authored
-
Lucain authored
Switch from hf_hub_url to hf_hub_download in remaining occurences
-
Raushan Turganbay authored
* fix special tokens in generatioon * fix test * add warning * fix the check * warn once * fix
-
Raushan Turganbay authored
* make mamba use cache * uss cache naming as in mamba * fix musicgen
-
Zhiyuan Chen authored
-
Ranggi Hwang authored
* SwitchTransformer MoE layer performance improvement * make fixup * comments about shapes * make fixup
-
graham authored
no need for explicit EXTRA_TOKENS
-
- 05 Jun, 2024 12 commits
-
-
amyeroberts authored
Skip failing tests for now
-
Cyril Vallez authored
* Fix contrastive_search for new cache structure, and improve performance by removing inneficient torch.stack(torch.split(x, top_k, dim=0)) * Fix _contrastive_search for non-standard cache using ellipsis slicing * Fix all outputs.logits memory leaks for all decoding strategies! * Fix small error in _contrastive_search() * Make all necessary change and revert for the new class * Apply coding style * Remove pipes in type hints for compatibility * correct type hint * apply style * Use DynamicCache by default and solve conflicts * Fix rebase issues * Add `_supports_dynamic_cache_class` in models for models that support DynamicCache but not other caches to make DynamicCache the default for more models * Create generation config to return legacy format by default, or to choose not to * style * Fix case when use_cache is False * Remove default DynamicCache in assiste_decoding if assistant_model does not support it + fix _seen_tokens when cropping cache * Update prepare_inputs_for_generation() for case with empty DynamicCache * Correct return of args in _assisted_decoding * Remove EfficientDynamicCache as it is no longer needed * Correct mistake in generation config * Move cache logic of assisted decoding to AssistedCandidateGenerator.__init__ * change DynamicCache function names from "split" to "batch_split" for readability + apply coding style * Remove `_supports_dynamic_cache_class` attribute after rebase * Correct missing line lost in conflict resolution during rebasing * Add special case for Jamba * Fix jamba test * Coding style * coding style * Correct missing import in rebasing * Simplify _validate_model_kwargs based on removal of _supports_dynamic_cache attribute * Simplify code paths in _contrastive_search * coding style * Update docstrings of cache methods * Update prepare_inputs_for_generation() -> past_key_values are always Cache objects
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Dhaivat Bhatt authored
-
bastrob authored
* add flaubert tokenization test, enrich inheritance in FlaubertTokenizer. * fix quality code ci * ensure parameter consistency * fix ci * fix copyright year and flatten vocab list. * fix style
-
Huazhong Ji authored
-
Vaibhav Srivastav authored
* doc: add info about wav2vec2 bert in older wav2vec2 models. * apply suggestions from review. * forward contrib credits from review --------- Co-authored-by:Sanchit Gandhi <sanchit-gandhi@users.noreply.github.com>
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/deebert Bumps [transformers](https://github.com/huggingface/transformers) from 3.5.1 to 4.38.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v3.5.1...v4.38.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
amyeroberts authored
* Move label validation checks - fail early * Remove some formatting changes - add back labels change wav2vec2
-
Yih-Dar authored
* benchmark workflow * benchmark workflow * benchmark workflow * benchmark workflow * build * build * build * build * build * build * build * build * build * build * build * build * build * build --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
James Braza authored
Fixed torch definition error
-
Yury Sulsky authored
The StoppingCriteriaList allocates is_done without specifying dtype=torch.bool. On XLA this allocates a float tensor and causes a failure on the following line: is_done = is_done | criteria(input_ids, scores, **kwargs) by attempting to OR float with bool.
-
- 04 Jun, 2024 12 commits
-
-
dependabot[bot] authored
Bump transformers in /examples/research_projects/vqgan-clip Bumps [transformers](https://github.com/huggingface/transformers) from 4.26.0 to 4.38.0. - [Release notes](https://github.com/huggingface/transformers/releases) - [Commits](https://github.com/huggingface/transformers/compare/v4.26.0...v4.38.0 ) --- updated-dependencies: - dependency-name: transformers dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Yih-Dar authored
* build * build * build * build * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
amyeroberts authored
* Move out common validation * Add missing backbone config arguments
-
Younes Belkada authored
* deprecate blip * mention deprecation on docs
-
Yih-Dar authored
* fix * fix * fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Manuel Faysse authored
-
Jacklanda authored
✨ Add new line switch before logging "***** Running {description} *****". Signed-off-by:jacklanda <yonyonlau@gmail.com>
-
amyeroberts authored
* Fix pipeline tests - torch imports * Frameowrk dependant float conversion
-
Chujie Zheng authored
* fix logits dtype * Add bf16/fp16 tests for text_classification pipeline * Update test_pipelines_text_classification.py * fix * fix
-
Kristen Pereira authored
* Added interpolate pos encoding feature and test to deit * Added interpolate pos encoding feature and test for deit TF model * readded accidentally delted test for multi_gpu * storing only patch_size instead of entire config and removed commented code * Update modeling_tf_deit.py to remove extra line Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> --------- Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Raushan Turganbay authored
video-llava can handle more frames
-
Max Strobel authored
* fix(PatchTST): Wrong dropout used for PretainHead * feat(PatchTST): remove unused config.dropout --------- Co-authored-by:Strobel Maximilian (IFAG PSS SIS SCE ACM) <Maximilian.Strobel@infineon.com>
-