- 12 Dec, 2022 18 commits
-
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
amyeroberts authored
* Add decorator for flaky tests * Fix up
-
Sylvain Gugger authored
* Disambiguate test for required_input in tokenization base file. * Add test for size
-
Sylvain Gugger authored
-
Ariel Ekgren authored
* Add templates for gpt-sw3 * Add templates for gpt-sw3 * Added sentencepiece tokenizer * intermediate commit with many changes * fixed conflicts * Init commit for tokenization port * Tokenization progress * Remove fast tokenizer * Clean up and rename spm.model -> spiece.model * Remove TF -> PT conversion script template, Clean up Megatron -> PT script * Optimize encode & decode performance * added new attention * added new attention * attention for gpt-sw3 working * attention good * Cache is now working * fixed attention mask so that it works with causal attention * fixed badbmm bug for cpu and caching * updated config with correct parameters * Refactor and leave optimizations as separate functions to avoid breaking expected functionality * Fix special tokens mapping for both tokenizers * cleaning up of code and comments * HF compatible attention outputs * Tokenizer now passing tests, add documentation * Update documentation * reverted back to base implementation after checking that it is identical to pretrained model * updated gpt-sw3 config * updated conversion script * aligned parameters with gpt-sw3 config * changed default scale_attn_by_inverse_layer_idx to true * removed flag from conversion script * added temporary model path * reverted back to functioning convert script * small changes to default config * updated tests for gpt-sw3 * make style, make quality, minor cleanup * Change local paths to testing online repository * Change name: GptSw3 -> GPTSw3 * Remove GPTSw3TokenizerFast references * Use official model repository and add more model sizes * Added reference to 6.7b model * Add GPTSw3DoubleHeadsModel to IGNORE_NON_AUTO_CONFIGURED, like GPT2DoubleHeadsModel * Remove pointers to non-existing TFGPTSw3 * Add GPTSw3 to docs/_toctree.yml * Remove TF artifacts from GPTSw3 in __init__ files * Update README:s with 'make fix-copies' * Add 20b model to archive list * Add documentation for GPT-Sw3 * Fix typo in documentation for GPT-Sw3 * Do 'make fix-copies' again after having updated docs * Fix some typos in docs * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/configuration_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update tests/models/gpt_sw3/test_tokenization_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/modeling_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Resolve comments from PR feedback * Resolve more comments from PR feedback, also set use_cache=True in convert script * Add '# Copied from' comments for GPTSw3 modeling * Set 'is_parallelizable = False' * Remove '# Copied from' where code was modified and add 'with x->y' when appropriate * Remove parallelize in mdx * make style, make quality * Update GPTSw3Config default values and corresponding documentation * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/__init__.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Clean up and protect GPTSw3Tokenizer imports with is_sentencepiece_available * Make style, make quality * Add dummy object for GPTSw3Tokenizer via 'make fix-copies' * make fix-copies * Remove GPTSw3 modeling classes * make style, make quality * Add GPTSw3 auto-mappings for other GPT2 heads * Update docs/source/en/model_doc/gpt-sw3.mdx Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/convert_megatron_to_pytorch.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/gpt_sw3/tokenization_gpt_sw3.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Remove old TODO-comment * Add example usage to GPTSw3Tokenizer docstring * make style, make quality * Add implementation details and example usage to gpt-sw3.mdx Co-authored-by:
JoeyOhman <joeyoh@kth.se> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
amyeroberts authored
* Add require_vision decorator * Fixup * Use requires_backends * Add requires_backend to utils functions
-
Steven Liu authored
* clarify docstring * make style
-
Matt authored
* Convert tokenizer outputs for Keras in doc example * Das deutsche Beispiel auch korrigieren
-
Juanjo do Olmo authored
* Create and translate to Spanish debugging.mdx * solved typo error in a header * Update debugging.mdx * Update debugging.mdx * Update docs/source/es/debugging.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/debugging.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/debugging.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/debugging.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/debugging.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update _toctree.yml Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sourab Mangrulkar authored
-
stanleycai95 authored
-
Ian C authored
* Initial commit * Add type hints for two major classes * Run make fixup * Fix output type for Whisper * Run isort to fix imports
-
Nicolas Patry authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Peter authored
*
✨ add accelerate support for LongT5 models Signed-off-by:peter szemraj <peterszemraj@gmail.com> * fix `accelerate` tests * Trigger CI test Signed-off-by:
peter szemraj <peterszemraj@gmail.com> Co-authored-by:
younesbelkada <younesbelkada@gmail.com>
-
Alberto Mario Ceballos-Arroyo authored
* Fix minor typo in question_answering.mdx * Fixes minor typo in the english version of tasks/asr.mdx * Update _toctree.yml * Translate add_new_pipeline.mdx into Spanish * Fixes some typos in the English version of add_new_pipeline.mdx * Translate asr.mdx into Spanish * Fixes small typos in add_new_pipeline.mdx * Update docs/source/es/add_new_pipeline.mdx Suggestion by @osanseviero Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/add_new_pipeline.mdx Suggestion by @osanseviero: use "biblioteca" instead of "librería." Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/tasks/asr.mdx Suggestion by @osanseviero. Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/add_new_pipeline.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/add_new_pipeline.mdx Suggestion by @osanseviero. Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/add_new_pipeline.mdx Suggestion by @osanseviero. Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/add_new_pipeline.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/tasks/asr.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/tasks/asr.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update docs/source/es/tasks/asr.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com> * Update asr.mdx Co-authored-by:
Omar Sanseviero <osanseviero@gmail.com>
-
Salvo Cavallaro authored
-
- 09 Dec, 2022 7 commits
-
-
Sylvain Gugger authored
* Fix rendering issue in quicktour * Separate in two blocks
-
Younes Belkada authored
* fix last slow test * revert deletion * Update src/transformers/models/vit_hybrid/modeling_vit_hybrid.py
-
amyeroberts authored
-
amyeroberts authored
* Replace FE references with IPs * Update processor tests * Update src/transformers/models/clip/processing_clip.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/clip/processing_clip.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update warning messages v4.27 -> v5 * Fixup * Update Chinese CLIP processor * Add feature_extractor property * Add attributes * Add tests Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Michael Benayoun authored
* Change transformers.onnx to use optimum.exporters.onnx * Update doc * Remove print * Fix transformers.onnx cli * Update documentation * Update documentation * Small fixes * Fix log message * Apply suggestions * Update src/transformers/onnx/__main__.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Apply suggestions * Add missing line break * Ran make fix-copies * Update src/transformers/onnx/__main__.py Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> * Update src/transformers/onnx/__main__.py Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com> Co-authored-by:
Michael Benayoun <michael@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
lewtun <lewis.c.tunstall@gmail.com>
-
NielsRogge authored
* Improve ResNet backbone * Improve Bit backbone * Improve docstrings * Fix default stage * Apply suggestions from code review Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
- 08 Dec, 2022 15 commits
-
-
Nathan Raw authored
*
🚧 wip video classification pipeline *🚧 wip - add is_decord_available check *🐛 add missing import *✅ add tests *🔧 add decord to setup extras *🚧 add is_decord_available *✨ add video-classification pipeline *📝 add video classification pipe to docs *🐛 add missing VideoClassificationPipeline import *📌 add decord install in test runner *✅ fix url inputs to video-classification pipeline *✨ updates from review *📝 add video cls pipeline to docs *📝 add docstring *🔥 remove unused import *🔥 remove some code *📝 docfix -
amyeroberts authored
* Add deprecation warning when image FE instantiated * Update src/transformers/models/beit/feature_extraction_beit.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update v2.7 -> v5 and add for new IPs * Add message to Chinese CLIP Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
IMvision12 authored
* Create test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py * Update test_tokenization_led.py
-
amyeroberts authored
* fix donut image processor * Update test values * Apply lower bound on resizing size * Add in missing size param * Resolve resize channel_dimension bug * Update src/transformers/image_transforms.py
-
Yih-Dar authored
* fix 1 * fix 2 * fix 3 * fix 4 Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
jeffhataws authored
-
Younes Belkada authored
* fix failing `accelerate` tests * make fixup * smaller values * even lower
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Migrate torchdynamo to torch.compile * Add docstring and generic option * Properly use the function... * Reorg args
-
dependabot[bot] authored
Bumps [certifi](https://github.com/certifi/python-certifi) from 2020.6.20 to 2022.12.7. - [Release notes](https://github.com/certifi/python-certifi/releases) - [Commits](https://github.com/certifi/python-certifi/compare/2020.06.20...2022.12.07 ) --- updated-dependencies: - dependency-name: certifi dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
dependabot[bot] authored
Bumps [certifi](https://github.com/certifi/python-certifi) from 2021.10.8 to 2022.12.7. - [Release notes](https://github.com/certifi/python-certifi/releases) - [Commits](https://github.com/certifi/python-certifi/compare/2021.10.08...2022.12.07 ) --- updated-dependencies: - dependency-name: certifi dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
dependabot[bot] authored
Bumps [certifi](https://github.com/certifi/python-certifi) from 2020.6.20 to 2022.12.7. - [Release notes](https://github.com/certifi/python-certifi/releases) - [Commits](https://github.com/certifi/python-certifi/compare/2020.06.20...2022.12.07 ) --- updated-dependencies: - dependency-name: certifi dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
-
Yih-Dar authored
* add BackboneBaseModel * add BackboneBaseModel * Rename to BackboneMixin * remove nn.Module Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Matt authored
-
Sylvain Gugger authored
* Update the list of contributors to reflect current organization * Proper indent
-