- 13 May, 2022 2 commits
-
-
Younes Belkada authored
* try fixes * Revert "try fixes" This reverts commit a8ad75ef69d4fc03a402ef61bd034b018aa8555e. * add correct shape * add correct path
-
Rafael Zimmer authored
* Added translation of installation.mdx to Portuguese, as well as default templates of _toctree.yml and _config.py * [ build_documentation.yml ] - Updated doc_builder to build documentation in Portuguese. [ pipeline_tutorial.mdx ] - Created translation for the pipeline_tutorial.mdx. * [ build_pr_documentation.yml ] - Added pt language to pr_documentation builder. [ pipeline_tutorial.mdx ] - Grammar changes. * [ accelerate.mdx ] - Translated to Portuguese the acceleration tutorial. * [ multilingual.mdx ] - Added portuguese translation for multilingual tutorial. [ training.mdx ] - Added portuguese translation for training tutorial. * [ preprocessing.mdx ] - WIP * Update _toctree.yml * Adding Pré-processamento to _toctree.yml * Update accelerate.mdx * Nits and eliminate preprocessing file while it is ready Co-authored-by:Omar U. Espejel <espejelomar@gmail.com>
-
- 12 May, 2022 17 commits
-
-
fxmarty authored
* Fix typo * Force rerun workflows Co-authored-by:Felix Marty <felix@huggingface.co>
-
Sylvain Gugger authored
* Black preview * Fixup too! * Fix check copies * Use the same version as the CI * Bump black
-
Suraj Patil authored
-
Lysandre Debut authored
-
Matt authored
* Add test to ensure models can take int64 inputs * is_integer is an attribute, not a method * Fix test when some inputs aren't tensors * Add casts to blenderbot and blenderbot-small * Add casts to the other failing models
-
Lysandre Debut authored
-
Sayak Paul authored
* Update data2vec.mdx * Update data2vec.mdx * Update docs/source/en/model_doc/data2vec.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Li Dong authored
## Motivation We are going to use a new blob account to store the checkpoints. ## Modification Modify the azure blob storage URLs for BEiT checkpoints.
-
Younes Belkada authored
* First version - OPT model * Final changes - putting use cache to False * few changes - remove commented block * few changes - remove unecessary files * fix style issues * few changes - remove a test file - added the logits test * Update src/transformers/models/auto/tokenization_auto.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * add gen tests * few changes - rm mask filling example on docstring * few changes - remove useless args * some changes - more tests should pass now - needs to clean more - documentation still needs to be done * fix code quality * major changes - change attention architecture to BART-like - modify some tests - style fix * rm useless classes - remove opt for: - QA - cond generation - seq classif * Removed autodoc calls to non-existant classes TOkenizers are not implemented * Update src/transformers/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/auto/modeling_tf_auto.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Replaced OPTTokeniser with GPT2 tokenizer * added GPT2Tokenizer.from_pretrained("patrickvonplaten/opt_gpt2_tokenizer") * Removed OPTTokenizer * make style * Make style replaces ``` ...).unsqueeze(``` by ``` >>>).unsqueeze(``` * make repo consistency * Removed PretrainedOPTModel * fix opt.mdx removed other heads * fix init, removed 3 heads * removed heads * finished cleaning head * removed seauence classif and question answering * removed unused imports * removed useless dummy object for QA, SC and CG * removed tests for removed useless dummy object for QA, SC and CG * Removed head_mask using encoder layers which don't exist * fixed test * fix line * added OPT to toctree * Updated model path with pushed weigths * fix model path * fixed code quality * fixed embeddings and generation tests * update paths * clean comments * removed OPTClassificationHead for sentence classification * renamed hidden layer * renamed num layers to standard num_hidden_layers * num_attention_heads fix * changes for 125m * add first version for 125m * add first version - flax * add new version * causal LM output * replace output type with BaseModelOutputWithPastAndCrossAttentions * revert working config from 150m to 350m * clean * removed decoder input ids * fixed embed dim * more embed_dim issues * make style + removed enc_dec test * update falx model * removed troublesome copy * added is_encoder_decoder=False to config * added set_input emb fuinction to model class * requires torch on embed test * use head mask instead of decoder head mask input param solves a test * 8 test remaining, update * Updated create_and_check_decoder_model_past_large_inputs * Make style * update op tokenizer with condition * make style * See if I can push * some clean up * remove linear head hack * save intermediate * save correct attention * add copied from from bart * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * fix part of the reviewss Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * same changes in naming / conversion * correct mask * more fixes * delete FlaxOPT and TfOPT * clean traces of Flax and Tf * fix mask * fixed positionnal embedding length when past key value is provoded * get 125m, 6.7b to work * Added do_layer_norm * solved mismatch in load dictionnary * clean up preapre opt input dict * fixed past key value as bool * fix previus * fixed return dict False tuple issue * All tests are passing * Make style * Ignore OPTDecoder non tested * make fix-copies * make repo consistency * small fix * removed uselss @torch.no_grad decorator * make styl;e * fix previous opt test * style * make style * added opt documentation * update OPT_PRETRAINED_MODEL_ARCHIVE_LIST * up * more fixes * model & config work * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * added comment on padding hack (+2) * cleaup * review update * docstring for missing arg * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/__init__.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * update pretrained map * update path and tests * make style * styling * make consistency * add gpt2 tok new * more tok fixes * Update src/transformers/models/auto/tokenization_auto.py * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update tests/models/opt/test_modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update based on reviews * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * make style * make tokenizer auto tests pass * apply Lysandre suggestion * finish tests * add some good tokenizer tests * improve docs slighly Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
ArthurZucker <arthur.zucker@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Michael Benayoun authored
* Support tracing for ViT * Swin support * Fix copies * Fix type annotation issue * Removed unused import
-
Omar U. Espejel authored
* Fix contents in index.mdx to match docs' sidebar * Eliminates api section from contents
-
Omar Sanseviero authored
-
Omar U. Espejel authored
* Change nits in Spanish for quicktour.mdx - Add tasks names in English too. - Fix small nits in Spanish * Translate index.mdx to Spanish * Translate body of index. * Translated the compatible models list (not the papers´ names). Since this should not be updated manually, I can come back to the original text. * Add models and a dataset for Spanish in the code exmaples * Replaced the English models to Spanish versions. * Add index to _toctree.yml and fix Spanish * Fix double ““ error * Change negative example in ASR example * make style * Debug style in quicktour.mdx
-
Jorge Loayza R authored
* Spanish translation of fast_tokenizers.mdx * add fast_tokenizers to the spanish _toctree.yml * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Joaq authored
* Spanish version of language_modeling.mdx doc file * modification to toctree.yml file * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Correct position of Guías conceptuales Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
jkmg authored
* adding philosophy.mdx translation to Spanish * adding philosophy.mdx translation to Spanish * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * philosophy translation to Spanish * Update _toctree.yml * Update _toctree.yml * nits Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Shijie Wu authored
-
- 11 May, 2022 13 commits
-
-
Amanpreet Singh authored
* [WIP] Add FLAVA model This PR aims to add [FLAVA](ihttps://arxiv.org/abs/2112.04482) model to the transformers repo. Following checklist delineates the list of things to be done for this PR to be complete: [x] Flava init [x] Flava base models [x] Flava layers [x] Flava Configs [x] Flava encoders [x] Flava pretraining models [ ] Flava classification/retrieval models (To be added in a separate PR) [x] Documentation updates [x] Imports updates [x] Argstring updates [x] Flava pretrained checkpoints [x] Flava tests [x] Flava processors [x] Sanity check [x] Lint
-
Antoni Baum authored
-
Suraj Patil authored
-
Antoni Baum authored
* Ensure tensors are at least 1d for pad and concat * Compatibility * Fix * Fix * Add test * Retrigger CI * Consistency with master * Retrigger CI
-
Manuel R. Ciosici authored
* Fix markdown code block * Use consistent spelling for self-attention * Fix typos and phrasing * Fix code style
-
Antoni Baum authored
* Remove unneeded columns for IterableDataset * Add test * Update trainer tests * Edit docstring * Lint * Apply feedback * Apply feedback
-
hasan salim kanmaz authored
* add seed worker and set_deterministic_seed_for_cuda function to enforce reproducability * change function name to enable determinism, add docstrings, reproducability support for tf * change function name to enable_determinism_for_distributed_training * revert changes in set_seed and call set_seed within enable_full_determinism * add one position argument for seed_worker function * add full_determinism flag in training args and call enable_full_determinism when it is true * add enable_full_determinism to documentation * apply make fixup after the last commit * Update src/transformers/training_args.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Martin Pömsl authored
* Create RetriBERT tests folder * Add missing RetriBERT tokenizer test file * Apply style corrections * Add non-english filter * Update tests/retribert/test_tokenization_retribert.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * Update tests/retribert/test_tokenization_retribert.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * Move test files to new directory * Update import path for testing utils to new test file structure Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com>
-
Heng Kuan Wee authored
Co-authored-by:kuanwee.heng <kuanwee.heng@aaqua.live>
-
Sylvain Gugger authored
-
arampacha authored
* propagate attention_mask dtype * fixup&style
-
jianan-gu authored
* add torch SGD and Adagrad optimizer bits * refine naming Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Suraj Patil authored
* remove duplicate example * remove code block
-
- 10 May, 2022 8 commits
-
-
Leon Derczynski authored
* unhardcode pretrained model path, make it a class var * add tests for mobilebert tokenizer * allow tempfiles for vocab & merge similarity test to autodelete * add explanatory comments * remove unused imports, let make style do its.. thing * remove inheritance and use BERT tok tests for MobileBERT * Update tests/mobilebert/test_tokenization_mobilebert.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * amend class names, remove unused import, add fix for mobilebert's hub pathname * unhardcode pretrained model path, make it a class var * add tests for mobilebert tokenizer * allow tempfiles for vocab & merge similarity test to autodelete * add explanatory comments * remove unused imports, let make style do its.. thing * remove inheritance and use BERT tok tests for MobileBERT * Update tests/mobilebert/test_tokenization_mobilebert.py Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com> * amend class names, remove unused import, add fix for mobilebert's hub pathname * amend paths for model tests being in models/ subdir of /tests * explicitly rm test from prev path Co-authored-by:
SaulLu <55560583+SaulLu@users.noreply.github.com>
-
Jason Phang authored
-
Sylvain Gugger authored
-
Nicolas Brousse authored
* add support for MLFLOW_FLATTEN_PARAMS * ensure key is str * fix style and update warning msg * Empty commit to trigger CI * fix bug in check_inits.py * add unittest for flatten_dict utils * fix 'NoneType' object is not callable on __del__ * add generic flatten_dict unittest to SPECIAL_MODULE_TO_TEST_MAP * fix style
-
Stas Bekman authored
-
Patrick Haller authored
-
Stas Bekman authored
* model zoo take 2 * add deberta * new param for zero2 * doc update * doc update * add layoutlm * bump deepspeed * add deberta-v2, funnel, longformer * new models * style * add t5_v1 * update TAPAS status * reorg problematic models * move doc to another PR * style * fix checkpoint check test * making progress on more models running * cleanup * new version * cleanup
-
Stas Bekman authored
* [trainer] sharded _load_best_model probably needs a test? * undo delete
-