- 16 May, 2022 7 commits
-
-
Loubna Ben Allal authored
* add pretokenization arguments * add pretokenization script * add support for pretokenized data * reformat code * fix run command for training * fix model call from config * remove a package * add comments on pretokenization in the readme * remove explicit parallelization Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * update readme Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * update readme -remove username Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * update readme -remove username Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * keep data parallelization * reformat code * reformat code * update readme * reformat code * Update examples/research_projects/codeparrot/README.md Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> Co-authored-by:
Loubna ben allal <loubnabenallal@gmail.com>
-
Loubna Ben Allal authored
* add new preprocessing arguments * add new filters * add new filters to readme * fix config and test count, update function names and docstrings * reformat code * update readme * Update readme * rename config_test filter Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * rename few_assignments filter Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * rename tokenizer in arguments Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> * rename functions and add limit_line argument for config_test filter * update threshold for config_test filter Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> Co-authored-by:
Loubna ben allal <loubnabenallal@gmail.com>
-
cavdard authored
* adding partial checkpoint support for optimizer state * formatted trainer.py * Refactoring based on comments * reformatting * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/trainer.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Cavdar <dcavdar@a07817b12d7e.ant.amazon.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Kenneth Enevoldsen authored
* fixed bug run_mlm_flax_stream.py Fixed bug caused by an update to tokenizer keys introduced in recent transformers versions (between `4.6.2` and `4.18.0`) where additional keys were introduced to the tokenizer output. * Update run_mlm_flax_stream.py * adding missing paranthesis * formatted to black * remove cols from dataset instead * reformat to black * moved rem. columns to map * formatted to black Co-authored-by:KennethEnevoldsen <kennethcenevolsen@gmail.com>
-
Stas Bekman authored
* [doc] performance/scalability revamp * link the new docs * no : * mixed precision * work on the first doc * expand the main doc * Trigger CI * style * revamp single GPU training section * work on training performance * remove files not used anymore or will be added later * final touches * fix rebase * Add hardware section to toctree * fix toctree again * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * remove `fast_tokenizers` entry that was copied in rebase * add warning about DP vs DDP * remove todo * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * fix missing closure of codeblock * Update docs/source/en/perf_train_gpu_many.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * sync with #16860 * update toc Co-authored-by:
leandro <leandro.vonwerra@spoud.io> Co-authored-by:
Leandro von Werra <lvwerra@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Joao Gante authored
-
cloudhan authored
Change config.encoder_ffn_dim -> config.decoder_ffn_dim for decoder.
-
- 13 May, 2022 13 commits
-
-
Ignacio Talavera authored
* file copied and toctree updated * Intro and configuration translated * model section translated * enter hotfix * Translation over, correction pending * Typos and corrections * Update docs/source/es/create_a_model.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/create_a_model.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/create_a_model.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/create_a_model.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Gerardo Huerta Robles authored
* Translated version of model_sharing to spanish * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Update docs/source_es/model_sharing.mdx * Addind model sharing to _toctree.yml Co-authored-by:Omar U. Espejel <espejelomar@gmail.com>
-
Fellip Silva Alves authored
* [ fast_tokenizers.mdx ] - Added translation to portuguese to tutorial * Delete docs/source/pt-br directory * [ fast_tokenizers.mdx ] - Continuing work on file * [ fast_tokenizers.mdx ] - Continuing work on file * Add fast tokenizers to _toctree.yml * Eliminated config and toctree.yml * Nits in fast_tokenizers.mdx Co-authored-by:Omar U. Espejel <espejelomar@gmail.com>
-
Yih-Dar authored
* add PR title to push CI report * add link Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* update push ci * install git-python * update comment * update deepspeed jobs * fix report * skip 2 more tests that require fairscale * Fix changes in test_fetcher.py (to deal with `setup.py` is changed) * set RUN_PT_TF_CROSS_TESTS=1 and final clean-up * remove SIGOPT_API_TOKEN * remove echo "$matrix_folders" Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Patrick von Platen authored
* correct some stuff * fix doc tests * make style
-
Younes Belkada authored
* try fixes * Revert "try fixes" This reverts commit a8ad75ef69d4fc03a402ef61bd034b018aa8555e. * add correct shape * add correct path
-
Rafael Zimmer authored
* Added translation of installation.mdx to Portuguese, as well as default templates of _toctree.yml and _config.py * [ build_documentation.yml ] - Updated doc_builder to build documentation in Portuguese. [ pipeline_tutorial.mdx ] - Created translation for the pipeline_tutorial.mdx. * [ build_pr_documentation.yml ] - Added pt language to pr_documentation builder. [ pipeline_tutorial.mdx ] - Grammar changes. * [ accelerate.mdx ] - Translated to Portuguese the acceleration tutorial. * [ multilingual.mdx ] - Added portuguese translation for multilingual tutorial. [ training.mdx ] - Added portuguese translation for training tutorial. * [ preprocessing.mdx ] - WIP * Update _toctree.yml * Adding Pré-processamento to _toctree.yml * Update accelerate.mdx * Nits and eliminate preprocessing file while it is ready Co-authored-by:Omar U. Espejel <espejelomar@gmail.com>
-
- 12 May, 2022 17 commits
-
-
fxmarty authored
* Fix typo * Force rerun workflows Co-authored-by:Felix Marty <felix@huggingface.co>
-
Sylvain Gugger authored
* Black preview * Fixup too! * Fix check copies * Use the same version as the CI * Bump black
-
Suraj Patil authored
-
Lysandre Debut authored
-
Matt authored
* Add test to ensure models can take int64 inputs * is_integer is an attribute, not a method * Fix test when some inputs aren't tensors * Add casts to blenderbot and blenderbot-small * Add casts to the other failing models
-
Lysandre Debut authored
-
Sayak Paul authored
* Update data2vec.mdx * Update data2vec.mdx * Update docs/source/en/model_doc/data2vec.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Li Dong authored
## Motivation We are going to use a new blob account to store the checkpoints. ## Modification Modify the azure blob storage URLs for BEiT checkpoints.
-
Younes Belkada authored
* First version - OPT model * Final changes - putting use cache to False * few changes - remove commented block * few changes - remove unecessary files * fix style issues * few changes - remove a test file - added the logits test * Update src/transformers/models/auto/tokenization_auto.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * add gen tests * few changes - rm mask filling example on docstring * few changes - remove useless args * some changes - more tests should pass now - needs to clean more - documentation still needs to be done * fix code quality * major changes - change attention architecture to BART-like - modify some tests - style fix * rm useless classes - remove opt for: - QA - cond generation - seq classif * Removed autodoc calls to non-existant classes TOkenizers are not implemented * Update src/transformers/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/__init__.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Update src/transformers/models/auto/modeling_tf_auto.py Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> * Replaced OPTTokeniser with GPT2 tokenizer * added GPT2Tokenizer.from_pretrained("patrickvonplaten/opt_gpt2_tokenizer") * Removed OPTTokenizer * make style * Make style replaces ``` ...).unsqueeze(``` by ``` >>>).unsqueeze(``` * make repo consistency * Removed PretrainedOPTModel * fix opt.mdx removed other heads * fix init, removed 3 heads * removed heads * finished cleaning head * removed seauence classif and question answering * removed unused imports * removed useless dummy object for QA, SC and CG * removed tests for removed useless dummy object for QA, SC and CG * Removed head_mask using encoder layers which don't exist * fixed test * fix line * added OPT to toctree * Updated model path with pushed weigths * fix model path * fixed code quality * fixed embeddings and generation tests * update paths * clean comments * removed OPTClassificationHead for sentence classification * renamed hidden layer * renamed num layers to standard num_hidden_layers * num_attention_heads fix * changes for 125m * add first version for 125m * add first version - flax * add new version * causal LM output * replace output type with BaseModelOutputWithPastAndCrossAttentions * revert working config from 150m to 350m * clean * removed decoder input ids * fixed embed dim * more embed_dim issues * make style + removed enc_dec test * update falx model * removed troublesome copy * added is_encoder_decoder=False to config * added set_input emb fuinction to model class * requires torch on embed test * use head mask instead of decoder head mask input param solves a test * 8 test remaining, update * Updated create_and_check_decoder_model_past_large_inputs * Make style * update op tokenizer with condition * make style * See if I can push * some clean up * remove linear head hack * save intermediate * save correct attention * add copied from from bart * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * fix part of the reviewss Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * same changes in naming / conversion * correct mask * more fixes * delete FlaxOPT and TfOPT * clean traces of Flax and Tf * fix mask * fixed positionnal embedding length when past key value is provoded * get 125m, 6.7b to work * Added do_layer_norm * solved mismatch in load dictionnary * clean up preapre opt input dict * fixed past key value as bool * fix previus * fixed return dict False tuple issue * All tests are passing * Make style * Ignore OPTDecoder non tested * make fix-copies * make repo consistency * small fix * removed uselss @torch.no_grad decorator * make styl;e * fix previous opt test * style * make style * added opt documentation * update OPT_PRETRAINED_MODEL_ARCHIVE_LIST * up * more fixes * model & config work * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * added comment on padding hack (+2) * cleaup * review update * docstring for missing arg * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * Update src/transformers/models/opt/__init__.py Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> * update pretrained map * update path and tests * make style * styling * make consistency * add gpt2 tok new * more tok fixes * Update src/transformers/models/auto/tokenization_auto.py * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/model_doc/opt.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update tests/models/opt/test_modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/models/opt/modeling_opt.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update based on reviews * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * make style * make tokenizer auto tests pass * apply Lysandre suggestion * finish tests * add some good tokenizer tests * improve docs slighly Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Arthur <48595927+ArthurZucker@users.noreply.github.com> Co-authored-by:
ArthurZucker <arthur.zucker@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-
Michael Benayoun authored
* Support tracing for ViT * Swin support * Fix copies * Fix type annotation issue * Removed unused import
-
Omar U. Espejel authored
* Fix contents in index.mdx to match docs' sidebar * Eliminates api section from contents
-
Omar Sanseviero authored
-
Omar U. Espejel authored
* Change nits in Spanish for quicktour.mdx - Add tasks names in English too. - Fix small nits in Spanish * Translate index.mdx to Spanish * Translate body of index. * Translated the compatible models list (not the papers´ names). Since this should not be updated manually, I can come back to the original text. * Add models and a dataset for Spanish in the code exmaples * Replaced the English models to Spanish versions. * Add index to _toctree.yml and fix Spanish * Fix double ““ error * Change negative example in ASR example * make style * Debug style in quicktour.mdx
-
Jorge Loayza R authored
* Spanish translation of fast_tokenizers.mdx * add fast_tokenizers to the spanish _toctree.yml * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/fast_tokenizers.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Joaq authored
* Spanish version of language_modeling.mdx doc file * modification to toctree.yml file * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/language_modeling.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Correct position of Guías conceptuales Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
jkmg authored
* adding philosophy.mdx translation to Spanish * adding philosophy.mdx translation to Spanish * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * Update docs/source/es/philosophy.mdx Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com> * philosophy translation to Spanish * Update _toctree.yml * Update _toctree.yml * nits Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
Shijie Wu authored
-
- 11 May, 2022 3 commits
-
-
Amanpreet Singh authored
* [WIP] Add FLAVA model This PR aims to add [FLAVA](ihttps://arxiv.org/abs/2112.04482) model to the transformers repo. Following checklist delineates the list of things to be done for this PR to be complete: [x] Flava init [x] Flava base models [x] Flava layers [x] Flava Configs [x] Flava encoders [x] Flava pretraining models [ ] Flava classification/retrieval models (To be added in a separate PR) [x] Documentation updates [x] Imports updates [x] Argstring updates [x] Flava pretrained checkpoints [x] Flava tests [x] Flava processors [x] Sanity check [x] Lint
-
Antoni Baum authored
-
Suraj Patil authored
-