- 14 Mar, 2022 1 commit
-
-
lewtun authored
* Make Camembert great again! * Add Camembert to TensorFlow ONNX tests
-
- 13 Mar, 2022 1 commit
-
-
Thomas Chaigneau authored
* Add missing type hints for all flavors of LayoutLMv2 PyTorch models. * Fixed return types and added type hints for LayoutLM. * Fix removed arguments which breaks tests.
-
- 12 Mar, 2022 6 commits
-
-
James Barry authored
-
João Gustavo A. Amorim authored
-
p-mishra1 authored
-
Abdelrhman-Hosny authored
-
Omar Sanseviero authored
-
Stas Bekman authored
* [WIP] add support for bf16 mode * prep for bf16 * prep for bf16 * fix; zero2/bf16 is ok * check bf16 is available * test fixes * enable zero3_bf16 * config files * docs * split stage_dtype; merge back to non-dtype-specific config file * fix doc * cleanup * cleanup * bfloat16 => bf16 to match the PR changes * s/zero_gather_fp16_weights_on_model_save/zero_gather_16bit_weights_on_model_save/; s/save_fp16_model/save_16bit_model/ * test fixes/skipping * move * fix * Update docs/source/main_classes/deepspeed.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * backticks * cleanup * cleanup * cleanup * new version * add note about grad accum in bf16 Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 11 Mar, 2022 18 commits
-
-
Jeff Rasley authored
* gather z3 params for new_lm_head * Update src/transformers/modeling_utils.py Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Steven Liu authored
*
📝 first draft of audio/vision guides *✨ make fixup *🖍 fix typo *🖍 close parentheses *🖍 apply feedback *🖍 apply feedback, make fixup *🖍 more fixup for perceiver *🖍 apply feedback *✨ make fixup *🖍 fix data collator -
Yih-Dar authored
* fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Thomas Chaigneau authored
* Add missing type hints for all flavors of RoBERTa PyTorch models. * Fixed type hints for all classes and fixed return types.
-
Lysandre Debut authored
* Rebuild deepspeed * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
João Gustavo A. Amorim authored
-
Steven Liu authored
*
📝 first draft *🖍 apply feedback *🖍 apply feedback -
Kevin Bondzio authored
* add possibility to softly regulate length when using sampling method in model.generate() function * fix test config, fix formatting * fix rag integration, fix docstyling * fix wrong docstring * change param to tuple, add test * fix old param in rag_model, remove unused import * change test according to new param * fix formatting * fix test case * fix doc style * move start_length calculation to Logitprocessor * add possibility to softly regulate length when using sampling method in model.generate() function * fix rag integration, fix docstyling * fix test config, fix formatting * change param to tuple, add test * fix old param in rag_model, remove unused import * add possibility to softly regulate length when using sampling method in model.generate() function * change param to tuple, add test * fix old param in rag_model, remove unused import * remove unused import * fix small errors * fix test * add possibility to softly regulate length when using sampling method in model.generate() function * fix test config, fix formatting * fix rag integration, fix docstyling * change param to tuple, add test * fix old param in rag_model, remove unused import * change test according to new param * fix test case * move start_length calculation to Logitprocessor * add possibility to softly regulate length when using sampling method in model.generate() function * fix rag integration, fix docstyling * fix test config, fix formatting * change param to tuple, add test * fix old param in rag_model, remove unused import * add possibility to softly regulate length when using sampling method in model.generate() function * fix test config, fix formatting * fix rag integration, fix docstyling * add possibility to softly regulate length when using sampling method in model.generate() function * fix rag integration, fix docstyling * change param to tuple, add test * fix old param in rag_model, remove unused import * fix small errors * Update src/transformers/generation_utils.py Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update src/transformers/generation_utils.py * Update src/transformers/generation_utils.py * fix docstring, add type ind model rag * fix docstrings * introduce seq_length variable for cleaner code * fix black formatting * add input_ids_seq_length to modeling_rag * add input_ids_seq_length to test * retrigger checks * retrigger checks Co-authored-by:
Kevin Bondzio <kev@AIM-LAP-02.local> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Kevin Bondzio <kev@AIM-LAP-02.fritz.box>
-
Patrick von Platen authored
-
feifang24 authored
-
Matt authored
* Adding type annotations for TFRoBERTa * Add type hints to TFRobertaModel too
-
Matt authored
* Add type annotations for BERT and copies * make fixup
-
Sylvain Gugger authored
-
Sylvain Gugger authored
-
Lysandre Debut authored
-
Funtowicz Morgan authored
* Remove assertion over possible activation functions * Same for TF and Flax
-
Sylvain Gugger authored
-
Yih-Dar authored
* fix name Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 10 Mar, 2022 14 commits
-
-
David S. Batista authored
-
lewtun authored
* Fix duplicate arguments passed to dummy inputs in ONNX export * Fix M2M100 ONNX config * Ensure we check PreTrained model only if torch is available * Remove TensorFlow tests for models without PyTorch parity
-
Suraj Patil authored
* support not sharing embeddings * update modeling * update tokenizer * fix conversion script * always use self.shared * boom boom * begin tests * update tests * fix resize_decoder_token_embeddings * address Patrick's comments * style * update conversion script * fix conversion script * fix tokenizer * better name target vocab * add integration test for tokenizer with two vocabs * style * address Patrick's comments * add integration test for model
-
Lysandre Debut authored
* Support for torch 1.11 * Address Sylvain's comment
-
Sanchit Gandhi authored
* Fix Bug in Flax Seq2Seq Models * incorporate suggested changes
-
Joao Gante authored
* MVP * apply decorator to TFBertModel * finish updating bert * update rembert (copy-linked to bert) * update roberta (copy-linked to bert); Fix args * Now working for non-text modalities
-
Sylvain Gugger authored
-
Sylvain Gugger authored
* Build the doc in a seperate folder then move it * Allow job * Is this it? * Dislike comments? * Copy instead of move * Removing version built * Typos * No variable * Take _versions.yml into account * Finish main job and add dev job * Forgot the run * Fix syntax error * Execute builder from the repo * Typo
-
Yih-Dar authored
* fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Sanchit Gandhi authored
* Fix Bug in Flax-Speech-Encoder-Decoder Test * change thresholds for CPU precision
-
Suraj Patil authored
-
NielsRogge authored
* Add attentions_option to common tester * Fix tests, apply suggestion * Apply suggestion from code review Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Patrick von Platen authored
* Move generate docs * up * Update docs/source/_toctree.yml * correct * correct some stuff * correct tests * more fixes * finish generate * add to doc stest * finish * finalize * add warning to generate method
-
André Storhaug authored
"uvicorn" is misspelled as "unicorn".
-