- 14 Mar, 2022 27 commits
-
-
Kamal Raj authored
-
Joydeep Bhattacharjee authored
-
Sylvain Gugger authored
-
Martin Pan authored
-
Jacob Dineen authored
* clip typhinting #16059 * removed optional type annotations for dataclass in CLIPOutput * type annotation fixes per Rocket - Clip Torch
-
Lysandre Debut authored
* Runtime -> Devel * Torch before DeepSpeed
-
Kamal Raj authored
* Add missing type hints - ELECTRA TF * bool -> Optional[bool]
-
Dan Tegzes authored
* Add type hints for SqueezeBert PyTorch * fixed unused List err * style fixes
-
Hyeonsoo Lee authored
-
Bhavika Tekwani authored
* Add type hints for LukeModel * Add type hints for entitypairclassification * Remove blank space Co-authored-by:bhavika <bhavika@debian-BULLSEYE-live-builder-AMD64>
-
Michael Benayoun authored
* Can choose framework for ONNX export * Fix docstring
-
Pepijn Boers authored
* Add type hints for TFDistilBert * Update src/transformers/models/distilbert/modeling_tf_distilbert.py Co-authored-by:Matt <Rocketknight1@users.noreply.github.com>
-
Bhavika Tekwani authored
* First pass * Fixup * Fix broken tests * Make unpack_inputs the first decorator
-
Bhavika Tekwani authored
* Replace input_processing * move unpack_inputs
-
Yih-Dar authored
* Fix checkpoint name in docstring example Co-authored-by:
ydshieh <ydshieh@users.noreply.github.com> Co-authored-by:
Patrick von Platen <patrick.v.platen@gmail.com>
-
Kamal Raj authored
-
Kamal Raj authored
-
Yhary Arias authored
* Spanish translation of the file training.mdx * Settings - Spanish translation of the file training.mdx * Latest changes to the Spanish translation of the training.mdx file * Delete Hugging.mdx * Last changes to the training fil Espanish version * Latest modifications * Latest changes, document ready for PR * Nits Co-authored-by:
Yhary Arias <yharystefa@gmail.com> Co-authored-by:
Omar U. Espejel <espejelomar@gmail.com>
-
NielsRogge authored
Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
Merve Noyan authored
-
Minh Chien Vu authored
* Change unpacking of TF mobilebert inputs to use decorator * Move unpack_inputs as the top decorator * make fixup Co-authored-by:ChienVM <chien_vm@detomo.co.jp>
-
Yih-Dar authored
Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* Make TF pt-tf equivalence test more aggressive * Fix for TFConvNextModelTest and TFTransfoXLModelTest * fix kwargs for outputs * clean-up * Add docstring for check_outputs() * remove: need to rename encoder-decoder * clean-up * send PyTorch things to the correct device * Add back the accidentally removed test case in test_pt_tf_model_equivalence() * Fix: change to tuple before calling check_outputs() * Fix: tfo could be a list * use to_tuple() * allow tfo only to be tuple or tensor * allow tfo to be list or tuple for now + style change * minor fix * remove np.copy and update comments * tfo -> tf_output, same for pt * Add more detailed comment * remove the incorrect comment Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
tiedemann authored
Configuration `tied-embeddings-all` implies `tied-embeddings-src`
-
Sanchit Gandhi authored
Fix Loading of Flax(Speech)EncoderDecoderModel kwargs from PreTrained Encoder-Decoder Checkpoints (#16056) * Fix Loading of Flax(Speech)EncoderDecoderModel kwargs from PreTrained Encoder-Decoder Checkpoints * change wording
-
Omar Sanseviero authored
-
lewtun authored
* Make Camembert great again! * Add Camembert to TensorFlow ONNX tests
-
- 13 Mar, 2022 1 commit
-
-
Thomas Chaigneau authored
* Add missing type hints for all flavors of LayoutLMv2 PyTorch models. * Fixed return types and added type hints for LayoutLM. * Fix removed arguments which breaks tests.
-
- 12 Mar, 2022 6 commits
-
-
James Barry authored
-
João Gustavo A. Amorim authored
-
p-mishra1 authored
-
Abdelrhman-Hosny authored
-
Omar Sanseviero authored
-
Stas Bekman authored
* [WIP] add support for bf16 mode * prep for bf16 * prep for bf16 * fix; zero2/bf16 is ok * check bf16 is available * test fixes * enable zero3_bf16 * config files * docs * split stage_dtype; merge back to non-dtype-specific config file * fix doc * cleanup * cleanup * bfloat16 => bf16 to match the PR changes * s/zero_gather_fp16_weights_on_model_save/zero_gather_16bit_weights_on_model_save/; s/save_fp16_model/save_16bit_model/ * test fixes/skipping * move * fix * Update docs/source/main_classes/deepspeed.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * backticks * cleanup * cleanup * cleanup * new version * add note about grad accum in bf16 Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 11 Mar, 2022 6 commits
-
-
Jeff Rasley authored
* gather z3 params for new_lm_head * Update src/transformers/modeling_utils.py Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
Steven Liu authored
*
📝 first draft of audio/vision guides *✨ make fixup *🖍 fix typo *🖍 close parentheses *🖍 apply feedback *🖍 apply feedback, make fixup *🖍 more fixup for perceiver *🖍 apply feedback *✨ make fixup *🖍 fix data collator -
Yih-Dar authored
* fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Thomas Chaigneau authored
* Add missing type hints for all flavors of RoBERTa PyTorch models. * Fixed type hints for all classes and fixed return types.
-
Lysandre Debut authored
* Rebuild deepspeed * Apply suggestions from code review Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com> Co-authored-by:
Stas Bekman <stas00@users.noreply.github.com>
-
João Gustavo A. Amorim authored
-