- 23 Mar, 2022 1 commit
-
-
Lysandre Debut authored
* Updates the default branch from master to main * Links from `master` to `main` * Typo * Update examples/flax/README.md Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 12 Mar, 2022 1 commit
-
-
Stas Bekman authored
* [WIP] add support for bf16 mode * prep for bf16 * prep for bf16 * fix; zero2/bf16 is ok * check bf16 is available * test fixes * enable zero3_bf16 * config files * docs * split stage_dtype; merge back to non-dtype-specific config file * fix doc * cleanup * cleanup * bfloat16 => bf16 to match the PR changes * s/zero_gather_fp16_weights_on_model_save/zero_gather_16bit_weights_on_model_save/; s/save_fp16_model/save_16bit_model/ * test fixes/skipping * move * fix * Update docs/source/main_classes/deepspeed.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * backticks * cleanup * cleanup * cleanup * new version * add note about grad accum in bf16 Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 09 Feb, 2022 1 commit
-
-
Lysandre Debut authored
* Upgrade black to version ~=22.0 * Check copies * Fix code
-
- 10 Jan, 2022 1 commit
-
-
Patrick von Platen authored
* up * up * up * up * up * up * improve * up * up * Update src/transformers/trainer.py * up * up * up
-
- 11 Nov, 2021 1 commit
-
-
Stas Bekman authored
-
- 22 Oct, 2021 1 commit
-
-
Antonio Carlos Falc茫o Petri authored
-
- 14 Oct, 2021 1 commit
-
-
Patrick von Platen authored
-
- 08 Aug, 2021 1 commit
-
-
Patrick von Platen authored
-
- 30 Jul, 2021 1 commit
-
-
21jun authored
help for `ModelArguments.gradient_checkpointing` should be "If True, use gradient checkpointing to save memory at the expense of slower backward pass." not "Whether to freeze the feature extractor layers of the model." (which is duplicated from `freeze_feature_extractor` arg)
-
- 23 Jul, 2021 1 commit
-
-
Stas Bekman authored
-
- 15 Jul, 2021 1 commit
-
-
Patrick von Platen authored
* fix_torch_device_generate_test * remove @ * start adding tests * correct wav2vec2 pretraining * up * up Co-authored-by:Patrick von Platen <patrick@huggingface.co>
-
- 25 Jun, 2021 1 commit
-
-
Stas Bekman authored
-
- 14 Jun, 2021 1 commit
-
-
Stas Bekman authored
* consistent nn. and nn.functional: p4 examples * restore
-
- 09 Jun, 2021 2 commits
-
-
Anton Lozhkov authored
* Working quantizer forward * Working quantizer forward * Clean up unused model parts, test reproducibility * Working quantizer forward * Clean up unused model parts, test reproducibility * Remove custom outputs from the shared ones * correct conversion * correct bug * add first pretrain script * save intermediate * static shapes * save intermediate * finish first pretrain script version * more refactor * remove wanddb * refactor more * improve test * correct perplexity compute bug * finish model implementation * add to docs * finish docs * finish pretraining script * finish pretraining script * remove wandb * finish PR for merge * finish config * finish * make deepspeed work * Apply suggestions from code review Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * apply suggestions * fix flaky test Co-authored-by:
patrickvonplaten <patrick.v.platen@gmail.com> Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Stas Bekman authored
-
- 08 Jun, 2021 1 commit
-
-
Stas Bekman authored
* wip * wip - but working with https://github.com/microsoft/DeepSpeed/pull/1044 * cleanup * workaround * working 5/8 modes * solve fp32 distributed zero3 * style * sync * sync * rework * deprecation * cleanup * https://github.com/microsoft/DeepSpeed/pull/1044 pr was merged * clean up * add a guide * more prose * more prose * fix * more prose * sub_group_size was too big * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * refactor * bug fix * make the true check explicit * new deepspeed release Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
- 12 May, 2021 1 commit
-
-
Philip May authored
-
- 14 Apr, 2021 1 commit
-
-
Nithin Holla authored
Co-authored-by:nithin19 <nithin@amberscript.com>
-
- 30 Mar, 2021 1 commit
-
-
Yih-Dar authored
-
- 22 Mar, 2021 2 commits
-
-
Qiushi Pan authored
Fix typo.
-
Patrick von Platen authored
-
- 21 Mar, 2021 4 commits
-
-
Suraj Patil authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Suraj Patil authored
-
- 19 Mar, 2021 3 commits
-
-
Julien Chaumond authored
* wording/typos tweaks * Make model upload instructions simpler
-
Patrick von Platen authored
-
Patrick von Platen authored
* finish * fix * fix * fix * fix
-
- 18 Mar, 2021 6 commits
-
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
-
Patrick von Platen authored
* upload * upload fine-tuning script * improve * adapt * Apply suggestions from code review * correct * upload * finalize * remove @ * correct typos
-
Suraj Patil authored
* add initial script * finish script * add shell script example * accept chars_to_ignor as cl arg * align the script with other example scripts * add torchaudio dep
-
Mohamed El-Geish authored
* wav2vec2: support datasets other than LibriSpeech * Formatting run_asr.py to pass code quality test * bundled orthography options and added verbose logs * fixing a typo in timit fine-tuning script * update comment for clarity * resize_lm_head and load custom vocab from file * adding a max_duration_in_seconds filter * do not assign `duration_filter` lambda, use a def * log untransliterated text as well * fix base model for arabic * fix duration filter when target_sr is not set * drop duration_in_seconds when unneeded * script for wav2vec2-large-lv60-timit-asr * fix for "tha" in arabic corpus (huggingface#10581) * adding more options to work with common_voice * PR feedback (huggingface#10581) * small README change
-
- 05 Mar, 2021 1 commit
-
-
Patrick von Platen authored
-
- 01 Mar, 2021 1 commit
-
-
Patrick von Platen authored
* add encode labels function to tokenizer * start adding finetuning * init dropout * upload * correct convert script * apply changes * fix second typo * make first dummy training run * adapt convert script * push confg for comparison * remove conf * finish training * adapt data collator * add research folder * update according to fairseq feedback * some minor corrections * refactor masking indices a bit * some minor changes * clean tokenizer * finish clean-up * remove previous logic * update run script * correct training * finish changes * finish model * correct bug * fix training a bit more * add some tests * finish gradient checkpointing * finish example * correct gradient checkpointing * improve tokenization method * revert changes in tokenizer * revert general change * adapt fine-tuning * update * save intermediate test * Update README.md * finish finetuning * delete conversion script * Update src/transformers/models/wav2vec2/configuration_wav2vec2.py * Update src/transformers/models/wav2vec2/processing_wav2vec2.py Co-authored-by:
Lysandre Debut <lysandre@huggingface.co> * finish wav2vec2 script * finish wav2vec2 fine-tuning * finalize test * correct test * adapt tests * finish * remove test file Co-authored-by:
Lysandre Debut <lysandre@huggingface.co>
-