- 06 Feb, 2023 3 commits
-
-
Jinen Setpal authored
updated documentation
-
jianan-gu authored
* Update perf_train_cpu.mdx * Update perf_train_cpu.mdx * Update perf_train_cpu.mdx * Update docs/source/en/perf_train_cpu.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update perf_train_cpu.mdx * Update perf_train_cpu.mdx * Update perf_train_cpu.mdx * Update perf_train_cpu.mdx --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Yih-Dar authored
* fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 03 Feb, 2023 11 commits
-
-
Yih-Dar authored
* fix * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
agossard authored
For IterableDataset, return DataLoader using self._train_batch_size. This is consistent with how we generate a regular DataLoader, and leads to the correct args.per_device_train_batch_size eventually ending up on each GPU.
-
Matt authored
* Add tutorial doc for TF + TPU * Fix all those extra asterisks in the markdown * Use the actual Tip formatting * Remove unnecessary spaces * Reformat checklist * Fix checklist and reformat tips slightly * Update docs/source/en/perf_train_tpu_tf.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/perf_train_tpu_tf.mdx Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Update docs/source/en/perf_train_tpu_tf.mdx Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Update docs/source/en/perf_train_tpu_tf.mdx Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com> * Add link to TPU notebook in the notebooks list * Add links to the TPU notebook in the tutorial doc * Make the markdown table a bit less wild * Fix notebook link * More notebook links * More fixes to wild tables --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> Co-authored-by:
Sayak Paul <spsayakpaul@gmail.com>
-
Darren Tuit authored
exclude deleted files from fixup script
-
Matthijs Hollemans authored
* make SpeechT5 model by copying Wav2Vec2 * add paper to docs * whoops added docs in wrong file * remove SpeechT5Tokenizer + put CTC back in the name * remove deprecated class * remove unused docstring * delete SpeechT5FeatureExtractor, use Wav2Vec2FeatureExtractor instead * remove classes we don't need right now * initial stab at speech encoder prenet * add more speech encoder prenet stuff * improve SpeechEncoderPrenet * add encoder (not finished yet) * add relative position bias to self-attention * add encoder CTC layers * fix formatting * add decoder from BART, doesn't work yet * make it work with generate loop * wrap the encoder into a speech encoder class * wrap the decoder in a text decoder class * changed my mind * changed my mind again ;-) * load decoder weights, make it work * add weights for text decoder postnet * add SpeechT5ForCTC model that uses only the encoder * clean up EncoderLayer and DecoderLayer * implement _init_weights in SpeechT5PreTrainedModel * cleanup config + Encoder and Decoder * add head + cross attention masks * improve doc comments * fixup * more cleanup * more fixup * TextDecoderPrenet works now, thanks Kendall * add CTC loss * add placeholders for other pre/postnets * add type annotation * fix freeze_feature_encoder * set padding tokens to 0 in decoder attention mask * encoder attention mask downsampling * remove features_pen calculation * disable the padding tokens thing again * fixup * more fixup * code review fixes * rename encoder/decoder wrapper classes * allow checkpoints to be loaded into SpeechT5Model * put encoder into wrapper for CTC model * clean up conversion script * add encoder for TTS model * add speech decoder prenet * add speech decoder post-net * attempt to reconstruct the generation loop * add speech generation loop * clean up generate_speech * small tweaks * fix forward pass * enable always dropout on speech decoder prenet * sort declaration * rename models * fixup * fix copies * more fixup * make consistency checker happy * add Seq2SeqSpectrogramOutput class * doc comments * quick note about loss and labels * add HiFi-GAN implementation (from Speech2Speech PR) * rename file * add vocoder to TTS model * improve vocoder * working on tokenizer * more better tokenizer * add CTC tokenizer * fix decode and batch_code in CTC tokenizer * fix processor * two processors and feature extractors * use SpeechT5WaveformFeatureExtractor instead of Wav2Vec2 * cleanup * more cleanup * even more fixup * notebooks * fix log-mel spectrograms * support reduction factor * fixup * shift spectrograms to right to create decoder inputs * return correct labels * add labels for stop token prediction * fix doc comments * fixup * remove SpeechT5ForPreTraining * more fixup * update copyright headers * add usage examples * add SpeechT5ProcessorForCTC * fixup * push unofficial checkpoints to hub * initial version of tokenizer unit tests * add slow test * fix failing tests * tests for CTC tokenizer * finish CTC tokenizer tests * processor tests * initial test for feature extractors * tests for spectrogram feature extractor * fixup * more fixup * add decorators * require speech for tests * modeling tests * more tests for ASR model * fix imports * add fake tests for the other models * fixup * remove jupyter notebooks * add missing SpeechT5Model tests * add missing tests for SpeechT5ForCTC * add missing tests for SpeechT5ForTextToSpeech * sort tests by name * fix Hi-Fi GAN tests * fixup * add speech-to-speech model * refactor duplicate speech generation code * add processor for SpeechToSpeech model * add usage example * add tests for speech-to-speech model * fixup * enable gradient checkpointing for SpeechT5FeatureEncoder * code review * push_to_hub now takes repo_id * improve doc comments for HiFi-GAN config * add missing test * add integration tests * make number of layers in speech decoder prenet configurable * rename variable * rename variables * add auto classes for TTS and S2S * REMOVE CTC!!! * S2S processor does not support save/load_pretrained * fixup * these models are now in an auto mapping * fix doc links * rename HiFiGAN to HifiGan, remove separate config file * REMOVE auto classes * there can be only one * fixup * replace assert * reformat * feature extractor can process input and target at same time * update checkpoint names * fix commit hash
-
Kashif Rasul authored
* no dot scale gradient in bf16 mode * fix since args.fp16 might be none * fixed typo * typo * only do if grad scaling is true * self.amp_dtype == torch.float16 is true * put back prop when fsdp is not none
-
Yih-Dar authored
fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Avi Singhal authored
* updated resources for LayoutLM * Apply suggestions from code review Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com> * fixed formatting, removed extra section --------- Co-authored-by:
Steven Liu <59462357+stevhliu@users.noreply.github.com>
-
Yih-Dar authored
* * Remove unused type_vocab_size * Remove unused initializer_factor * Remove unused n_embd * Remove unused scale_embedding * Remove unused scale_attn_weights * fix * fix * Remove unused head_hidden_scale * Remove unused activation_dropout --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Pavel Denisov authored
Add accepting `.generate()` calls with `inputs_embeds` on BLOOM models
-
Joao Gante authored
-
- 02 Feb, 2023 12 commits
-
-
Erwann Millon authored
* Add VQGAN-CLIP research project * fixed style issues * Update examples/research_projects/vqgan-clip/README.md Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/VQGAN_CLIP.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/requirements.txt Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/README.md Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/VQGAN_CLIP.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/VQGAN_CLIP.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/VQGAN_CLIP.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * Update examples/research_projects/vqgan-clip/loaders.py Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * replace CLIPProcessor with tokenizer, change asserts to exceptions * rm unused import * remove large files (jupyter notebook linked in readme, imgs migrated to hf dataset) * add tokenizers dependency * Remove comment Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com> * rm model checkpoints --------- Co-authored-by:
Erwann Millon <erwann@Erwanns-MacBook-Air.local> Co-authored-by:
amyeroberts <22614925+amyeroberts@users.noreply.github.com>
-
Steven Liu authored
* first draft of audio section * make style * first draft of computer vision section * add convnext and encoder tasks * finish up nlp tasks * minor edits * add arch images, more edits * fix image links * apply sanchit feedback * model naming convention * apply niels vit feedback * replace detr for segmentation with mask2former * apply feedback * apply feedback
-
Jorge C. Gomes authored
input_ids_seq_length doesn't exist in the GenerationConfig, it exists as local variable in the function. Setting exponential_decay_length_penalty therefore results in an error: `AttributeError: 'GenerationConfig' object has no attribute 'input_ids_seq_length'` This simple change fixes this issue, and the exponential_decay_length_penalty works as expected.
-
Steven Liu authored
fix formatting
-
Yih-Dar authored
* fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Yih-Dar authored
* Allow to add more information * fix style --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Younes Belkada authored
* force `memory_efficient_backward=True` * enhancements - trainer support - add new flag * some changes - internal changes in `Trainer` - small refactor * make quality * Fixes - add new testing util - add new test - change test in Trainer * fix CI test * educate users on how to ft 8bit models * more checks * fix `logger` error * Apply suggestions from code review Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * adapt from review * fix * add comment * use return instead --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Clémentine Fourrier authored
* [FIX] path for Graphormer checkpoint * [FIX] Test suite for graphormer * [FIX] Update graphormer default num_classes
-
Joel Lamy-Poirier authored
* gelu_python_tanh * rename * Version check, add test * Pr comment
-
Matt authored
* Add distinct section names for PyTorch and TF * Remove extra space
-
Shikhar Tuli authored
Co-authored-by:Shreshth Tuli <shreshthtuli@gmail.com>
-
Yih-Dar authored
Use torch 1.13.1 in push/scheduled CI Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
- 01 Feb, 2023 7 commits
-
-
Joao Gante authored
-
amyeroberts authored
* TF image classification script * Update requirements * Fix up * Add tests * Update test fetcher Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * Fix directory path * Adding `zero-shot-object-detection` pipeline doctest. (#20274) * Adding `zero-shot-object-detection` pipeline doctest. * Remove nested_simplify. * Add generate kwargs to `AutomaticSpeechRecognitionPipeline` (#20952) * Add generate kwargs to AutomaticSpeechRecognitionPipeline * Add test for generation kwargs * Trigger CI * Data collator returns np * Update feature extractor -> image processor * Bug fixes - updates to reflect changes in API * Update flags to match PT & run faster * Update instructions - Maria's comment * Update examples/tensorflow/image-classification/README.md * Remove slow decorator --------- Co-authored-by:
Nicolas Patry <patry.nicolas@protonmail.com> Co-authored-by:
bofeng huang <bofenghuang7@gmail.com> Co-authored-by:
Sylvain Gugger <Sylvain.gugger@gmail.com>
-
Jinen Setpal authored
* integrated logger * bugifx * added data * bugfix * model + state artifacts should log * fixed paths * i lied, trying again * updated function call * typo this is painful :( what a stupid error * typo this is painful :( what a stupid error * pivoted to adding a directory * silly path bug * multiple experiments * migrated to getattr * syntax fix * syntax fix * fixed repo pointer * fixed path error * added dataset if dataloader is present, uploaded artifacts * variable in scope * removed unnecessary line * updated error type Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * trimmed unused variables, imports * style formatting * removed type conversion reliance Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com> * reverted accidental line deletion --------- Co-authored-by:
Sylvain Gugger <35901082+sgugger@users.noreply.github.com>
-
Sylvain Gugger authored
* Skip batches fast with Accelerate * remove debug statement * Hack seed reload at the right time * Reorganize RNG sync * Fix accelerate version comp
-
raghavanone authored
* Fix the input embeds issue with tests * Fix black and isort issue * Clean up tests * Add slow tag to the test introduced * Incorporate PR feedbacks
-
Maria Khalusova authored
moved LiLT under multimodal models
-
Patrick von Platen authored
* Bump onnx in /examples/research_projects/decision_transformer Bumps [onnx](https://github.com/onnx/onnx) from 1.11.0 to 1.13.0. - [Release notes](https://github.com/onnx/onnx/releases) - [Changelog](https://github.com/onnx/onnx/blob/main/docs/Changelog.md) - [Commits](https://github.com/onnx/onnx/compare/v1.11.0...v1.13.0 ) --- updated-dependencies: - dependency-name: onnx dependency-type: direct:production ... Signed-off-by:
dependabot[bot] <support@github.com> * adapt * finish * Update examples/research_projects/decision_transformer/requirements.txt * up * add tests * Apply suggestions from code review Co-authored-by:
Lucain <lucainp@gmail.com> Co-authored-by:
Pedro Cuenca <pedro@huggingface.co> * fix test --------- Signed-off-by:
dependabot[bot] <support@github.com> Co-authored-by:
dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Co-authored-by:
Lucain <lucainp@gmail.com> Co-authored-by:
Pedro Cuenca <pedro@huggingface.co>
-
- 31 Jan, 2023 7 commits
-
-
Yih-Dar authored
* fix Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
Joao Gante authored
Generate: fix TF XLA tests on models with `max_position_embeddings` or `max_target_positions` (#21389)
-
Yih-Dar authored
* remove unused classifier_dropout * remove unused dropout * remove unused pooler_fn * remove unnecessary is_encoder_decoder * remove unnecessary drop_rate * remove unused classifier_dropout * remove unused classifier_dropout * remove unused dropout * remove unused dropout * remove unused summary_* attributes * remove unused tie_word_embeddings * remove unused summary_* attributes * fix --------- Co-authored-by:ydshieh <ydshieh@users.noreply.github.com>
-
raghavanone authored
* Add support of backward_prefetch and forward_prefetch * Fix format issue * Fix isort issue * Fix doc style issue * Update src/transformers/trainer.py Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com> * Update src/transformers/training_args.py Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com> * Fix black issue * Fix doc-style issue * Make additional fsdp parameters into fsdp config * Fix black issue * Remove unused imports * Fix doc style issues * Incorporate PR feedbacks * Remove unused imports * Fix tests * Fix tests * Fix tests * Fix tests * Fix tests * Update src/transformers/training_args.py Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com> * Fix tests * Incorporate PR feedbacks * Incorporate PR feedbacks * Fix black issues --------- Co-authored-by:
Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
-
Quentin Lhoest authored
* simplify column_names in run_clm * simplify column_names in run_mlm * minor
-
NielsRogge authored
* Improve docs * Add DETA resources --------- Co-authored-by:Niels Rogge <nielsrogge@Nielss-MacBook-Pro.local>
-
regisss authored
Do not log the generation config for each iteration
-