- 15 Aug, 2024 2 commits
-
-
Yoach Lacombe authored
* Update README.md * Update README.md * Update README.md * update configs and readme * fix training and eval single gpus and long audios errors * fix error transcriptions none * fix trascription null wer * Update README.md * Update README.md --------- Co-authored-by: yoach@huggingface.co <Yoach Lacombe>
-
Yoach Lacombe authored
* Update README.md * Update README.md * Update README.md * update configs and readme * fix training and eval single gpus and long audios errors * fix error transcriptions none * fix trascription null wer --------- Co-authored-by: yoach@huggingface.co <Yoach Lacombe>
-
- 13 Aug, 2024 1 commit
-
-
UncleCode authored
-
- 08 Aug, 2024 1 commit
-
-
Yoach Lacombe authored
* bump version to v0.2 * adapt readme * Update README.md * update README * add inference tips + streamer class * update readme * Update README.md * Apply suggestions from code review Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update README * Apply suggestions from code review Co-authored-by:
Vaibhav Srivastav <vaibhavs10@gmail.com> --------- Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> Co-authored-by:
Vaibhav Srivastav <vaibhavs10@gmail.com>
-
- 07 Aug, 2024 2 commits
-
-
Yoach Lacombe authored
-
eustlb authored
* add rope * don't include padding in rope * possibly use cross-attn for prompt * fix rope * fix cross-attn * fix self-attn * fix dummy model * clean-up rope * first gqa implementation * fix wer eval * feat: add flash attention and spda * chore: add README for flash attention * chore: add benchmark script * chore: add benchmark attention approach * multi node and fix wer and fix compile * Update modeling_parler_tts.py * fix FA2, SDPA and add cross-attn MHA and attention type forcing * better cross_attention key values number of heads default + add training arguments for attn implementation * fix audio padding when torch compile or pad_to_max_length=True * correct multi node * make rope faster * fix encoder sdpa * fix training with cross attention + with FAZ * use fp32 as default model dtype + fix generation when using FA2 with autocast * remove redundant passes in generate + clean and fix attentions * fix edge case in WER evaluation when longform generation * better multi-node mapping and saving / add eval dataloader num workers * remove old benchmarks * faster audio encoding + checkpointing + fix generation step * unpin trfms * remove CFG * imports and constants Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * attention modifications to handle static cach Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * decoder layer modification to handle static cache Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * ParlerTTSPreTrainedModel modifs to handle static cache Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * ParlerTTSDecoder modifs to handle static cache Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * ParlerTTSModel + ParlerTTSForCausalLM modfis Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * ParlerTTSForConditionalGeneration modifs Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * decoder_attention_mask for static cache Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * create inputs_embeds early to have a good cache initialization Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * _get_cache method Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * init the cache Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * ensure good device Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * pin tfrms version Co-Authored-By:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> * fix attention_mask FA2 * remove unnecessary method * Update parler_tts/modeling_parler_tts.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * Update parler_tts/modeling_parler_tts.py Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com> * remove unnecessary imports * replace the hardcoded cache_position with a more elegant approach * make style * unpin transformers * pin transformers * pin torch * refactor + unpin torch * Update parler_tts/modeling_parler_tts.py Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> * update training script to match 11b209e1 * Update parler_tts/modeling_parler_tts.py Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> * ensure compatibility with trfms 4.43.3, changes taken from #31980 on trfms * fix input_ids_length * warning full attention mask creation * changes for training compatibility --------- Co-authored-by:
sanchit-gandhi <sanchit@huggingface.co> Co-authored-by:
Yoach Lacombe <yoach.lacombe@gmail.com> Co-authored-by:
Yoach Lacombe <52246514+ylacombe@users.noreply.github.com> Co-authored-by:
sang-nguyen-ts <sang.nguyen@trustingsocial.com> Co-authored-by: yoach@huggingface.co <Yoach Lacombe> Co-authored-by:
sang-nguyen-ts <sang-nguyen-ts@users.noreply.github.com> Co-authored-by:
Sanchit Gandhi <93869735+sanchit-gandhi@users.noreply.github.com>
-
- 31 Jul, 2024 1 commit
-
-
Yoach Lacombe authored
* add RoPe * don't include padding in rope * possibly use cross-attn for prompt * fix rope * fix cross-attn * fix self-attn * fix dummy model * clean-up rope * first gqa implementation * fix wer eval * feat: add flash attention and spda * chore: add README for flash attention * chore: add benchmark script * chore: add benchmark attention approach * multi node and fix wer and fix compile * Update modeling_parler_tts.py * fix FA2, SDPA and add cross-attn MHA and attention type forcing * better cross_attention key values number of heads default + add training arguments for attn implementation * fix audio padding when torch compile or pad_to_max_length=True * correct multi node * make rope faster * fix encoder sdpa * fix training with cross attention + with FAZ * use fp32 as default model dtype + fix generation when using FA2 with autocast * remove redundant passes in generate + clean and fix attentions * fix edge case in WER evaluation when longform generation * better multi-node mapping and saving / add eval dataloader num workers * remove old benchmarks * faster audio encoding + checkpointing + fix generation step * better eval + add right padding + fix eval loss compute * correct README * correct config docstrings * remove comment * make style --------- Co-authored-by:
sanchit-gandhi <sanchit@huggingface.co> Co-authored-by:
sang-nguyen-ts <sang.nguyen@trustingsocial.com> Co-authored-by: yoach@huggingface.co <Yoach Lacombe>
-
- 30 May, 2024 2 commits
-
-
Sanchit Gandhi authored
[eval] fix trigger for english normaliser
-
sanchit-gandhi authored
-
- 23 May, 2024 2 commits
-
-
Yoach Lacombe authored
Fix WER transcription
-
Yoach Lacombe authored
-
- 22 May, 2024 9 commits
-
-
Yoach Lacombe authored
Fix naming and add min new tokens
-
Yoach Lacombe authored
-
Yoach Lacombe authored
-
Yoach Lacombe authored
[Training] Small nits
-
Yoach Lacombe authored
[training] compute normalised wer
-
Yoach Lacombe authored
-
sanchit-gandhi authored
-
sanchit-gandhi authored
-
sanchit-gandhi authored
-
- 18 May, 2024 1 commit
-
-
Sanchit Gandhi authored
[fix] Add fixed sampling rate to feature extractor
-
- 14 May, 2024 7 commits
-
-
choihk authored
-
Yoach Lacombe authored
Pr/wauplin/18
-
Yoach Lacombe authored
-
Yoach Lacombe authored
-
Yoach Lacombe authored
-
Yoach Lacombe authored
add mps and xpu to examples
-
Yoach Lacombe authored
Simple re-organization of training script
-
- 09 May, 2024 4 commits
-
-
Sanchit Gandhi authored
[generation] use private greedy/sampling methods
-
sanchit-gandhi authored
-
sanchit-gandhi authored
-
sanchit-gandhi authored
-
- 30 Apr, 2024 4 commits
-
-
Yoach Lacombe authored
Add colab link for fine-tuning
-
Yoach Lacombe authored
-
Yoach Lacombe authored
-
Yoach Lacombe authored
-
- 25 Apr, 2024 2 commits
- 24 Apr, 2024 2 commits