Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
chenpangpang
parler-tts
Commits
0e5f2734
"configs/vscode:/vscode.git/clone" did not exist on "8bf2f5a453a66513fb36cb09f66ff7e8dbd9086e"
Commit
0e5f2734
authored
Apr 09, 2024
by
Yoach Lacombe
Browse files
remove useless parameters from training config
parent
b10e5625
Changes
6
Hide whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
10 additions
and
12 deletions
+10
-12
helpers/model_init_scripts/init_dummy_model.py
helpers/model_init_scripts/init_dummy_model.py
+3
-0
helpers/model_init_scripts/init_model_300M.py
helpers/model_init_scripts/init_model_300M.py
+3
-0
helpers/training_configs/librispeech_tts_r_300M_dummy.json
helpers/training_configs/librispeech_tts_r_300M_dummy.json
+0
-3
helpers/training_configs/starting_point_0.01.json
helpers/training_configs/starting_point_0.01.json
+0
-3
training/TRAINING.md
training/TRAINING.md
+0
-2
training/run_parler_tts_training.py
training/run_parler_tts_training.py
+4
-4
No files found.
helpers/model_init_scripts/init_dummy_model.py
View file @
0e5f2734
...
@@ -60,5 +60,8 @@ if __name__ == "__main__":
...
@@ -60,5 +60,8 @@ if __name__ == "__main__":
model
.
generation_config
.
max_length
=
int
(
30
*
model
.
audio_encoder
.
config
.
frame_rate
)
model
.
generation_config
.
max_length
=
int
(
30
*
model
.
audio_encoder
.
config
.
frame_rate
)
model
.
generation_config
.
do_sample
=
True
# True
model
.
generation_config
.
do_sample
=
True
# True
model
.
generation_config
.
guidance_scale
=
1
# 3.0
model
.
generation_config
.
guidance_scale
=
1
# 3.0
model
.
config
.
pad_token_id
=
encodec_vocab_size
model
.
config
.
decoder_start_token_id
=
encodec_vocab_size
+
1
model
.
save_pretrained
(
os
.
path
.
join
(
args
.
save_directory
,
"tiny-model"
))
model
.
save_pretrained
(
os
.
path
.
join
(
args
.
save_directory
,
"tiny-model"
))
helpers/model_init_scripts/init_model_300M.py
View file @
0e5f2734
...
@@ -60,5 +60,8 @@ if __name__ == "__main__":
...
@@ -60,5 +60,8 @@ if __name__ == "__main__":
model
.
generation_config
.
max_length
=
int
(
30
*
model
.
audio_encoder
.
config
.
frame_rate
)
model
.
generation_config
.
max_length
=
int
(
30
*
model
.
audio_encoder
.
config
.
frame_rate
)
model
.
generation_config
.
do_sample
=
True
# True
model
.
generation_config
.
do_sample
=
True
# True
model
.
generation_config
.
guidance_scale
=
1
# 3.0
model
.
generation_config
.
guidance_scale
=
1
# 3.0
model
.
config
.
pad_token_id
=
encodec_vocab_size
model
.
config
.
decoder_start_token_id
=
encodec_vocab_size
+
1
model
.
save_pretrained
(
os
.
path
.
join
(
args
.
save_directory
,
"parler-tts-untrained-300M/"
))
model
.
save_pretrained
(
os
.
path
.
join
(
args
.
save_directory
,
"parler-tts-untrained-300M/"
))
helpers/training_configs/librispeech_tts_r_300M_dummy.json
View file @
0e5f2734
...
@@ -37,9 +37,6 @@
...
@@ -37,9 +37,6 @@
"preprocessing_num_workers"
:
8
,
"preprocessing_num_workers"
:
8
,
"pad_token_id"
:
1024
,
"decoder_start_token_id"
:
1025
,
"do_train"
:
true
,
"do_train"
:
true
,
"num_train_epochs"
:
50
,
"num_train_epochs"
:
50
,
"gradient_accumulation_steps"
:
1
,
"gradient_accumulation_steps"
:
1
,
...
...
helpers/training_configs/starting_point_0.01.json
View file @
0e5f2734
...
@@ -39,9 +39,6 @@
...
@@ -39,9 +39,6 @@
"preprocessing_num_workers"
:
8
,
"preprocessing_num_workers"
:
8
,
"pad_token_id"
:
1024
,
"decoder_start_token_id"
:
1025
,
"do_train"
:
true
,
"do_train"
:
true
,
"num_train_epochs"
:
40
,
"num_train_epochs"
:
40
,
"gradient_accumulation_steps"
:
1
,
"gradient_accumulation_steps"
:
1
,
...
...
training/TRAINING.md
View file @
0e5f2734
...
@@ -124,8 +124,6 @@ accelerate launch ./training/run_parler_tts_training.py \
...
@@ -124,8 +124,6 @@ accelerate launch ./training/run_parler_tts_training.py \
--add_audio_samples_to_wandb
true
\
--add_audio_samples_to_wandb
true
\
--id_column_name
"id"
\
--id_column_name
"id"
\
--preprocessing_num_workers
8
\
--preprocessing_num_workers
8
\
--pad_token_id
1024
\
--decoder_start_token_id
1025
\
--do_train
true
\
--do_train
true
\
--num_train_epochs
50
\
--num_train_epochs
50
\
--gradient_accumulation_steps
1
\
--gradient_accumulation_steps
1
\
...
...
training/run_parler_tts_training.py
View file @
0e5f2734
...
@@ -235,11 +235,11 @@ class ModelArguments:
...
@@ -235,11 +235,11 @@ class ModelArguments:
metadata
=
{
"help"
:
"Whether to freeze the text encoder."
},
metadata
=
{
"help"
:
"Whether to freeze the text encoder."
},
)
)
do_sample
:
bool
=
field
(
do_sample
:
bool
=
field
(
default
=
Fals
e
,
default
=
Tru
e
,
metadata
=
{
"help"
:
"Whether to do sampling or greedy decoding."
},
metadata
=
{
"help"
:
"Whether to do sampling or greedy decoding."
},
)
)
temperature
:
float
=
field
(
temperature
:
float
=
field
(
default
=
0.4
,
default
=
1.0
,
metadata
=
{
"help"
:
"Temperature if sampling."
},
metadata
=
{
"help"
:
"Temperature if sampling."
},
)
)
max_length
:
int
=
field
(
max_length
:
int
=
field
(
...
@@ -1018,10 +1018,10 @@ def main():
...
@@ -1018,10 +1018,10 @@ def main():
{
{
"pad_token_id"
:
model_args
.
pad_token_id
"pad_token_id"
:
model_args
.
pad_token_id
if
model_args
.
pad_token_id
is
not
None
if
model_args
.
pad_token_id
is
not
None
else
model
.
config
.
pad_token_id
,
else
config
.
pad_token_id
,
"decoder_start_token_id"
:
model_args
.
decoder_start_token_id
"decoder_start_token_id"
:
model_args
.
decoder_start_token_id
if
model_args
.
decoder_start_token_id
is
not
None
if
model_args
.
decoder_start_token_id
is
not
None
else
model
.
config
.
decoder_start_token_id
,
else
config
.
decoder_start_token_id
,
}
}
)
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment