Unverified Commit d5239bab authored by Mikel Penagarikano's avatar Mikel Penagarikano Committed by GitHub
Browse files

Sync preprocesses before loading the processor at run_speech_recognition_ctc.py (#21926)

* Update run_speech_recognition_ctc.py

Make sure all processes wait until data is saved before loading the processor from the output_dit

* Make sure all processes wait until data is saved before loading the processor from the output_dit

* Update run_speech_recognition_ctc.py

* Update run_speech_recognition_seq2seq.py
parent f49b0762
...@@ -673,11 +673,14 @@ def main(): ...@@ -673,11 +673,14 @@ def main():
return metrics return metrics
# Now save everything to be able to create a single processor later # Now save everything to be able to create a single processor later
if is_main_process(training_args.local_rank): # make sure all processes wait until data is saved
# save feature extractor, tokenizer and config with training_args.main_process_first():
feature_extractor.save_pretrained(training_args.output_dir) # only the main process saves them
tokenizer.save_pretrained(training_args.output_dir) if is_main_process(training_args.local_rank):
config.save_pretrained(training_args.output_dir) # save feature extractor, tokenizer and config
feature_extractor.save_pretrained(training_args.output_dir)
tokenizer.save_pretrained(training_args.output_dir)
config.save_pretrained(training_args.output_dir)
try: try:
processor = AutoProcessor.from_pretrained(training_args.output_dir) processor = AutoProcessor.from_pretrained(training_args.output_dir)
......
...@@ -506,11 +506,14 @@ def main(): ...@@ -506,11 +506,14 @@ def main():
return {"wer": wer} return {"wer": wer}
# 9. Create a single speech processor # 9. Create a single speech processor
if is_main_process(training_args.local_rank): # make sure all processes wait until data is saved
# save feature extractor, tokenizer and config with training_args.main_process_first():
feature_extractor.save_pretrained(training_args.output_dir) # only the main process saves them
tokenizer.save_pretrained(training_args.output_dir) if is_main_process(training_args.local_rank):
config.save_pretrained(training_args.output_dir) # save feature extractor, tokenizer and config
feature_extractor.save_pretrained(training_args.output_dir)
tokenizer.save_pretrained(training_args.output_dir)
config.save_pretrained(training_args.output_dir)
processor = AutoProcessor.from_pretrained(training_args.output_dir) processor = AutoProcessor.from_pretrained(training_args.output_dir)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment