# Text-To-Audio The `stabilityai/stable-audio-open-1.0` pipeline generates audio from text prompts. ## Prerequisites If you use a gated model (e.g., `stabilityai/stable-audio-open-1.0`), ensure you have access: 1. **Accept Model License**: Visit the model page on Hugging Face (e.g., [stabilityai/stable-audio-open-1.0]) and accept the user agreement. 2. **Authenticate**: Log in to Hugging Face locally to access the gated model. ```bash huggingface-cli login ``` ## Local CLI Usage ```bash python text_to_audio.py \ --model stabilityai/stable-audio-open-1.0 \ --prompt "The sound of a hammer hitting a wooden surface" \ --negative_prompt "Low quality" \ --seed 42 \ --guidance_scale 7.0 \ --audio_length 10.0 \ --num_inference_steps 100 \ --output stable_audio_output.wav ``` Key arguments: - `--prompt`: text description (string). - `--negative_prompt`: negative prompt for classifier-free guidance. - `--seed`: integer seed for deterministic generation. - `--guidance_scale`: classifier-free guidance scale. - `--audio_length`: audio duration in seconds. - `--num_inference_steps`: diffusion sampling steps.(more steps = higher quality, slower). - `--output`: path to save the generated WAV file.