Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
Qwen3-TTS_pytorch
Commits
be69ba06
"vscode:/vscode.git/clone" did not exist on "50cfbb726b26700d143ce5bb55c0b5e86de7a1e6"
Commit
be69ba06
authored
Mar 02, 2026
by
weishb
Browse files
update README.md
parent
bcdb7109
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
16 additions
and
7 deletions
+16
-7
README.md
README.md
+16
-7
No files found.
README.md
View file @
be69ba06
...
@@ -84,15 +84,12 @@ python test_model_12hz_base.py
...
@@ -84,15 +84,12 @@ python test_model_12hz_base.py
```
```
### vllm
### vllm
#### 单机推理
#### 单机推理
VoiceDesign
启动服务
```
bash
```
bash
#
以VoiceDesign为例子,CustomVoice和Voice Clone需要使用不同的模型
#
# serve启动
VLLM_USE_V1
=
0 python
-m
vllm.entrypoints.openai.api_server
--model
Qwen3-TTS/Qwen3-TTS-12Hz-1.7B-VoiceDesign
--served-model-name
qwen3-tts
--host
0.0.0.0
--port
8000
--trust-remote-code
--dtype
bfloat16
--disable-async-output-proc
VLLM_USE_V1
=
0 python
-m
vllm.entrypoints.openai.api_server
--model
Qwen3-TTS/Qwen3-TTS-12Hz-1.7B-VoiceDesign
--served-model-name
qwen3-tts
--host
0.0.0.0
--port
8000
--trust-remote-code
--dtype
bfloat16
--disable-async-output-proc
```
调用服务:
## client访问
```
VoiceDesign
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
-H
"Content-Type: application/json"
\
-H
"Content-Type: application/json"
\
-o
output.wav
\
-o
output.wav
\
...
@@ -111,8 +108,14 @@ curl -sS http://127.0.0.1:8000/v1/audio/speech \
...
@@ -111,8 +108,14 @@ curl -sS http://127.0.0.1:8000/v1/audio/speech \
},
},
"response_format":"wav"
"response_format":"wav"
}'
}'
```
CustomVoice
CustomVoice
```
bash
## serve启动
VLLM_USE_V1
=
0 python
-m
vllm.entrypoints.openai.api_server
--model
Qwen3-TTS/Qwen3-TTS-12Hz-1.7B-CustomVoice
--served-model-name
qwen3-tts
--host
0.0.0.0
--port
8000
--trust-remote-code
--dtype
bfloat16
--disable-async-output-proc
## client访问
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
-H
"Content-Type: application/json"
\
-H
"Content-Type: application/json"
\
-o
output_customvoice.wav
\
-o
output_customvoice.wav
\
...
@@ -132,8 +135,14 @@ curl -sS http://127.0.0.1:8000/v1/audio/speech \
...
@@ -132,8 +135,14 @@ curl -sS http://127.0.0.1:8000/v1/audio/speech \
},
},
"response_format":"wav"
"response_format":"wav"
}'
}'
```
Voice Clone
Voice Clone
```
bash
## serve启动
VLLM_USE_V1
=
0 python
-m
vllm.entrypoints.openai.api_server
--model
Qwen3-TTS/Qwen3-TTS-12Hz-1.7B-Base
--served-model-name
qwen3-tts
--host
0.0.0.0
--port
8000
--trust-remote-code
--dtype
bfloat16
--disable-async-output-proc
## client访问
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
curl
-sS
http://127.0.0.1:8000/v1/audio/speech
\
-H
"Content-Type: application/json"
\
-H
"Content-Type: application/json"
\
-o
output_clone_icl.wav
\
-o
output_clone_icl.wav
\
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment