Unverified Commit 169d8c7f authored by Jaylin Lee's avatar Jaylin Lee Committed by GitHub
Browse files

[bugfix] Fix some docs' bug in 'serving' (#109)

* [bugfix] Fix some docs' bug in 'serving'

* [bugfix] Fix some docs' bug in 'serving'
parent b6bb8ce2
......@@ -93,13 +93,13 @@ bash workspace/service_docker_up.sh
Then, you can communicate with the inference server by command line,
```shell
python3 lmdeploy.serve.client {server_ip_addresss}:33337 internlm
python3 -m lmdeploy.serve.client {server_ip_addresss}:33337 internlm
```
or webui,
```
python3 lmdeploy.app {server_ip_addresss}:33337 internlm
python3 -m lmdeploy.app {server_ip_addresss}:33337 internlm
```
![](https://github.com/InternLM/lmdeploy/assets/67539920/08d1e6f2-3767-44d5-8654-c85767cec2ab)
......
......@@ -92,13 +92,13 @@ bash workspace/service_docker_up.sh
你可以通过命令行方式与推理服务进行对话:
```shell
python3 lmdeploy.serve.client {server_ip_addresss}:33337 internlm
python3 -m lmdeploy.serve.client {server_ip_addresss}:33337 internlm
```
也可以通过 WebUI 方式来对话:
```
python3 lmdeploy.app {server_ip_addresss}:33337 internlm
python3 -m lmdeploy.app {server_ip_addresss}:33337 internlm
```
![](https://github.com/InternLM/lmdeploy/assets/67539920/08d1e6f2-3767-44d5-8654-c85767cec2ab)
......
......@@ -60,7 +60,7 @@ python3 -m fastchat.model.apply_delta \
--target-model-path /path/to/vicuna-7b \
--delta-path lmsys/vicuna-7b-delta-v1.1
python3 lmdeploy/serve/turbomind/deploy.py vicuna-7B /path/to/vicuna-7b hf
python3 -m lmdeploy.serve.turbomind.deploy vicuna-7B /path/to/vicuna-7b hf
bash workspace/service_docker_up.sh
```
......@@ -76,7 +76,7 @@ python3 -m fastchat.model.apply_delta \
--target-model-path /path/to/vicuna-13b \
--delta-path lmsys/vicuna-13b-delta-v1.1
python3 lmdeploy/serve/turbomind/deploy.py vicuna-13B /path/to/vicuna-13b hf
python3 -m lmdeploy.serve.turbomind.deploy vicuna-13B /path/to/vicuna-13b hf
bash workspace/service_docker_up.sh
```
......
......@@ -8,7 +8,7 @@
<summary><b>7B</b></summary>
```shell
python3 lmdeploy.serve.turbomind.deploy llama-7B /path/to/llama-7b llama \
python3 -m lmdeploy.serve.turbomind.deploy llama-7B /path/to/llama-7b llama \
--tokenizer_path /path/to/tokenizer/model
bash workspace/service_docker_up.sh
```
......@@ -19,7 +19,7 @@ bash workspace/service_docker_up.sh
<summary><b>13B</b></summary>
```shell
python3 lmdeploy.serve.turbomind.deploy llama-13B /path/to/llama-13b llama \
python3 -m lmdeploy.serve.turbomind.deploy llama-13B /path/to/llama-13b llama \
--tokenizer_path /path/to/tokenizer/model --tp 2
bash workspace/service_docker_up.sh
```
......@@ -30,7 +30,7 @@ bash workspace/service_docker_up.sh
<summary><b>30B</b></summary>
```shell
python3 lmdeploy.serve.turbomind.deploy llama-32B /path/to/llama-30b llama \
python3 -m lmdeploy.serve.turbomind.deploy llama-32B /path/to/llama-30b llama \
--tokenizer_path /path/to/tokenizer/model --tp 4
bash workspace/service_docker_up.sh
```
......@@ -41,7 +41,7 @@ bash workspace/service_docker_up.sh
<summary><b>65B</b></summary>
```shell
python3 lmdeploy.serve.turbomind.deploy llama-65B /path/to/llama-65b llama \
python3 -m lmdeploy.serve.turbomind.deploy llama-65B /path/to/llama-65b llama \
--tokenizer_path /path/to/tokenizer/model --tp 8
bash workspace/service_docker_up.sh
```
......@@ -60,7 +60,7 @@ python3 -m fastchat.model.apply_delta \
--target-model-path /path/to/vicuna-7b \
--delta-path lmsys/vicuna-7b-delta-v1.1
python3 lmdeploy/serve/turbomind/deploy.py vicuna-7B /path/to/vicuna-7b hf
python3 -m lmdeploy.serve.turbomind.deploy vicuna-7B /path/to/vicuna-7b hf
bash workspace/service_docker_up.sh
```
......@@ -76,7 +76,7 @@ python3 -m fastchat.model.apply_delta \
--target-model-path /path/to/vicuna-13b \
--delta-path lmsys/vicuna-13b-delta-v1.1
python3 lmdeploy/serve/turbomind/deploy.py vicuna-13B /path/to/vicuna-13b hf
python3 -m lmdeploy.serve.turbomind.deploy vicuna-13B /path/to/vicuna-13b hf
bash workspace/service_docker_up.sh
```
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment