Commit d765e83e authored by chenych's avatar chenych
Browse files

Add train loss

parent 0179c9b5
...@@ -89,7 +89,7 @@ python infer_vllm.py --model_name mistralai/Ministral-8B-Instruct-2410 ...@@ -89,7 +89,7 @@ python infer_vllm.py --model_name mistralai/Ministral-8B-Instruct-2410
1. 启动server 1. 启动server
```bash ```bash
vllm serve mistralai/Ministral-8B-Instruct-2410 --tokenizer_mode mistral --config_format mistral --load_format mistral vllm serve mistralai/Ministral-8B-Instruct-2410 --tokenizer-mode mistral --config-format mistral --load-format mistral
``` ```
2. client测试: 2. client测试:
...@@ -110,12 +110,20 @@ curl --location 'http://<your-node-url>:8000/v1/chat/completions' \ ...@@ -110,12 +110,20 @@ curl --location 'http://<your-node-url>:8000/v1/chat/completions' \
``` ```
## result ## result
Prompt: "Do we need to think for 10 seconds to find the answer of 1 + 1?"
<div align=center> <div align=center>
<img src="./doc/results.png"/> <img src="./doc/results.png"/>
</div> </div>
### 精度 ### 精度
DCU与GPU精度一致,推理框架:pytorch。 训练框架:[Llama-Factory](https://developer.sourcefind.cn/codes/OpenDAS/llama-factory)
训练脚本:[ministral_lora_sft.yaml](llama-factory/train_lora/ministral_lora_sft.yaml)
| device | train_loss |
|:----------:|:-------:|
| DCU K100AI | 0.9068 |
| GPU A800 | |
## 应用场景 ## 应用场景
### 算法类别 ### 算法类别
......
...@@ -17,12 +17,13 @@ preprocessing_num_workers: 16 ...@@ -17,12 +17,13 @@ preprocessing_num_workers: 16
dataloader_num_workers: 4 dataloader_num_workers: 4
### output ### output
output_dir: saves/ministral/full/sft output_dir: saves/ministral-8B/full/sft
logging_steps: 10 logging_steps: 10
save_steps: 500 save_steps: 500
plot_loss: true plot_loss: true
overwrite_output_dir: true overwrite_output_dir: true
save_only_model: false save_only_model: false
report_to: none # choices: [none, wandb, tensorboard, swanlab, mlflow]
### train ### train
per_device_train_batch_size: 1 per_device_train_batch_size: 1
......
...@@ -20,12 +20,13 @@ preprocessing_num_workers: 16 ...@@ -20,12 +20,13 @@ preprocessing_num_workers: 16
dataloader_num_workers: 4 dataloader_num_workers: 4
### output ### output
output_dir: saves/ministral/lora/sft output_dir: saves/ministral-8B/lora/sft
logging_steps: 10 logging_steps: 10
save_steps: 500 save_steps: 500
plot_loss: true plot_loss: true
overwrite_output_dir: true overwrite_output_dir: true
save_only_model: false save_only_model: false
report_to: none # choices: [none, wandb, tensorboard, swanlab, mlflow]
### train ### train
per_device_train_batch_size: 1 per_device_train_batch_size: 1
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment