Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
ModelZoo
Ministral_pytorch
Commits
d765e83e
Commit
d765e83e
authored
Jun 12, 2025
by
chenych
Browse files
Add train loss
parent
0179c9b5
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
14 additions
and
4 deletions
+14
-4
README.md
README.md
+10
-2
llama-factory/train_full/ministral_full_sft.yaml
llama-factory/train_full/ministral_full_sft.yaml
+2
-1
llama-factory/train_lora/ministral_lora_sft.yaml
llama-factory/train_lora/ministral_lora_sft.yaml
+2
-1
No files found.
README.md
View file @
d765e83e
...
@@ -89,7 +89,7 @@ python infer_vllm.py --model_name mistralai/Ministral-8B-Instruct-2410
...
@@ -89,7 +89,7 @@ python infer_vllm.py --model_name mistralai/Ministral-8B-Instruct-2410
1.
启动server
1.
启动server
```
bash
```
bash
vllm serve mistralai/Ministral-8B-Instruct-2410
--tokenizer
_
mode
mistral
--config
_
format
mistral
--load
_
format
mistral
vllm serve mistralai/Ministral-8B-Instruct-2410
--tokenizer
-
mode
mistral
--config
-
format
mistral
--load
-
format
mistral
```
```
2.
client测试:
2.
client测试:
...
@@ -110,12 +110,20 @@ curl --location 'http://<your-node-url>:8000/v1/chat/completions' \
...
@@ -110,12 +110,20 @@ curl --location 'http://<your-node-url>:8000/v1/chat/completions' \
```
```
## result
## result
Prompt: "Do we need to think for 10 seconds to find the answer of 1 + 1?"
<div
align=
center
>
<div
align=
center
>
<img
src=
"./doc/results.png"
/>
<img
src=
"./doc/results.png"
/>
</div>
</div>
### 精度
### 精度
DCU与GPU精度一致,推理框架:pytorch。
训练框架:
[
Llama-Factory
](
https://developer.sourcefind.cn/codes/OpenDAS/llama-factory
)
训练脚本:
[
ministral_lora_sft.yaml
](
llama-factory/train_lora/ministral_lora_sft.yaml
)
| device | train_loss |
|:----------:|:-------:|
| DCU K100AI | 0.9068 |
| GPU A800 | |
## 应用场景
## 应用场景
### 算法类别
### 算法类别
...
...
llama-factory/train_full/ministral_full_sft.yaml
View file @
d765e83e
...
@@ -17,12 +17,13 @@ preprocessing_num_workers: 16
...
@@ -17,12 +17,13 @@ preprocessing_num_workers: 16
dataloader_num_workers
:
4
dataloader_num_workers
:
4
### output
### output
output_dir
:
saves/ministral/full/sft
output_dir
:
saves/ministral
-8B
/full/sft
logging_steps
:
10
logging_steps
:
10
save_steps
:
500
save_steps
:
500
plot_loss
:
true
plot_loss
:
true
overwrite_output_dir
:
true
overwrite_output_dir
:
true
save_only_model
:
false
save_only_model
:
false
report_to
:
none
# choices: [none, wandb, tensorboard, swanlab, mlflow]
### train
### train
per_device_train_batch_size
:
1
per_device_train_batch_size
:
1
...
...
llama-factory/train_lora/ministral_lora_sft.yaml
View file @
d765e83e
...
@@ -20,12 +20,13 @@ preprocessing_num_workers: 16
...
@@ -20,12 +20,13 @@ preprocessing_num_workers: 16
dataloader_num_workers
:
4
dataloader_num_workers
:
4
### output
### output
output_dir
:
saves/ministral/lora/sft
output_dir
:
saves/ministral
-8B
/lora/sft
logging_steps
:
10
logging_steps
:
10
save_steps
:
500
save_steps
:
500
plot_loss
:
true
plot_loss
:
true
overwrite_output_dir
:
true
overwrite_output_dir
:
true
save_only_model
:
false
save_only_model
:
false
report_to
:
none
# choices: [none, wandb, tensorboard, swanlab, mlflow]
### train
### train
per_device_train_batch_size
:
1
per_device_train_batch_size
:
1
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment