Commit e2c0b4f7 authored by wxj's avatar wxj
Browse files

Update README.md

parent edb026d5
Pipeline #2654 passed with stage
...@@ -214,15 +214,26 @@ DATA_PATH="/datasets/oscar-1GB-llama_text_document" ...@@ -214,15 +214,26 @@ DATA_PATH="/datasets/oscar-1GB-llama_text_document"
将hf格式转为pt格式 将hf格式转为pt格式
```shell ```shell
python tools/checkpoint/convert.py \ python tools/checkpoint/convert.py \
--model-type GPT \ --model-type GPT \
--loader llama_mistral \ --loader llama_mistral \
--saver megatron \ --saver megatron \
--target-tensor-parallel-size 1 \ --target-tensor-parallel-size 1 \
--checkpoint-type hf \ --target-pipeline-parallel-size 2 \
--model-size llama2-7Bf \ --checkpoint-type hf \
--load-dir /models/llama2/Llama-2-7b-hf/ \ --model-size llama2-7Bf \
--save-dir ./Llama-2-7b-megatron-lm-0108 \ --load-dir /data/model_weights/Llama-2-7b-hf/ \
--tokenizer-model /models/llama2/Llama-2-7b-hf --save-dir ./tmp_modelconvert \
--tokenizer-model /data/model_weights/Llama-2-7b-hf/
```
然后在训练的脚本上添加微调的参数
```shell
FINETUNE_ARGS=(
# --finetune
# --pretrained-checkpoint $CHECKPOINT_PATH
--load $CHECKPOINT_PATH
--no-load-optim
--no-load-rng
)
``` ```
# 参考 # 参考
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment