Unverified Commit 57a3c4db authored by kingkingofall's avatar kingkingofall Committed by GitHub
Browse files

[chat]fix readme (#3429)

* fix stage 2

fix stage 2

* add torch
parent 7d8d8256
......@@ -57,7 +57,7 @@ You can run the `examples/train_rm.sh` to start a reward model training.
You can also use the following cmd to start training a reward model.
```
torchrun --standalone --nproc_per_node=4 train_reward_model.py
torchrun --standalone --nproc_per_node=4 train_reward_model.py \
--pretrain "/path/to/LLaMa-7B/" \
--model 'llama' \
--strategy colossalai_zero2 \
......
......@@ -51,6 +51,7 @@ Please ensure you have downloaded HF-format model weights of LLaMA models.
Usage:
```python
import torch
from transformers import LlamaForCausalLM
USE_8BIT = True # use 8-bit quantization; otherwise, use fp16
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment