Commit 86913514 authored by Rayyyyy's avatar Rayyyyy
Browse files

Modify chat.py in 70B

parent 0aea7dd1
...@@ -5,9 +5,6 @@ export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 # 可自行修改为指定显卡号 ...@@ -5,9 +5,6 @@ export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 # 可自行修改为指定显卡号
export HSA_FORCE_FINE_GRAIN_PCIE=1 export HSA_FORCE_FINE_GRAIN_PCIE=1
export USE_MIOPEN_BATCHNORM=1 export USE_MIOPEN_BATCHNORM=1
export MASTER_ADDR=localhost
export MASTER_PORT=12355
export RANK=0
echo "Starting ..." echo "Starting ..."
# 8B # 8B
......
#!/bin/bash #!/bin/bash
echo "Export params ..." echo "Export params ..."
export HIP_VISIBLE_DEVICES=1,2 export HIP_VISIBLE_DEVICES=1,2 # 可自行修改为指定显卡号
export HSA_FORCE_FINE_GRAIN_PCIE=1 export HSA_FORCE_FINE_GRAIN_PCIE=1
export USE_MIOPEN_BATCHNORM=1 export USE_MIOPEN_BATCHNORM=1
......
import os
import sys import sys
import fire import fire
import warnings
from typing import List, Optional from typing import List, Optional
from llama import Dialog, Llama from llama import Dialog, Llama
warnings.filterwarnings('ignore', category=UserWarning)
def main( def main(
ckpt_dir: str, ckpt_dir: str,
...@@ -24,11 +27,15 @@ def main( ...@@ -24,11 +27,15 @@ def main(
try: try:
# Continue util the user decides to stop # Continue util the user decides to stop
while True: while True:
user_input = input("You: ") local_rank = int(os.environ.get("LOCAL_RANK", 0))
# Allow the user to quit the dialogue if local_rank > 0:
if user_input.lower() in ['stop', 'exit']: dialogs.append({"role": "user", "content": "None"})
break else:
dialogs.append({"role": "user", "content": user_input}) user_input = input("You: ")
# Allow the user to quit the dialogue
if user_input.lower() in ['stop', 'exit']:
break
dialogs.append({"role": "user", "content": user_input})
# Generate response based on the current dialog context # Generate response based on the current dialog context
results = generator.chat_completion( results = generator.chat_completion(
[dialogs], [dialogs],
......
ulimit -u 200000
export OMP_NUM_THREADS=1
export NCCL_DEBUG=INFO
export MIOPEN_FIND_MODE=3
export HSA_FORCE_FINE_GRAIN_PCIE=1
export MIOPEN_COMPILE_PARALLEL_LEVEL=1
export NCCL_PLUGIN_P2P=ucx
export NCCL_SOCKET_IFNAME=ib0
export NCCL_P2P_LEVEL=5
export NCCL_NET_PLUGIN=none
echo "START TIME: $(date)"
hostfile=./hostfile
np=$(cat $hostfile|sort|uniq |wc -l)
np=$(($np*8))
nodename=$(cat $hostfile |sed -n "1p")
dist_url=`echo $nodename | awk '{print $1}'`
which mpirun
mpirun -np $np --allow-run-as-root --hostfile hostfile --bind-to none --mca btl_tcp_if_include $dist_url run_train_single.sh
echo "END TIME: $(date)"
#!/bin/bash
export HSA_FORCE_FINE_GRAIN_PCIE=1
export MIOPEN_FIND_MODE=3
export MIOPEN_COMPILE_PARALLEL_LEVEL=1
export NCCL_PLUGIN_P2P=ucx
export NCCL_SOCKET_IFNAME=ib0
export NCCL_P2P_LEVEL=5
export NCCL_IB_HCA=mlx5_0
export NCCL_DEBUG=INFO
export NCCL_NET_PLUGIN=none
lrank=$OMPI_COMM_WORLD_LOCAL_RANK
echo "LRANK===============================$lrank"
RANK=$OMPI_COMM_WORLD_RANK
WORLD_SIZE=$OMPI_COMM_WORLD_SIZE
export HIP_VISIBLE_DEVICES=0,1,2,3
LR=1e-5
APP="python3 ../main.py \
--deepspeed ../deepspeed.json \
--do_train \
--train_file AdvertiseGen/train.json \
--prompt_column content \
--response_column summary \
--model_name_or_path THUDM/chatglm-6b \
--output_dir ./output_ft/pretrain \
--overwrite_output_dir \
--max_source_length 64 \
--max_target_length 64 \
--per_device_train_batch_size 1 \
--per_device_eval_batch_size 1 \
--gradient_accumulation_steps 1 \
--predict_with_generate \
--max_steps 2000 \
--logging_steps 5 \
--save_steps 1000 \
--learning_rate $LR \
--fp16 \
--local_rank $lrank "
case ${lrank} in
[0])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_0:1
export UCX_IB_PCI_BW=mlx5_0:50Gbs
numactl --cpunodebind=0 --membind=0 ${APP}
;;
[1])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_1:1
export UCX_IB_PCI_BW=mlx5_1:50Gbs
numactl --cpunodebind=0 --membind=0 ${APP}
;;
[2])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_2:1
export UCX_IB_PCI_BW=mlx5_2:50Gbs
numactl --cpunodebind=0 --membind=0 ${APP}
;;
[3])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_3:1
export UCX_IB_PCI_BW=mlx5_3:50Gbs
numactl --cpunodebind=0 --membind=0 ${APP}
;;
[4])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_4:1
export UCX_IB_PCI_BW=mlx5_4:50Gbs
numactl --cpunodebind=3 --membind=3 ${APP}
;;
[5])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_5:1
export UCX_IB_PCI_BW=mlx5_5:50Gbs
numactl --cpunodebind=3 --membind=3 ${APP}
;;
[6])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_6:1
export UCX_IB_PCI_BW=mlx5_6:50Gbs
numactl --cpunodebind=3 --membind=3 ${APP}
;;
[7])
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
export UCX_NET_DEVICES=mlx5_7:1
export UCX_IB_PCI_BW=mlx5_7:50Gbs
numactl --cpunodebind=3 --membind=3 ${APP}
;;
esac
#!/bin/bash #!/bin/bash
echo "Export params ..." echo "Export params ..."
export HIP_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 # 可自行修改为指定显卡号
export HIP_VISIBLE_DEVICES=0 # 自行修改为训练的卡号和数量
export HSA_FORCE_FINE_GRAIN_PCIE=1 export HSA_FORCE_FINE_GRAIN_PCIE=1
export USE_MIOPEN_BATCHNORM=1 export USE_MIOPEN_BATCHNORM=1
...@@ -13,8 +12,8 @@ torchrun --nproc_per_node 1 example_text_completion.py \ ...@@ -13,8 +12,8 @@ torchrun --nproc_per_node 1 example_text_completion.py \
--tokenizer_path Meta-Llama-3-8B/original/tokenizer.model \ --tokenizer_path Meta-Llama-3-8B/original/tokenizer.model \
--max_seq_len 128 --max_batch_size 4 --max_seq_len 128 --max_batch_size 4
# Meta-Llama-3-8B-Instruct 模型 # Meta-Llama-3-70B-Instruct 模型
# torchrun --nproc_per_node 1 example_chat_completion.py \ # torchrun --nproc_per_node 8 example_chat_completion.py \
# --ckpt_dir ./Meta-Llama-3-8B-Instruct/original/ \ # --ckpt_dir /data/Meta-llama3-models/Meta-Llama-3-70B-Instruct/original/ \
# --tokenizer_path ./Meta-Llama-3-8B-Instruct/original/tokenizer.model \ # --tokenizer_path /data/Meta-llama3-models/Meta-Llama-3-70B-Instruct/original/tokenizer.model \
# --max_seq_len 512 --max_batch_size 6 # --max_seq_len 512 --max_batch_size 6
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment