#!/bin/bash export HIP_VISIBLE_DEVICES=6 python src/train.py \ --stage sft --do_train \ --model_name_or_path /home/practice/deepseek-math-7b-base \ --dataset alpaca_en_demo --template deepseek --finetuning_type lora --lora_target q_proj,v_proj \ --output_dir saves/deepseek/lora/sft \ --overwrite_cache --per_device_train_batch_size 2 --gradient_accumulation_steps 32 --lr_scheduler_type cosine \ --logging_steps 10 --save_steps 1000 --learning_rate 1e-4 --num_train_epochs 3.0 --plot_loss --fp16