[hongzhen@worker-0 finetune]$ sh test.sh
Traceback (most recent call last):
File "src/train_bash.py", line 2, in <module>
from llmtuner import run_exp
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/__init__.py", line 3, in <module>
from .api import create_app
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/api/__init__.py", line 1, in <module>
from .app import create_app
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/api/app.py", line 8, in <module>
from ..chat import ChatModel
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/chat/__init__.py", line 2, in <module>
from .chat_model import ChatModel
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/chat/chat_model.py", line 5, in <module>
from ..hparams import get_infer_args
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/hparams/__init__.py", line 6, in <module>
from .parser import get_eval_args, get_infer_args, get_train_args
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/hparams/parser.py", line 14, in <module>
from ..extras.misc import check_dependencies, get_current_device
File "/public/home/hongzhen/apprepo/qwen1.5-pytorch/finetune/src/llmtuner/extras/misc.py", line 6, in <module>
from peft import PeftModel
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/peft/__init__.py", line 22, in <module>
from .auto import (
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/peft/auto.py", line 31, in <module>
from .config import PeftConfig
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/peft/config.py", line 23, in <module>
from .utils import CONFIG_NAME, PeftType, TaskType
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/peft/utils/__init__.py", line 21, in <module>
from .loftq_utils import replace_lora_weights_loftq
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/peft/utils/loftq_utils.py", line 35, in <module>
import bitsandbytes as bnb
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/__init__.py", line 6, in <module>
from . import cuda_setup, utils, research
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/research/__init__.py", line 1, in <module>
from . import nn
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/research/nn/__init__.py", line 1, in <module>
from .modules import LinearFP8Mixed, LinearFP8Global
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/research/nn/modules.py", line 8, in <module>
from bitsandbytes.optim import GlobalOptimManager
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/optim/__init__.py", line 6, in <module>
from bitsandbytes.cextension import COMPILED_WITH_CUDA
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/cextension.py", line 13, in <module>
setup.run_cuda_setup()
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/cuda_setup/main.py", line 121, in run_cuda_setup
binary_name, cudart_path, cc, cuda_version_string = evaluate_cuda_setup()
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/cuda_setup/main.py", line 347, in evaluate_cuda_setup
cuda_version_string = get_cuda_version()
File "/public/home/hongzhen/.local/lib/python3.8/site-packages/bitsandbytes/cuda_setup/main.py", line 317, in get_cuda_version
major, minor = map(int, torch.version.cuda.split("."))
AttributeError: 'NoneType' object has no attribute 'split'