Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in / Register
  • L Llama1and2_lmdeploy
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Repository
    • Repository
    • Files
    • Commits
    • Branches
    • Tags
    • Contributors
    • Graph
    • Compare
  • Issues 2
    • Issues 2
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 0
    • Merge requests 0
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Deployments
    • Deployments
    • Environments
    • Releases
  • Monitor
    • Monitor
    • Incidents
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Infrastructure Registry
  • Analytics
    • Analytics
    • CI/CD
    • Repository
    • Value stream
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Graph
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
Collapse sidebar
  • ModelZoo
  • Llama1and2_lmdeploy
  • Issues
  • #1

Closed
Open
Created Jun 17, 2024 by ncic_liuyao@ncic_liuyao

lmdeploy0.0.13 error RuntimeError: [TM][ERROR] Assertion fail

docker pull image.sourcefind.cn:5000/dcu/admin/base/custom:lmdeploy0.0.13_dtk23.04_torch1.13_py38

用上面拉取docker镜像的ID替换

主机端路径

容器映射路径

docker run -it --name llama --shm-size=1024G --device=/dev/kfd --device=/dev/dri/ --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --ulimit memlock=-1:-1 --ipc=host --network host --group-add video -v : /bin/bash

running: 运行 LLama2-13b

模型转换

lmdeploy convert --model_name llama2 --model_path /path/to/model --model_format hf --tokenizer_path None --dst_path ./workspace_llama2-13b --tp 1

bash界面运行

lmdeploy chat turbomind --model_path ./workspace_llama2-13b --tp 1

error: [root@master141-k100 llama_lmdeploy]# lmdeploy chat turbomind --model_path ./workspace --tp 8 Traceback (most recent call last): File "/usr/local/bin/lmdeploy", line 33, in sys.exit(load_entry_point('lmdeploy==0.0.13', 'console_scripts', 'lmdeploy')()) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/cli/cli.py", line 61, in run fire.Fire(cli, name='lmdeploy') File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 141, in Fire component_trace = _Fire(component, args, parsed_flag_args, context, name) File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 475, in _Fire component, remaining_args = _CallAndUpdateTrace( File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace component = fn(*varargs, **kwargs) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/cli/chat.py", line 85, in turbomind run_turbomind_model(model_path, File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/turbomind/chat.py", line 92, in main tm_model = tm.TurboMind(model_path, eos_id=tokenizer.eos_token_id, tp=tp) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/turbomind/turbomind.py", line 128, in init self.nccl_params = model.create_nccl_params(self.node_id) RuntimeError: [TM][ERROR] Assertion fail: /data/Imdeploy/data_2304/0.0.13/src/lmdeploy/src/turbomind/triton_backend/transformer_triton_backend.cpp:32

[root@master141-k100 llama_lmdeploy]# lmdeploy chat turbomind --model_path ./workspace --tp 8 Traceback (most recent call last): File "/usr/local/bin/lmdeploy", line 33, in sys.exit(load_entry_point('lmdeploy==0.0.13', 'console_scripts', 'lmdeploy')()) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/cli/cli.py", line 61, in run fire.Fire(cli, name='lmdeploy') File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 141, in Fire component_trace = _Fire(component, args, parsed_flag_args, context, name) File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 475, in _Fire component, remaining_args = _CallAndUpdateTrace( File "/usr/local/lib/python3.8/site-packages/fire/core.py", line 691, in _CallAndUpdateTrace component = fn(*varargs, **kwargs) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/cli/chat.py", line 85, in turbomind run_turbomind_model(model_path, File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/turbomind/chat.py", line 92, in main tm_model = tm.TurboMind(model_path, eos_id=tokenizer.eos_token_id, tp=tp) File "/usr/local/lib/python3.8/site-packages/lmdeploy-0.0.13-py3.8-linux-x86_64.egg/lmdeploy/turbomind/turbomind.py", line 128, in init self.nccl_params = model.create_nccl_params(self.node_id) RuntimeError: [TM][ERROR] Assertion fail: /data/Imdeploy/data_2304/0.0.13/src/lmdeploy/src/turbomind/triton_backend/transformer_triton_backend.cpp:32

Assignee
Assign to
Time tracking