Commit 23a29a97 authored by mashun1's avatar mashun1
Browse files

Update README.md

parent 53d4b553
......@@ -25,7 +25,7 @@ Multi-Head Attention是一种并行注意力机制,它通过多个子空间中
### Docker(方法一)
docker pull image.sourcefind.cn:5000/dcu/admin/base/pytorch:2.3.0-ubuntu22.04-dtk24.04.3-py3.10
docker pull image.sourcefind.cn:5000/dcu/admin/base/pytorch:2.4.1-ubuntu22.04-dtk25.04-py3.10-fixpy
docker run --shm-size 100g --network=host --name=wan --privileged --device=/dev/kfd --device=/dev/dri --group-add video --cap-add=SYS_PTRACE --security-opt seccomp=unconfined -v 项目地址(绝对路径):/home/ -v /opt/hyhal:/opt/hyhal:ro -it <your IMAGE ID> bash
......@@ -49,16 +49,14 @@ Multi-Head Attention是一种并行注意力机制,它通过多个子空间中
1、关于本项目DCU显卡所需的特殊深度学习库可从光合开发者社区下载安装: https://developer.sourcefind.cn/tool/
```
DTK驱动:dtk24.04.3
DTK驱动:dtk25.04
python:python3.10
torch:2.3.0
torchvision:0.18.1
triton:2.1.0
torch:2.4.1
triton:3.0
flash-attn:2.6.1
deepspeed:0.14.2
apex:1.3.0
xformers:0.0.25
transformers:4.48.0
apex:1.4.0
transformers:4.46.3
```
2、其他非特殊库直接按照requirements.txt安装
......@@ -90,7 +88,7 @@ cd -
2、数据验证(可选)
```bash
cd actionstudio/examples
cd actionstudio/examples/trainings
bash sft_data_verifier.sh
```
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment