Commit eedfcef8 authored by mashun1's avatar mashun1
Browse files

Update README.md

parent dff8d698
......@@ -27,16 +27,8 @@
pip install -r requirements.txt
cd xformers && pip install xformers==0.0.23 --no-deps && bash patch_xformers.rocm.sh (whl.zip文件中)
# 以下按需安装
yum install epel-release -y
yum localinstall --nogpgcheck https://download1.rpmfusion.org/free/el/rpmfusion-free-release-7.noarch.rpm -y
yum install ffmpeg ffmpeg-devel libsm6 libxext6 -y
sudo apt install ffmpeg
### Dockerfile(方法二)
......@@ -47,30 +39,20 @@
pip install -r requirements.txt
pip install flash_attn-2.0.4_torch2.1_dtk2310-cp38-cp38-linux_x86_64.whl (whl.zip文件中)
pip install triton-2.1.0%2Bgit34f8189.abi0.dtk2310-cp38-cp38-manylinux2014_x86_64.whl (开发者社区下载)
cd xformers && pip install xformers==0.0.23 --no-deps && bash patch_xformers.rocm.sh (whl.zip文件中)
# 以下按需安装
yum install epel-release -y
yum localinstall --nogpgcheck https://download1.rpmfusion.org/free/el/rpmfusion-free-release-7.noarch.rpm -y
yum install ffmpeg ffmpeg-devel libsm6 libxext6 -y
sudo apt install ffmpeg
### Anaconda (方法三)
1、关于本项目DCU显卡所需的特殊深度学习库可从光合开发者社区下载安装:
https://developer.hpccube.com/tool/
DTK驱动:dtk23.10.1
python:python3.8
DTK驱动:dtk24.04.1
python:python3.10
torch:2.1.0
torchvision:0.16.0
triton:2.1.0
xformers:0.0.25
flash-attn:2.0.4
Tips:以上dtk驱动、python、torch等DCU相关工具版本需要严格一一对应
......@@ -78,8 +60,6 @@ Tips:以上dtk驱动、python、torch等DCU相关工具版本需要严格一
pip install -r requirements.txt
cd xformers && pip install xformers==0.0.23 --no-deps && bash patch_xformers.rocm.sh (whl.zip文件中)
# 按需安装
conda install -c conda-forge ffmpeg
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment