Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
52f5370c
"torchvision/git@developer.sourcefind.cn:OpenDAS/vision.git" did not exist on "c59f0474a68c53ae567e548d5beeee9820050044"
Commit
52f5370c
authored
Apr 17, 2024
by
Jeremy
Browse files
add support for custom gpu build flags for llama.cpp
parent
7c000ec3
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
15 additions
and
3 deletions
+15
-3
llm/generate/gen_linux.sh
llm/generate/gen_linux.sh
+15
-3
No files found.
llm/generate/gen_linux.sh
View file @
52f5370c
...
@@ -172,8 +172,15 @@ if [ -d "${CUDA_LIB_DIR}" ]; then
...
@@ -172,8 +172,15 @@ if [ -d "${CUDA_LIB_DIR}" ]; then
# Disabling has minimal performance effect while maintaining compatibility.
# Disabling has minimal performance effect while maintaining compatibility.
ARM64_DEFS
=
"-DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_AVX512=off -DLLAMA_CUDA_F16=off"
ARM64_DEFS
=
"-DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_AVX512=off -DLLAMA_CUDA_F16=off"
fi
fi
CMAKE_DEFS
=
"-DLLAMA_CUDA=on -DLLAMA_CUDA_FORCE_MMQ=on -DCMAKE_CUDA_ARCHITECTURES=
${
CMAKE_CUDA_ARCHITECTURES
}
${
COMMON_CMAKE_DEFS
}
${
CMAKE_DEFS
}
${
ARM64_DEFS
}
"
# Users building from source can tune the exact flags we pass to cmake for configuring llama.cpp
CMAKE_DEFS
=
"
${
CMAKE_DEFS
}
${
OLLAMA_CUSTOM_GPU_DEPS
}
"
if
[
-n
"
${
OLLAMA_CUSTOM_GPU_DEFS
}
"
]
;
then
echo
"OLLAMA_CUSTOM_GPU_DEFS=
\"
${
OLLAMA_CUSTOM_GPU_DEFS
}
\"
"
CMAKE_CUDA_DEFS
=
"-DLLAMA_CUDA=on -DCMAKE_CUDA_ARCHITECTURES=
${
CMAKE_CUDA_ARCHITECTURES
}
${
OLLAMA_CUSTOM_GPU_DEFS
}
"
echo
"Building custom GPU"
else
CMAKE_CUDA_DEFS
=
"-DLLAMA_CUDA=on -DLLAMA_CUDA_FORCE_MMQ=on -DCMAKE_CUDA_ARCHITECTURES=
${
CMAKE_CUDA_ARCHITECTURES
}
"
fi
CMAKE_DEFS
=
"
${
COMMON_CMAKE_DEFS
}
${
CMAKE_DEFS
}
${
ARM64_DEFS
}
${
CMAKE_CUDA_DEFS
}
"
BUILD_DIR
=
"../build/linux/
${
ARCH
}
/cuda
${
CUDA_VARIANT
}
"
BUILD_DIR
=
"../build/linux/
${
ARCH
}
/cuda
${
CUDA_VARIANT
}
"
EXTRA_LIBS
=
"-L
${
CUDA_LIB_DIR
}
-lcudart -lcublas -lcublasLt -lcuda"
EXTRA_LIBS
=
"-L
${
CUDA_LIB_DIR
}
-lcudart -lcublas -lcublasLt -lcuda"
build
build
...
@@ -218,7 +225,12 @@ if [ -d "${ROCM_PATH}" ]; then
...
@@ -218,7 +225,12 @@ if [ -d "${ROCM_PATH}" ]; then
fi
fi
init_vars
init_vars
CMAKE_DEFS
=
"
${
COMMON_CMAKE_DEFS
}
${
CMAKE_DEFS
}
-DLLAMA_HIPBLAS=on -DCMAKE_C_COMPILER=
$ROCM_PATH
/llvm/bin/clang -DCMAKE_CXX_COMPILER=
$ROCM_PATH
/llvm/bin/clang++ -DAMDGPU_TARGETS=
$(
amdGPUs
)
-DGPU_TARGETS=
$(
amdGPUs
)
"
CMAKE_DEFS
=
"
${
COMMON_CMAKE_DEFS
}
${
CMAKE_DEFS
}
-DLLAMA_HIPBLAS=on -DCMAKE_C_COMPILER=
$ROCM_PATH
/llvm/bin/clang -DCMAKE_CXX_COMPILER=
$ROCM_PATH
/llvm/bin/clang++ -DAMDGPU_TARGETS=
$(
amdGPUs
)
-DGPU_TARGETS=
$(
amdGPUs
)
"
CMAKE_DEFS
=
"
${
CMAKE_DEFS
}
${
OLLAMA_CUSTOM_GPU_DEPS
}
"
# Users building from source can tune the exact flags we pass to cmake for configuring llama.cpp
if
[
-n
"
${
OLLAMA_CUSTOM_GPU_DEFS
}
"
]
;
then
echo
"OLLAMA_CUSTOM_GPU_DEFS=
\"
${
OLLAMA_CUSTOM_GPU_DEFS
}
\"
"
CMAKE_DEFS
=
"
${
CMAKE_DEFS
}
${
OLLAMA_CUSTOM_GPU_DEFS
}
"
echo
"Building custom GPU"
fi
BUILD_DIR
=
"../build/linux/
${
ARCH
}
/rocm
${
ROCM_VARIANT
}
"
BUILD_DIR
=
"../build/linux/
${
ARCH
}
/rocm
${
ROCM_VARIANT
}
"
EXTRA_LIBS
=
"-L
${
ROCM_PATH
}
/lib -L/opt/amdgpu/lib/x86_64-linux-gnu/ -Wl,-rpath,
\$
ORIGIN/../../rocm/ -lhipblas -lrocblas -lamdhip64 -lrocsolver -lamd_comgr -lhsa-runtime64 -lrocsparse -ldrm -ldrm_amdgpu"
EXTRA_LIBS
=
"-L
${
ROCM_PATH
}
/lib -L/opt/amdgpu/lib/x86_64-linux-gnu/ -Wl,-rpath,
\$
ORIGIN/../../rocm/ -lhipblas -lrocblas -lamdhip64 -lrocsolver -lamd_comgr -lhsa-runtime64 -lrocsparse -ldrm -ldrm_amdgpu"
build
build
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment