llm: Remove GGML_CUDA_NO_PEER_COPY for ROCm (#7174)
This workaround logic in llama.cpp is causing crashes for users with less system memory than VRAM.
Showing
Please register or sign in to comment
This workaround logic in llama.cpp is causing crashes for users with less system memory than VRAM.