- 05 Jan, 2024 6 commits
-
-
Jeffrey Morgan authored
* gpu: read memory info from all cuda devices * add `LOOKUP_SIZE` constant * better constant name * address comments
-
Bruce MacDonald authored
-
Matt Williams authored
-
Michael Yang authored
update Dockerfile.build
-
Matt Williams authored
Signed-off-by:Matt Williams <m@technovangelist.com>
-
Michael Yang authored
-
- 04 Jan, 2024 19 commits
-
-
Daniel Hiltgen authored
Cleaup stale submodule
-
Daniel Hiltgen authored
If the tree has a stale submodule, make sure we clean it up first
-
Daniel Hiltgen authored
Revamp code layout for the llm directory and llama.cpp submodule
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
-
Brian Murray authored
-
Daniel Hiltgen authored
Load dynamic cpu lib on windows
-
Daniel Hiltgen authored
On linux, we link the CPU library in to the Go app and fall back to it when no GPU match is found. On windows we do not link in the CPU library so that we can better control our dependencies for the CLI. This fixes the logic so we correctly fallback to the dynamic CPU library on windows.
-
Bruce MacDonald authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
* update cmake flags for intel macOS * remove `LLAMA_K_QUANTS` * put back `CMAKE_OSX_DEPLOYMENT_TARGET` and disable `LLAMA_F16C`
-
Daniel Hiltgen authored
Improve maintainability of Radeon card list
-
Daniel Hiltgen authored
Fail fast on WSL1 while allowing on WSL2
-
Daniel Hiltgen authored
Fix CPU only builds
-
Daniel Hiltgen authored
Go embed doesn't like when there's no matching files, so put a dummy placeholder in to allow building without any GPU support If no "server" library is found, it's safely ignored at runtime.
-
Daniel Hiltgen authored
This prevents users from accidentally installing on WSL1 with instructions guiding how to upgrade their WSL instance to version 2. Once running WSL2 if you have an NVIDIA card, you can follow their instructions to set up GPU passthrough and run models on the GPU. This is not possible on WSL1.
-
- 03 Jan, 2024 13 commits
-
-
Daniel Hiltgen authored
This moves the list of AMD GPUs to an easier to maintain list which should make it easier to update over time.
-
Daniel Hiltgen authored
Add ollama user to render group for Radeon support
-
Daniel Hiltgen authored
For the ROCm libraries to access the driver, we need to add the ollama user to the render group.
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
-
Daniel Hiltgen authored
Fix windows system memory lookup
-
Daniel Hiltgen authored
This refines the gpu package error handling and fixes a bug with the system memory lookup on windows.
-
Daniel Hiltgen authored
Refactor how we augment llama.cpp and refine windows native build
-
Bruce MacDonald authored
-
Cole Gillespie authored
-
Jeffrey Morgan authored
-
Patrick Devine authored
-
Jeffrey Morgan authored
-
- 02 Jan, 2024 2 commits
-
-
Daniel Hiltgen authored
This one log line was triggering a single line llama.log to be generated in the pwd of the server
-
Daniel Hiltgen authored
-