- 23 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 01 May, 2024 1 commit
-
-
alwqx authored
-
- 09 Apr, 2024 2 commits
-
-
Blake Mizerany authored
-
Blake Mizerany authored
This commit introduces a more friendly way to build Ollama dependencies and the binary without abusing `go generate` and removing the unnecessary extra steps it brings with it. This script also provides nicer feedback to the user about what is happening during the build process. At the end, it prints a helpful message to the user about what to do next (e.g. run the new local Ollama).
-
- 26 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 25 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 09 Mar, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 07 Mar, 2024 3 commits
-
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
Jeffrey Morgan authored
-
John authored
Signed-off-by:hishope <csqiye@126.com>
-
- 21 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
The linux build now support parallel CPU builds to speed things up. This also exposes AMD GPU targets as an optional setting for advaced users who want to alter our default set.
-
- 20 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 18 Jan, 2024 3 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
A few obvious levels were adjusted, but generally everything mapped to "info" level.
-
Daniel Hiltgen authored
-
- 11 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
This reduces the built-in linux version to not use any vector extensions which enables the resulting builds to run under Rosetta on MacOS in Docker. Then at runtime it checks for the actual CPU vector extensions and loads the best CPU library available
-
- 25 Dec, 2023 1 commit
-
-
Daniel Hiltgen authored
-
- 22 Dec, 2023 1 commit
-
-
Daniel Hiltgen authored
By default builds will now produce non-debug and non-verbose binaries. To enable verbose logs in llama.cpp and debug symbols in the native code, set `CGO_CFLAGS=-g`
-
- 19 Dec, 2023 1 commit
-
-
Daniel Hiltgen authored
If someone checks out the ollama repo and doesn't install the CUDA library, this will ensure they can build a CPU only version
-
- 01 Oct, 2023 1 commit
-
-
Jiayu Liu authored
-
- 20 Sep, 2023 4 commits
-
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 14 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* enable packaging multiple cuda versions * use nvcc cuda version if available --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 12 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* linux gpu support * handle multiple gpus * add cuda docker image (#488) --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 30 Aug, 2023 1 commit
-
-
Bruce MacDonald authored
* remove c code * pack llama.cpp * use request context for llama_cpp * let llama_cpp decide the number of threads to use * stop llama runner when app stops * remove sample count and duration metrics * use go generate to get libraries * tmp dir for running llm
-
- 25 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 08 Aug, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 24 Jul, 2023 1 commit
-
-
Michael Yang authored
-
- 21 Jul, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 18 Jul, 2023 1 commit
-
-
Matt Williams authored
Signed-off-by:Matt Williams <m@technovangelist.com>
-
- 07 Jul, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 28 Jun, 2023 4 commits
-
-
Michael Yang authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
-
- 27 Jun, 2023 3 commits
-
-
Bruce MacDonald authored
-
Michael Chiang authored
-
Michael Chiang authored
-