- 07 Mar, 2024 3 commits
-
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
Jeffrey Morgan authored
-
John authored
Signed-off-by:hishope <csqiye@126.com>
-
- 05 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 01 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 25 Feb, 2024 1 commit
-
-
elthommy authored
Remove unused GPT4all Use nomic-embed-text as embedded model Fix a deprecation warning (__call__)
-
- 22 Feb, 2024 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 21 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 20 Feb, 2024 4 commits
-
-
Jeffrey Morgan authored
Add instructions to get public key on windows
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 19 Feb, 2024 2 commits
-
-
Patrick Devine authored
-
Daniel Hiltgen authored
-
- 16 Feb, 2024 1 commit
-
-
Tristan Rhodes authored
-
- 15 Feb, 2024 2 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This focuses on Windows first, but coudl be used for Mac and possibly linux in the future.
-
- 12 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 09 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 08 Feb, 2024 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 07 Feb, 2024 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 06 Feb, 2024 1 commit
-
-
Bruce MacDonald authored
-
- 05 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 02 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 29 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
Some users are new to containers and unsure where the server logs go
-
- 26 Jan, 2024 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 22 Jan, 2024 1 commit
-
-
Michael Yang authored
-
- 21 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
The linux build now support parallel CPU builds to speed things up. This also exposes AMD GPU targets as an optional setting for advaced users who want to alter our default set.
-
- 20 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 18 Jan, 2024 3 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
A few obvious levels were adjusted, but generally everything mapped to "info" level.
-
Daniel Hiltgen authored
-
- 12 Jan, 2024 1 commit
-
-
Tristram Oaten authored
After executing the `userdel ollama` command, I saw this message: ```sh $ sudo userdel ollama userdel: group ollama not removed because it has other members. ``` Which reminded me that I had to remove the dangling group too. For completeness, the uninstall instructions should do this too. Thanks!
-
- 11 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
This reduces the built-in linux version to not use any vector extensions which enables the resulting builds to run under Rosetta on MacOS in Docker. Then at runtime it checks for the actual CPU vector extensions and loads the best CPU library available
-
- 09 Jan, 2024 1 commit
-
-
Robin Glauser authored
Fixed assistant in the example response.
-