- 09 Mar, 2024 5 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Michael Yang authored
-
- 08 Mar, 2024 8 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
fix: default terminal width, height
-
Daniel Hiltgen authored
Refined ROCm troubleshooting docs
-
Bruce MacDonald authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 07 Mar, 2024 17 commits
-
-
Blake Mizerany authored
Also, document OLLAMA_HOST client semantics per command that honors it. This looks nicer than having a general puprose environment variable section in the root usage which was showing up after the "addition help topics" section outputed by Cobra's default template. It was decided this was easier to work with than using a custom template for Cobra right now.
-
Michael Yang authored
-
Michael Yang authored
fix ci
-
Michael Yang authored
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
Revamp ROCm support
-
Michael Yang authored
remove empty examples
-
Michael Yang authored
-
Michael Yang authored
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
Michael Yang authored
adjust download and upload concurrency based on available bandwidth
-
Daniel Hiltgen authored
Allow setting max vram for workarounds
-
Jeffrey Morgan authored
-
DJ Johnson authored
-
John authored
Signed-off-by:hishope <csqiye@126.com>
-
Patrick Devine authored
-
Daniel Hiltgen authored
Until we get all the memory calculations correct, this can provide and escape valve for users to workaround out of memory crashes.
-
- 06 Mar, 2024 2 commits
-
-
Blake Mizerany authored
Updates #2944
-
Leo authored
* Add Odin Runes to README Add Odin Runes to README This commit adds Odin Runes to the "Community Integrations" section of the README. Odin Runes is a Java-based GPT client designed to provide seamless interaction with GPT models, enhancing productivity in prompt engineering and text generation tasks. This addition highlights the integration between Odin Runes and Ollama, offering users the flexibility to leverage large language models locally within their development workflow. * Update README.md this commit applies the comments of the reviewer.
-
- 05 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 04 Mar, 2024 2 commits
-
-
Anders Rex authored
-
Timothy Graupmann authored
* Added community link for Ollama Copilot * Update README.md --------- Co-authored-by:Michael <mchiang0610@users.noreply.github.com>
-
- 01 Mar, 2024 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 29 Feb, 2024 3 commits
-
-
Daniel Hiltgen authored
Add ollama user to video group
-
Daniel Hiltgen authored
Add env var so podman will map cuda GPUs
-
Daniel Hiltgen authored
Omit build date from gzip headers
-