- 07 May, 2025 1 commit
-
-
Daniel Hiltgen authored
This reduces the size of our Windows installer payloads by ~256M by dropping support for nvidia drivers older than Feb 2023. Hardware support is unchanged. Linux default bundle sizes are reduced by ~600M to 1G.
-
- 08 Apr, 2025 1 commit
-
-
frob authored
* cleanup: remove OLLAMA_TMPDIR * cleanup: ollama doesn't use temporary executables anymore --------- Co-authored-by:Richard Lyons <frob@cloudstaff.com>
-
- 25 Mar, 2025 1 commit
-
-
copeland3300 authored
-
- 25 Feb, 2025 1 commit
-
-
frob authored
Co-authored-by:Richard Lyons <frob@cloudstaff.com>
-
- 07 Feb, 2025 1 commit
-
-
Azis Alvriyanto authored
-
- 10 Dec, 2024 1 commit
-
-
Stefan Weil authored
-
- 21 Nov, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 12 Nov, 2024 2 commits
-
-
Daniel Hiltgen authored
Docker uses the container filesystem for name resolution, so we can't guide users to use the name of the host group. Instead they must specify the numeric ID.
-
Daniel Hiltgen authored
GPU support can break on some systems after a while. This captures a known workaround to solve the problem.
-
- 11 Sep, 2024 1 commit
-
-
Daniel Hiltgen authored
This adds back a check which was lost many releases back to verify /dev/kfd permissions which when lacking, can lead to confusing failure modes of: "rocBLAS error: Could not initialize Tensile host: No devices found" This implementation does not hard fail the serve command but instead will fall back to CPU with an error log. In the future we can include this in the GPU discovery UX to show detected but unsupported devices we discovered.
-
- 05 Aug, 2024 1 commit
-
-
frob authored
Users using `journalctl` to get logs for issue logging sometimes don't realize that paging is causing information to be missed.
-
- 04 Jul, 2024 1 commit
-
-
Daniel Hiltgen authored
We haven't found a workaround, so for now recommend updating.
-
- 03 Jul, 2024 1 commit
-
-
Daniel Hiltgen authored
Refine the way we log GPU discovery to improve the non-debug output, and report more actionable log messages when possible to help users troubleshoot on their own.
-
- 19 Jun, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 23 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 21 May, 2024 1 commit
-
-
Patrick Devine authored
-
- 20 May, 2024 1 commit
-
-
alwqx authored
-
- 09 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 01 Apr, 2024 1 commit
-
-
Daniel Hiltgen authored
We may have users that run into problems with our current payload model, so this gives us an escape valve.
-
- 21 Mar, 2024 1 commit
-
-
Bruce MacDonald authored
-
- 15 Mar, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 11 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 07 Mar, 2024 2 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
- 15 Feb, 2024 1 commit
-
-
Daniel Hiltgen authored
This focuses on Windows first, but coudl be used for Mac and possibly linux in the future.
-
- 29 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
Some users are new to containers and unsure where the server logs go
-
- 11 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
This reduces the built-in linux version to not use any vector extensions which enables the resulting builds to run under Rosetta on MacOS in Docker. Then at runtime it checks for the actual CPU vector extensions and loads the best CPU library available
-
- 22 Dec, 2023 1 commit
-
-
Matt Williams authored
* Clean up documentation Will probably need to update with PRs for new release. Signed-off-by:
Matt Williams <m@technovangelist.com> * Correcting to fit in 0.1.15 changes Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * addressing comments Signed-off-by:
Matt Williams <m@technovangelist.com> * more api cleanup Signed-off-by:
Matt Williams <m@technovangelist.com> * its llava not llama Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Updated hosting to server and documented all env vars Signed-off-by:
Matt Williams <m@technovangelist.com> * remove last of the cli descriptions Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update further per conversation with jeff earlier today Signed-off-by:
Matt Williams <m@technovangelist.com> * cleanup the doc readme Signed-off-by:
Matt Williams <m@technovangelist.com> * move upgrade to faq Signed-off-by:
Matt Williams <m@technovangelist.com> * first change Signed-off-by:
Matt Williams <m@technovangelist.com> * updated Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * examples in parent Signed-off-by:
Matt Williams <m@technovangelist.com> * add exapmle for create model. Signed-off-by:
Matt Williams <m@technovangelist.com> * update faq Signed-off-by:
Matt Williams <m@technovangelist.com> * update create model api Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update the readme in docs Signed-off-by:
Matt Williams <m@technovangelist.com> * update a few more things Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/modelfile.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> --------- Signed-off-by:
Matt Williams <m@technovangelist.com> Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com>
-