"apps/vscode:/vscode.git/clone" did not exist on "1b9bc16b1a57c6e7957c3fd74d89d1b206fc5bb8"
- 07 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
* Bring back escape valve for llm libraries If the new discovery logic picks the wrong library, this gives users the ability to force a specific one using the same pattern as before. This can also potentially speed up bootstrap discovery if one of the libraries takes a long time to load and ultimately bind to no devices. For example unsupported AMD iGPUS can sometimes take a while to discover and rule out. * Bypass extra discovery on jetpack systems On at least Jetpack6, cuda_v12 appears to expose the iGPU, but crashes later on in cublasInit so if we detect a Jetpack, short-circuit and use that variant.
-
- 08 Sep, 2025 1 commit
-
-
Daniel Hiltgen authored
This debug setting can help troubleshoot obscure initialization failures.
-
- 28 Jul, 2025 1 commit
-
-
Yoshi authored
-
- 23 Jun, 2025 1 commit
-
-
Daniel Hiltgen authored
* Re-remove cuda v11 Revert the revert - drop v11 support requiring drivers newer than Feb 23 This reverts commit c6bcdc42. * Simplify layout With only one version of the GPU libraries, we can simplify things down somewhat. (Jetsons still require special handling) * distinct sbsa variant for linux arm64 This avoids accidentally trying to load the sbsa cuda libraries on a jetson system which results in crashes. * temporary prevent rocm+cuda mixed loading
-
- 13 May, 2025 1 commit
-
-
Daniel Hiltgen authored
Bring back v11 until we can better warn users that their driver is too old. This reverts commit fa393554.
-
- 07 May, 2025 1 commit
-
-
Daniel Hiltgen authored
This reduces the size of our Windows installer payloads by ~256M by dropping support for nvidia drivers older than Feb 2023. Hardware support is unchanged. Linux default bundle sizes are reduced by ~600M to 1G.
-
- 08 Apr, 2025 1 commit
-
-
frob authored
* cleanup: remove OLLAMA_TMPDIR * cleanup: ollama doesn't use temporary executables anymore --------- Co-authored-by:Richard Lyons <frob@cloudstaff.com>
-
- 25 Mar, 2025 1 commit
-
-
copeland3300 authored
-
- 25 Feb, 2025 1 commit
-
-
frob authored
Co-authored-by:Richard Lyons <frob@cloudstaff.com>
-
- 07 Feb, 2025 1 commit
-
-
Azis Alvriyanto authored
-
- 10 Dec, 2024 1 commit
-
-
Stefan Weil authored
-
- 21 Nov, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 12 Nov, 2024 2 commits
-
-
Daniel Hiltgen authored
Docker uses the container filesystem for name resolution, so we can't guide users to use the name of the host group. Instead they must specify the numeric ID.
-
Daniel Hiltgen authored
GPU support can break on some systems after a while. This captures a known workaround to solve the problem.
-
- 11 Sep, 2024 1 commit
-
-
Daniel Hiltgen authored
This adds back a check which was lost many releases back to verify /dev/kfd permissions which when lacking, can lead to confusing failure modes of: "rocBLAS error: Could not initialize Tensile host: No devices found" This implementation does not hard fail the serve command but instead will fall back to CPU with an error log. In the future we can include this in the GPU discovery UX to show detected but unsupported devices we discovered.
-
- 05 Aug, 2024 1 commit
-
-
frob authored
Users using `journalctl` to get logs for issue logging sometimes don't realize that paging is causing information to be missed.
-
- 04 Jul, 2024 1 commit
-
-
Daniel Hiltgen authored
We haven't found a workaround, so for now recommend updating.
-
- 03 Jul, 2024 1 commit
-
-
Daniel Hiltgen authored
Refine the way we log GPU discovery to improve the non-debug output, and report more actionable log messages when possible to help users troubleshoot on their own.
-
- 19 Jun, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 23 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 21 May, 2024 1 commit
-
-
Patrick Devine authored
-
- 20 May, 2024 1 commit
-
-
alwqx authored
-
- 09 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 01 Apr, 2024 1 commit
-
-
Daniel Hiltgen authored
We may have users that run into problems with our current payload model, so this gives us an escape valve.
-
- 21 Mar, 2024 1 commit
-
-
Bruce MacDonald authored
-
- 15 Mar, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 11 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 07 Mar, 2024 2 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
- 15 Feb, 2024 1 commit
-
-
Daniel Hiltgen authored
This focuses on Windows first, but coudl be used for Mac and possibly linux in the future.
-
- 29 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
Some users are new to containers and unsure where the server logs go
-
- 11 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
This reduces the built-in linux version to not use any vector extensions which enables the resulting builds to run under Rosetta on MacOS in Docker. Then at runtime it checks for the actual CPU vector extensions and loads the best CPU library available
-
- 22 Dec, 2023 1 commit
-
-
Matt Williams authored
* Clean up documentation Will probably need to update with PRs for new release. Signed-off-by:
Matt Williams <m@technovangelist.com> * Correcting to fit in 0.1.15 changes Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * addressing comments Signed-off-by:
Matt Williams <m@technovangelist.com> * more api cleanup Signed-off-by:
Matt Williams <m@technovangelist.com> * its llava not llama Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Updated hosting to server and documented all env vars Signed-off-by:
Matt Williams <m@technovangelist.com> * remove last of the cli descriptions Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update further per conversation with jeff earlier today Signed-off-by:
Matt Williams <m@technovangelist.com> * cleanup the doc readme Signed-off-by:
Matt Williams <m@technovangelist.com> * move upgrade to faq Signed-off-by:
Matt Williams <m@technovangelist.com> * first change Signed-off-by:
Matt Williams <m@technovangelist.com> * updated Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * examples in parent Signed-off-by:
Matt Williams <m@technovangelist.com> * add exapmle for create model. Signed-off-by:
Matt Williams <m@technovangelist.com> * update faq Signed-off-by:
Matt Williams <m@technovangelist.com> * update create model api Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update the readme in docs Signed-off-by:
Matt Williams <m@technovangelist.com> * update a few more things Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/modelfile.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> --------- Signed-off-by:
Matt Williams <m@technovangelist.com> Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com>
-