- 17 Nov, 2025 1 commit
-
-
Jeffrey Morgan authored
-
- 14 Nov, 2025 1 commit
-
-
Parth Sareen authored
-
- 13 Nov, 2025 2 commits
-
-
nicole pardal authored
-
Kowyo authored
-
- 12 Nov, 2025 1 commit
-
-
Daniel Hiltgen authored
* docs: vulkan information * Revert "CI: Set up temporary opt-out Vulkan support (#12614)" This reverts commit 8b6e5bae. * vulkan: temporary opt-in for Vulkan support Revert this once we're ready to enable by default. * win: add vulkan CI build
-
- 11 Nov, 2025 4 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
Some route endpoints return an empty response with a 200 OK. These should be documented in the OpenAPI doc. Note that the previous deletion response was not correct.
-
Sheikh authored
-
- 08 Nov, 2025 1 commit
-
-
Parth Sareen authored
-
- 07 Nov, 2025 2 commits
-
-
Daniel Hiltgen authored
* doc: re-add login autostart faq This appears to have been accidentally dropped during the doc migration. * docs: GPU updates lost on the doc update * review comments: improve windows login disable instructions
-
Tomoya Fujita authored
-
- 05 Nov, 2025 1 commit
-
-
nicole pardal authored
Co-authored-by:A-Akhil <akhilrahul70@gmail.com> This PR introduces a new ollama embed command that allows users to generate embeddings directly from the command line. Added ollama embed MODEL [TEXT...] command for generating text embeddings Supports both direct text arguments and stdin piping for scripted workflows Outputs embeddings as JSON arrays (one per line)
-
- 29 Oct, 2025 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 28 Oct, 2025 5 commits
-
-
Parth Sareen authored
-
Parth Sareen authored
-
Parth Sareen authored
-
Parth Sareen authored
This reverts commit 934dd9e1.
-
Parth Sareen authored
-
- 16 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
8.7 is Jetpack only, so no need on x86 builds 10.3 covers [G]B300
-
- 11 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
-
- 07 Oct, 2025 2 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
* Bring back escape valve for llm libraries If the new discovery logic picks the wrong library, this gives users the ability to force a specific one using the same pattern as before. This can also potentially speed up bootstrap discovery if one of the libraries takes a long time to load and ultimately bind to no devices. For example unsupported AMD iGPUS can sometimes take a while to discover and rule out. * Bypass extra discovery on jetpack systems On at least Jetpack6, cuda_v12 appears to expose the iGPU, but crashes later on in cublasInit so if we detect a Jetpack, short-circuit and use that variant.
-
- 02 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
Notable EOLs with this change: - MacOS v12 and v13 are no longer supported (v14+ required) - AMD gfx900 and gfx906 are no longer supported
-
- 01 Oct, 2025 1 commit
-
-
Daniel Hiltgen authored
This revamps how we discover GPUs in the system by leveraging the Ollama runner. This should eliminate inconsistency between our GPU discovery and the runners capabilities at runtime, particularly for cases where we try to filter out unsupported GPUs. Now the runner does that implicitly based on the actual device list. In some cases free VRAM reporting can be unreliable which can leaad to scheduling mistakes, so this also includes a patch to leverage more reliable VRAM reporting libraries if available. Automatic workarounds have been removed as only one GPU leveraged this, which is now documented. This GPU will soon fall off the support matrix with the next ROCm bump. Additional cleanup of the scheduler and discovery packages can be done in the future once we have switched on the new memory management code, and removed support for the llama runner.
-
- 22 Sep, 2025 2 commits
- 15 Sep, 2025 1 commit
-
-
Daniel Hiltgen authored
-
- 11 Sep, 2025 1 commit
-
-
Michael Yang authored
* feat: add field to truncate embeddings * add openai embeddings for dimensions
-
- 10 Sep, 2025 1 commit
-
-
Daniel Hiltgen authored
* Add support for upcoming NVIDIA Jetsons The latest Jetsons with JetPack 7 are moving to an SBSA compatible model and will not require building a JetPack specific variant. * cuda: bring back dual versions This adds back dual CUDA versions for our releases, with v11 and v13 to cover a broad set of GPUs and driver versions. * win: break up native builds in build_windows.ps1 * v11 build working on windows and linux * switch to cuda v12.8 not JIT * Set CUDA compression to size * enhance manual install linux docs
-
- 08 Sep, 2025 1 commit
-
-
Daniel Hiltgen authored
This debug setting can help troubleshoot obscure initialization failures.
-
- 15 Aug, 2025 1 commit
-
-
Thomas Pelster authored
-
- 14 Aug, 2025 1 commit
-
-
Daniel Hiltgen authored
Some users expect the rocm bundles to be self-sufficient, but are designed to be additive.
-
- 06 Aug, 2025 3 commits
-
-
Patrick Devine authored
-
Gao feng authored
update api.md to make it consist with code. https://github.com/ollama/ollama/blob/main/server/download.go#L447
-
Parth Sareen authored
-
- 05 Aug, 2025 1 commit
-
-
Jeffrey Morgan authored
-
- 28 Jul, 2025 1 commit
-
-
Yoshi authored
-