- 17 Nov, 2025 2 commits
-
-
Eva H authored
-
Jeffrey Morgan authored
-
- 16 Nov, 2025 6 commits
-
-
omahs authored
-
Joel Bryan Juliano authored
Kdeps is an AI framework for building Dockerized full-stack AI applications declaratively and uses Ollama LLM models on the backend
-
pierwill authored
Co-authored-by:pierwill <pierwill@users.noreply.github.com>
-
Vignesh Skanda authored
-
Laurențiu Nicola authored
-
Patrick Devine authored
This change adds a basic benchmarking test framework for Ollama which can be used to determine the prefill, eval, load duration, and total duration for running a given model or models.
-
- 14 Nov, 2025 2 commits
-
-
Daniel Hiltgen authored
Many failed GPU discovery issues recently can be traced to incorrect override settings. This extra logging should help quickly spot these and guide users to try unsetting them first.
-
Parth Sareen authored
-
- 13 Nov, 2025 9 commits
-
-
Michael Yang authored
-
Michael Yang authored
* use slice/chunks * bert * llama4 * gemma3n * gptoss * mistral3 * qwen3vl * qwen25vl * deepseek2 * remove unused ops
-
Parth Sareen authored
-
Michael Yang authored
* slice * chunk, chunksections
-
nicole pardal authored
-
Kowyo authored
-
Jeffrey Morgan authored
The code in this directory has been replaced with the new Go version in the 'app' directory.
-
Radhi authored
-
Jeffrey Morgan authored
-
- 12 Nov, 2025 3 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
* docs: vulkan information * Revert "CI: Set up temporary opt-out Vulkan support (#12614)" This reverts commit 8b6e5bae. * vulkan: temporary opt-in for Vulkan support Revert this once we're ready to enable by default. * win: add vulkan CI build
-
Daniel Hiltgen authored
This should be reverted once we update ggml past b6897
-
- 11 Nov, 2025 14 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
Some route endpoints return an empty response with a 200 OK. These should be documented in the OpenAPI doc. Note that the previous deletion response was not correct.
-
Jesse Gross authored
We currently assign model layers to GPUs according to free VRAM, which assumes that GPU performance is roughly equal. This does not work well for mixed dGPU and iGPU systems because iGPUs typically use system memory which is large but their performance is slow. This instead assigns layers to dGPUs first and then iGPUs. In the future, this could be generalized to have a more fine grained notion of GPU performance but dGPU vs. iGPU performance is the most extreme.
-
Jesse Gross authored
Originally, llamaServer represented old memory estimates, which could be used with either the old or new engine. ollamaServer was used only for the new estimates and new engine. Since these implementations did not map directly to engine, there was engine- specific code in common code paths. Now that new estimates are always used for the new engine, there is a direct mapping between server type and engine. This separates out most of the engine-specific code into the correct implementation to make things easier to understand.
-
Jesse Gross authored
Currently for both the old and new engines, there is code to calculate how much memory is required for a model and lay out the layers onto GPUs. This reuses the new engine's lay out code for the old engine as well, bringing them closer together. The old engine continues to use its current method of estimating required memory. This reduces maintainence effort and improves consistency, as new features only need to be implemented in one place. The newer code is also more accurate, especially with multiple GPUs.
-
Jesse Gross authored
We used to control the way that llama.cpp saw devices using CUDA_VISIBLE_DEVICES or similar. This would ensure that the layers offloaded to a device were actually the ones intended. This is particularly important because we might reorder devices based on free memory or performance. When we started explicitly scheduling layers, this logic went away but the llamarunner didn't have any way to set the correct order of devices. This meant that the correct number of layers would be assigned to a device but not necessarily the layers that were expected. This change sets up the devices correctly based on the offload information.
-
Eva H authored
-
Baptiste Jamin authored
Adds logprobs support to Ollama's API including support for Ollama's OpenAI-compatible API. By specifying the new 'logprobs' boolean parameter in the API, Ollama will return the log probabilities for each token generated. 'top_logprobs', an integer value can also be specified up to the value 20. When specified, the API will also provide the number of most likely tokens to return at each token position Co-authored-by:Baptiste Jamin <baptiste@crisp.chat>
-
Eva Ho authored
-
Sheikh authored
-
Eva Ho authored
-
Eva Ho authored
-
Eva Ho authored
-
- 10 Nov, 2025 1 commit
-
-
Eva H authored
-
- 08 Nov, 2025 3 commits
-
-
Bruce MacDonald authored
-
Patrick Devine authored
-
Parth Sareen authored
-