- 11 Nov, 2025 1 commit
-
-
Eva Ho authored
-
- 10 Nov, 2025 1 commit
-
-
Eva H authored
-
- 08 Nov, 2025 3 commits
-
-
Bruce MacDonald authored
-
Patrick Devine authored
-
Parth Sareen authored
-
- 07 Nov, 2025 2 commits
-
-
Daniel Hiltgen authored
* doc: re-add login autostart faq This appears to have been accidentally dropped during the doc migration. * docs: GPU updates lost on the doc update * review comments: improve windows login disable instructions
-
Tomoya Fujita authored
-
- 06 Nov, 2025 15 commits
-
-
Thomas Stocker authored
* Remove unnecessary macos 13 Patch * Remove unnecessary MacOs Version Guard patch * rename patchesw * remove again macos13 patch * rename files
-
Jeffrey Morgan authored
-
Saifeddine ALOUI authored
-
breatn authored
-
Jeffrey Morgan authored
-
Eva Ho authored
-
Eva H authored
feat: add support for WebP images in Ollama's app
-
Eva Ho authored
-
Daniel Hiltgen authored
-
Daniel Alejandro Coll Tejeda authored
-
7394112478 authored
-
mags0ft authored
-
Saifeddine ALOUI authored
-
Vincent Koc authored
-
Eva Ho authored
-
- 05 Nov, 2025 13 commits
-
-
Eva Ho authored
-
Eva Ho authored
-
Eva Ho authored
-
Daniel Hiltgen authored
-
Daniel Alejandro Coll Tejeda authored
-
nicole pardal authored
Co-authored-by:A-Akhil <akhilrahul70@gmail.com> This PR introduces a new ollama embed command that allows users to generate embeddings directly from the command line. Added ollama embed MODEL [TEXT...] command for generating text embeddings Supports both direct text arguments and stdin piping for scripted workflows Outputs embeddings as JSON arrays (one per line)
-
Daniel Hiltgen authored
The scheduler updates free VRAM based on current loaded models. This was mutating the persisted list of GPUs, and when coupled with the non-refreshing logic for Metal that lead to stale low VRAM reporting after unload. The fix is to make sure the GPU discovery always returns a copy so the schedulers GPU list is in fact ephemeral and doesn't leak any temporary adjustments back into the persistent list.
-
Patrick Devine authored
-
Daniel Hiltgen authored
The behavior change in 0.12.4 is the most likely the root cause of hangs some users are seeing. This reverts to the 0.12.3 code, with some added trace logging.
-
Youdon authored
-
Daniel Hiltgen authored
-
Grace authored
* routes/types: add tool call id --------- Co-authored-by:ParthSareen <parth.sareen@ollama.com>
-
Daniel Hiltgen authored
-
- 04 Nov, 2025 5 commits
-
-
Daniel Hiltgen authored
* discovery: only retry AMD GPUs CUDA and Vulkan don't crash on unsupported devices, so retry isn't necessary. This also refactors the code to shift the Library specific logic into the ml package. * review comments
-
virajwad authored
* PDH free memory skeleton * Add PDH printing * Add LUID support for Vulkan * wire luid from ggml-vulkan to mem-dxgi-pdh file * Fix to ggml-impl * Continue skeleton * Implemented ggml_dxgi_pdh_get_device_memory * fix comments * Fix - change value GB to bytes * add ifdefs to only support windows and not linux * modify error codes * Finished ggml_dxgi_pdh_init() function * completed ggml_dxgi_pdh_release() * Formatting changes, add static to functions * fix build errors * fix go build error * fix luid - now should match between dxgi and vulkan * Fix the free memory reporting (was using copy by value, change to reference) * keep only dxgi1_2.h * Modifications based on PR feedback * fix merge conflicts (2) and fix desc1.description printout * move dxgi + pdh api calls to before the vendor specific library calls * change from 3 samples to 1 sample for PDH * modify when old_mode is set * add fix for building MacOS * fix release and returns for other vendors * add patch file
-
Daniel Hiltgen authored
* app: add code for macOS and Windows apps under 'app' * app: add readme * app: windows and linux only for now * ci: fix ui CI validation --------- Co-authored-by:jmorganca <jmorganca@gmail.com>
-
Daniel Hiltgen authored
Also adjusts the vulkan windows build pattern to match recent changes in other backends so incremental builds are faster.
-
Jesse Gross authored
The initial implementation of qwen3-vl:235b exceeded the maximum graph size based on the number of tensors. Although this was later fixed through the use of the mrope operation, we are close to the limit in some cases. This updates to track the current llama.cpp usage of GGML.
-