1. 23 Dec, 2025 1 commit
  2. 18 Dec, 2025 1 commit
  3. 12 Dec, 2025 2 commits
    • Devon Rifkin's avatar
      docs: add docs for v1/responses and rework openai compat section (#13416) · 9f782285
      Devon Rifkin authored
      
      
      * docs: add docs for v1/responses and rework openai compat section
      
      I reworked the examples to be separated by topic and to be fully
      runnable (i.e., they now log output instead of just suggesting how a
      call might be made).
      
      We now use `<CodeGroup>`s so that each example has a dropdown on the
      docs site for users to choose, which makes the examples a lot more
      digestible (since you only see approx 1/3 of the code you used to).
      
      I also added a new tool to extract code examples into files so that it's
      easier to actually run them and check that they work.
      
      ## Example
      
      ```shell
      go run docs/tools/extract-examples/main.go docs/api/openai-compatibility.mdx
      ```
      
      Output:
      
      ```
      Extracting code examples to: /var/folders/vq/wfm2g6k917d3ldzpjdxc8ph00000gn/T/mdx-examples-3271754368
      
        - 01_basic.py
        - 01_basic.js
        - 01_basic.sh
        - 02_responses.py
        - 02_responses.js
        - 02_responses.sh
        - 03_vision.py
        - 03_vision.js
        - 03_vision.sh
      
      Extracted 9 file(s) to /var/folders/vq/wfm2g6k917d3ldzpjdxc8ph00000gn/T/mdx-examples-3271754368
      
      To run examples:
      
        cd /var/folders/vq/wfm2g6k917d3ldzpjdxc8ph00000gn/T/mdx-examples-3271754368
        npm install   # for JS examples
      
      then run individual files with `node file.js`, `python file.py`, `bash file.sh`
      ```
      
      In the future we should consider actually running the examples in CI and
      having some sort of acceptance test so we can automatically detect when
      our examples break. So this is just a start in that direction.
      
      * Update docs/api/openai-compatibility.mdx
      Co-authored-by: default avatarParth Sareen <parth.sareen@ollama.com>
      
      * Update docs/api/openai-compatibility.mdx
      Co-authored-by: default avatarParth Sareen <parth.sareen@ollama.com>
      
      ---------
      Co-authored-by: default avatarParth Sareen <parth.sareen@ollama.com>
      9f782285
    • Alexander Gusak's avatar
      docs: fix link to modelfile.mdx (#13220) · 93d45d7a
      Alexander Gusak authored
      93d45d7a
  4. 02 Dec, 2025 2 commits
  5. 29 Nov, 2025 1 commit
  6. 26 Nov, 2025 1 commit
  7. 18 Nov, 2025 1 commit
  8. 17 Nov, 2025 1 commit
  9. 14 Nov, 2025 1 commit
  10. 13 Nov, 2025 2 commits
  11. 12 Nov, 2025 1 commit
  12. 11 Nov, 2025 4 commits
  13. 08 Nov, 2025 1 commit
  14. 07 Nov, 2025 2 commits
  15. 05 Nov, 2025 1 commit
    • nicole pardal's avatar
      embeddings: added embedding command for cl (#12795) · 1ca608bc
      nicole pardal authored
      
      Co-authored-by: default avatarA-Akhil <akhilrahul70@gmail.com>
      
      This PR introduces a new ollama embed command that allows users to generate embeddings directly from the command line.
      
      Added ollama embed MODEL [TEXT...] command for generating text embeddings
      Supports both direct text arguments and stdin piping for scripted workflows
      
      Outputs embeddings as JSON arrays (one per line)
      1ca608bc
  16. 29 Oct, 2025 3 commits
  17. 28 Oct, 2025 5 commits
  18. 16 Oct, 2025 1 commit
  19. 11 Oct, 2025 1 commit
  20. 07 Oct, 2025 2 commits
    • Daniel Hiltgen's avatar
      303be930
    • Daniel Hiltgen's avatar
      Bring back escape valve for llm libraries and fix Jetpack6 crash (#12529) · bd15eba4
      Daniel Hiltgen authored
      * Bring back escape valve for llm libraries
      
      If the new discovery logic picks the wrong library, this gives users the
      ability to force a specific one using the same pattern as before. This
      can also potentially speed up bootstrap discovery if one of the libraries
      takes a long time to load and ultimately bind to no devices.  For example
      unsupported AMD iGPUS can sometimes take a while to discover and rule out.
      
      * Bypass extra discovery on jetpack systems
      
      On at least Jetpack6, cuda_v12 appears to expose the iGPU, but crashes later on in
      cublasInit so if we detect a Jetpack, short-circuit and use that variant.
      bd15eba4
  21. 02 Oct, 2025 1 commit
    • Daniel Hiltgen's avatar
      Update GGML to b6646 (#12245) · c68f367e
      Daniel Hiltgen authored
      Notable EOLs with this change:
      - MacOS v12 and v13 are no longer supported (v14+ required)
      - AMD gfx900 and gfx906 are no longer supported
      c68f367e
  22. 01 Oct, 2025 1 commit
    • Daniel Hiltgen's avatar
      Use runners for GPU discovery (#12090) · bc8909fb
      Daniel Hiltgen authored
      This revamps how we discover GPUs in the system by leveraging the Ollama
      runner.  This should eliminate inconsistency between our GPU discovery and the
      runners capabilities at runtime, particularly for cases where we try to filter
      out unsupported GPUs.  Now the runner does that implicitly based on the actual
      device list.  In some cases free VRAM reporting can be unreliable which can
      leaad to scheduling mistakes, so this also includes a patch to leverage more
      reliable VRAM reporting libraries if available.
      
      Automatic workarounds have been removed as only one GPU leveraged this, which
      is now documented. This GPU will soon fall off the support matrix with the next
      ROCm bump.
      
      Additional cleanup of the scheduler and discovery packages can be done in the
      future once we have switched on the new memory management code, and removed
      support for the llama runner.
      bc8909fb
  23. 22 Sep, 2025 2 commits
  24. 15 Sep, 2025 1 commit
  25. 11 Sep, 2025 1 commit