1. 06 Jan, 2026 1 commit
    • Devon Rifkin's avatar
      preserve tool definition and call JSON ordering (#13525) · e51dead6
      Devon Rifkin authored
      * preserve tool definition and call JSON ordering
      
      This is another iteration of
      <https://github.com/ollama/ollama/pull/12518>, but this time we've
      simplified things by relaxing the competing requirements of being
      compatible AND order-preserving with templates (vs. renderers). We
      maintain backwards compatibility at the cost of not guaranteeing order
      for templates. We plan on moving more and more models to renderers,
      which have been updated to use these new data types, and additionally
      we could add an opt-in way of templates getting an order-preserved list
      (e.g., via sibling template vars)
      
      * orderedmap_test: remove testify
      e51dead6
  2. 18 Dec, 2025 1 commit
  3. 04 Nov, 2025 1 commit
  4. 20 Jun, 2025 1 commit
  5. 18 Jun, 2025 1 commit
  6. 12 Jun, 2025 1 commit
  7. 29 Apr, 2025 1 commit
  8. 05 Mar, 2025 1 commit
    • Blake Mizerany's avatar
      server/internal/registry: take over pulls from server package (#9485) · e2252d0f
      Blake Mizerany authored
      This commit replaces the old pull implementation in the server package
      with the new, faster, more robust pull implementation in the registry
      package.
      
      The new endpoint, and now the remove endpoint too, are behind the
      feature gate "client2" enabled only by setting the OLLAMA_EXPERIMENT
      environment variable include "client2".
      
      Currently, the progress indication is wired to perform the same as the
      previous implementation to avoid making changes to the CLI, and because
      the status reports happen at the start of the download, and the end of
      the write to disk, the progress indication is not as smooth as it could
      be. This is a known issue and will be addressed in a future change.
      
      This implementation may be ~0.5-1.0% slower in rare cases, depending on
      network and disk speed, but is generally MUCH faster and more robust
      than the its predecessor in all other cases.
      e2252d0f
  9. 27 Feb, 2025 1 commit
    • Blake Mizerany's avatar
      server/internal: replace model delete API with new registry handler. (#9347) · 2412adf4
      Blake Mizerany authored
      This commit introduces a new API implementation for handling
      interactions with the registry and the local model cache. The new API is
      located in server/internal/registry. The package name is "registry" and
      should be considered temporary; it is hidden and not bleeding outside of
      the server package. As the commits roll in, we'll start consuming more
      of the API and then let reverse osmosis take effect, at which point it
      will surface closer to the root level packages as much as needed.
      2412adf4
  10. 29 Jan, 2025 1 commit
    • Michael Yang's avatar
      next build (#8539) · dcfb7a10
      Michael Yang authored
      
      
      * add build to .dockerignore
      
      * test: only build one arch
      
      * add build to .gitignore
      
      * fix ccache path
      
      * filter amdgpu targets
      
      * only filter if autodetecting
      
      * Don't clobber gpu list for default runner
      
      This ensures the GPU specific environment variables are set properly
      
      * explicitly set CXX compiler for HIP
      
      * Update build_windows.ps1
      
      This isn't complete, but is close.  Dependencies are missing, and it only builds the "default" preset.
      
      * build: add ollama subdir
      
      * add .git to .dockerignore
      
      * docs: update development.md
      
      * update build_darwin.sh
      
      * remove unused scripts
      
      * llm: add cwd and build/lib/ollama to library paths
      
      * default DYLD_LIBRARY_PATH to LD_LIBRARY_PATH in runner on macOS
      
      * add additional cmake output vars for msvc
      
      * interim edits to make server detection logic work with dll directories like lib/ollama/cuda_v12
      
      * remove unncessary filepath.Dir, cleanup
      
      * add hardware-specific directory to path
      
      * use absolute server path
      
      * build: linux arm
      
      * cmake install targets
      
      * remove unused files
      
      * ml: visit each library path once
      
      * build: skip cpu variants on arm
      
      * build: install cpu targets
      
      * build: fix workflow
      
      * shorter names
      
      * fix rocblas install
      
      * docs: clean up development.md
      
      * consistent build dir removal in development.md
      
      * silence -Wimplicit-function-declaration build warnings in ggml-cpu
      
      * update readme
      
      * update development readme
      
      * llm: update library lookup logic now that there is one runner (#8587)
      
      * tweak development.md
      
      * update docs
      
      * add windows cuda/rocm tests
      
      ---------
      Co-authored-by: default avatarjmorganca <jmorganca@gmail.com>
      Co-authored-by: default avatarDaniel Hiltgen <daniel@ollama.com>
      dcfb7a10
  11. 21 Dec, 2024 1 commit
  12. 20 Dec, 2024 1 commit
  13. 23 Nov, 2024 1 commit
  14. 22 Nov, 2024 1 commit
  15. 14 Nov, 2024 1 commit
  16. 18 Oct, 2024 1 commit
  17. 06 Jun, 2024 1 commit
  18. 21 May, 2024 1 commit
  19. 11 May, 2024 1 commit
  20. 15 Apr, 2024 1 commit
  21. 07 Mar, 2024 1 commit
  22. 24 Feb, 2024 1 commit
  23. 15 Feb, 2024 2 commits
  24. 19 Dec, 2023 1 commit
  25. 05 Dec, 2023 1 commit
  26. 14 Nov, 2023 1 commit
  27. 01 Nov, 2023 1 commit
  28. 25 Oct, 2023 2 commits
  29. 16 Oct, 2023 1 commit
  30. 06 Oct, 2023 1 commit
  31. 22 Sep, 2023 2 commits
  32. 05 Sep, 2023 1 commit
  33. 30 Aug, 2023 1 commit
    • Bruce MacDonald's avatar
      subprocess llama.cpp server (#401) · 42998d79
      Bruce MacDonald authored
      * remove c code
      * pack llama.cpp
      * use request context for llama_cpp
      * let llama_cpp decide the number of threads to use
      * stop llama runner when app stops
      * remove sample count and duration metrics
      * use go generate to get libraries
      * tmp dir for running llm
      42998d79
  34. 10 Aug, 2023 1 commit
  35. 08 Aug, 2023 1 commit
  36. 01 Aug, 2023 2 commits