1. 01 Nov, 2024 1 commit
  2. 18 Oct, 2024 1 commit
  3. 17 Oct, 2024 1 commit
  4. 06 Sep, 2024 1 commit
  5. 05 Sep, 2024 1 commit
    • Daniel Hiltgen's avatar
      Introduce GPU Overhead env var (#5922) · b05c9e83
      Daniel Hiltgen authored
      Provide a mechanism for users to set aside an amount of VRAM on each GPU
      to make room for other applications they want to start after Ollama, or workaround
      memory prediction bugs
      b05c9e83
  6. 20 Jun, 2024 1 commit
  7. 18 Jun, 2024 2 commits
    • Daniel Hiltgen's avatar
      Handle models with divergent layer sizes · 359b15a5
      Daniel Hiltgen authored
      The recent refactoring of the memory prediction assumed all layers
      are the same size, but for some models (like deepseek-coder-v2) this
      is not the case, so our predictions were significantly off.
      359b15a5
    • Daniel Hiltgen's avatar
      Tighten up memory prediction logging · 7784ca33
      Daniel Hiltgen authored
      Prior to this change, we logged the memory prediction multiple times
      as the scheduler iterates to find a suitable configuration, which can be
      confusing since only the last log before the server starts is actually valid.
      This now logs once just before starting the server on the final configuration.
      It also reports what library instead of always saying "offloading to gpu" when
      using CPU.
      7784ca33
  8. 14 Jun, 2024 3 commits
  9. 04 Jun, 2024 2 commits
  10. 24 May, 2024 1 commit
  11. 13 May, 2024 2 commits
  12. 10 May, 2024 1 commit
  13. 08 May, 2024 1 commit
  14. 07 May, 2024 1 commit
  15. 05 May, 2024 1 commit
    • Daniel Hiltgen's avatar
      Centralize server config handling · f56aa200
      Daniel Hiltgen authored
      This moves all the env var reading into one central module
      and logs the loaded config once at startup which should
      help in troubleshooting user server logs
      f56aa200
  16. 01 May, 2024 1 commit
  17. 26 Apr, 2024 1 commit
  18. 25 Apr, 2024 1 commit
  19. 24 Apr, 2024 1 commit
    • Daniel Hiltgen's avatar
      Add back memory escape valve · 5445aaa9
      Daniel Hiltgen authored
      If we get our predictions wrong, this can be used to
      set a lower memory limit as a workaround.  Recent multi-gpu
      refactoring accidentally removed it, so this adds it back.
      5445aaa9
  20. 23 Apr, 2024 1 commit
    • Daniel Hiltgen's avatar
      Request and model concurrency · 34b9db5a
      Daniel Hiltgen authored
      This change adds support for multiple concurrent requests, as well as
      loading multiple models by spawning multiple runners. The default
      settings are currently set at 1 concurrent request per model and only 1
      loaded model at a time, but these can be adjusted by setting
      OLLAMA_NUM_PARALLEL and OLLAMA_MAX_LOADED_MODELS.
      34b9db5a