1. 14 Jun, 2024 1 commit
    • Daniel Hiltgen's avatar
      Improve multi-gpu handling at the limit · 6fd04ca9
      Daniel Hiltgen authored
      Still not complete, needs some refinement to our prediction to understand the
      discrete GPUs available space so we can see how many layers fit in each one
      since we can't split one layer across multiple GPUs we can't treat free space
      as one logical block
      6fd04ca9
  2. 04 Jun, 2024 2 commits
  3. 24 May, 2024 1 commit
  4. 13 May, 2024 2 commits
  5. 10 May, 2024 1 commit
  6. 08 May, 2024 1 commit
  7. 07 May, 2024 1 commit
  8. 05 May, 2024 1 commit
    • Daniel Hiltgen's avatar
      Centralize server config handling · f56aa200
      Daniel Hiltgen authored
      This moves all the env var reading into one central module
      and logs the loaded config once at startup which should
      help in troubleshooting user server logs
      f56aa200
  9. 01 May, 2024 1 commit
  10. 26 Apr, 2024 1 commit
  11. 25 Apr, 2024 1 commit
  12. 24 Apr, 2024 1 commit
    • Daniel Hiltgen's avatar
      Add back memory escape valve · 5445aaa9
      Daniel Hiltgen authored
      If we get our predictions wrong, this can be used to
      set a lower memory limit as a workaround.  Recent multi-gpu
      refactoring accidentally removed it, so this adds it back.
      5445aaa9
  13. 23 Apr, 2024 1 commit
    • Daniel Hiltgen's avatar
      Request and model concurrency · 34b9db5a
      Daniel Hiltgen authored
      This change adds support for multiple concurrent requests, as well as
      loading multiple models by spawning multiple runners. The default
      settings are currently set at 1 concurrent request per model and only 1
      loaded model at a time, but these can be adjusted by setting
      OLLAMA_NUM_PARALLEL and OLLAMA_MAX_LOADED_MODELS.
      34b9db5a