- 06 May, 2024 11 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
- FROM /path/to/{safetensors,pytorch} - FROM /path/to/fp{16,32}.bin - FROM model:fp{16,32} -
Jeffrey Morgan authored
-
Daniel Hiltgen authored
The model processing was recently changed to be deferred but this test scenario hadn't been adjusted for that change in behavior.
-
Jeffrey Morgan authored
-
- 05 May, 2024 4 commits
-
-
Daniel Hiltgen authored
This moves all the env var reading into one central module and logs the loaded config once at startup which should help in troubleshooting user server logs
-
Jeffrey Morgan authored
-
Patrick Devine authored
-
Daniel Hiltgen authored
This also bumps up the default to be 50 queued requests instead of 10.
-
- 03 May, 2024 1 commit
-
-
Daniel Hiltgen authored
This gives us more headroom on the scheduler tests to tamp down some flakes.
-
- 01 May, 2024 6 commits
-
-
Michael Yang authored
-
Mark Ward authored
log when the waiting for the process to stop to help debug when other tasks execute during this wait. expire timer clear the timer reference because it will not be reused. close will clean up expireTimer if calling code has not already done this.
-
Mark Ward authored
fix runner expire during active use. Clearing the expire timer as it is used. Allowing the finish to assign an expire timer so that the runner will expire after no use.
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 30 Apr, 2024 1 commit
-
-
Bruce MacDonald authored
- return descriptive error messages when unauthorized to create blob or push a model - display the local public key associated with the request that was denied
-
- 29 Apr, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 28 Apr, 2024 1 commit
-
-
Daniel Hiltgen authored
Prior refactoring passes accidentally removed the logic to bypass VRAM checks for CPU loads. This adds that back, along with test coverage. This also fixes loaded map access in the unit test to be behind the mutex which was likely the cause of various flakes in the tests.
-
- 26 Apr, 2024 2 commits
-
-
Jeffrey Morgan authored
-
Blake Mizerany authored
-
- 25 Apr, 2024 2 commits
-
-
Jeffrey Morgan authored
* reload model if `num_gpu` changes * dont reload on -1 * fix tests
-
Daniel Hiltgen authored
-
- 24 Apr, 2024 4 commits
-
-
Bryce Reitano authored
-
Bryce Reitano authored
-
Bryce Reitano authored
-
Michael Yang authored
-
- 23 Apr, 2024 2 commits
-
-
Daniel Hiltgen authored
Give the go routine a moment to deliver the expired event
-
Daniel Hiltgen authored
This change adds support for multiple concurrent requests, as well as loading multiple models by spawning multiple runners. The default settings are currently set at 1 concurrent request per model and only 1 loaded model at a time, but these can be adjusted by setting OLLAMA_NUM_PARALLEL and OLLAMA_MAX_LOADED_MODELS.
-
- 21 Apr, 2024 1 commit
-
-
Cheng authored
-
- 15 Apr, 2024 2 commits
-
-
Patrick Devine authored
-
Jeffrey Morgan authored
* terminate subprocess if receiving `SIGINT` or `SIGTERM` signals while model is loading * use `unload` in signal handler
-
- 10 Apr, 2024 1 commit
-
-
Blake Mizerany authored
This is a quick fix to help users who are stuck on the "pull" step at 99%. In the near future we're introducing a new registry client that should/will hopefully be smarter. In the meantime, this should unblock the users hitting issue #1736.
-
- 08 Apr, 2024 1 commit
-
-
Michael Yang authored
-