- 19 Dec, 2023 2 commits
-
-
Daniel Hiltgen authored
Run the server.cpp directly inside the Go runtime via cgo while retaining the LLM Go abstractions.
-
Bruce MacDonald authored
- remove ggml runner - automatically pull gguf models when ggml detected - tell users to update to gguf in the case automatic pull fails Co-Authored-By:Jeffrey Morgan <jmorganca@gmail.com>
-
- 18 Dec, 2023 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 13 Dec, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 04 Dec, 2023 1 commit
-
-
Michael Yang authored
-
- 26 Nov, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 24 Nov, 2023 2 commits
-
-
Jing Zhang authored
* Support cuda build in Windows * Enable dynamic NumGPU allocation for Windows
-
Jongwook Choi authored
When CUDA peer access is enabled, multi-gpu inference will produce garbage output. This is a known bug of llama.cpp (or nvidia). Until the upstream bug is fixed, we can disable CUDA peer access temporarily to ensure correct output. See #961.
-
- 22 Nov, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 21 Nov, 2023 1 commit
-
-
Michael Yang authored
-
- 20 Nov, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 17 Nov, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 27 Oct, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 24 Oct, 2023 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 23 Oct, 2023 2 commits
-
-
Michael Yang authored
pin to 9e70cc03229df19ca2d28ce23cc817198f897278 for now since 438c2ca83045a00ef244093d27e9ed41a8cb4ea9 is breaking
-
Michael Yang authored
-
- 17 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 06 Oct, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
- this makes it easier to see that the subprocess is associated with ollama
-
- 21 Sep, 2023 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 20 Sep, 2023 6 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 18 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* subprocess improvements - increase start-up timeout - when runner fails to start fail rather than timing out - try runners in order rather than choosing 1 runner - embed metal runner in metal dir rather than gpu - refactor logging and error messages * Update llama.go * Update llama.go * simplify by using glob
-
- 14 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* enable packaging multiple cuda versions * use nvcc cuda version if available --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 12 Sep, 2023 2 commits
-
-
Bruce MacDonald authored
-
Bruce MacDonald authored
* linux gpu support * handle multiple gpus * add cuda docker image (#488) --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 07 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 06 Sep, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 05 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
-