- 27 Oct, 2023 3 commits
-
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 24 Oct, 2023 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 23 Oct, 2023 3 commits
-
-
Michael Yang authored
pin to 9e70cc03229df19ca2d28ce23cc817198f897278 for now since 438c2ca83045a00ef244093d27e9ed41a8cb4ea9 is breaking
-
Michael Yang authored
-
Michael Yang authored
ggufv3 adds support for big endianness, mainly for s390x architecture. while that's not currently supported for ollama, the change is simple. loosen version check to be more forward compatible. unless specified, gguf versions other v1 will be decoded into v2.
-
- 19 Oct, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
add error for falcon and starcoder vocab compatibility --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-
- 18 Oct, 2023 6 commits
-
-
Arne Müller authored
-
Arne Müller authored
-
Arne Müller authored
-
Bruce MacDonald authored
-
Arne Müller authored
-
Arne Müller authored
-
- 17 Oct, 2023 5 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Arne Müller authored
-
- 16 Oct, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
omitting `--n-gpu-layers` means use metal on macos which isn't correct since ollama uses `num_gpu=0` to explicitly disable gpu for file types that are not implemented in metal
-
Arne Müller authored
-
- 13 Oct, 2023 8 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
- remove new lines from llama.cpp error messages relayed to client - check api option types and return error on wrong type - change num layers from 95% VRAM to 92% VRAM
-
- 12 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* give direction to user when runner fails * also relay errors from timeout * increase timeout to 3 minutes
-
- 11 Oct, 2023 2 commits
-
-
Michael Yang authored
-
Bruce MacDonald authored
* prevent waiting on exited command * close llama runner once
-
- 10 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* check free memory not total * wait for subprocess to exit
-
- 06 Oct, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
- this makes it easier to see that the subprocess is associated with ollama
-
- 05 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-