- 23 Oct, 2023 1 commit
-
-
Michael Yang authored
-
- 19 Oct, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
add error for falcon and starcoder vocab compatibility --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-
- 18 Oct, 2023 6 commits
-
-
Arne Müller authored
-
Arne Müller authored
-
Arne Müller authored
-
Bruce MacDonald authored
-
Arne Müller authored
-
Arne Müller authored
-
- 17 Oct, 2023 5 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Arne Müller authored
-
- 16 Oct, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
omitting `--n-gpu-layers` means use metal on macos which isn't correct since ollama uses `num_gpu=0` to explicitly disable gpu for file types that are not implemented in metal
-
Arne Müller authored
-
- 13 Oct, 2023 8 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
- remove new lines from llama.cpp error messages relayed to client - check api option types and return error on wrong type - change num layers from 95% VRAM to 92% VRAM
-
- 12 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* give direction to user when runner fails * also relay errors from timeout * increase timeout to 3 minutes
-
- 11 Oct, 2023 2 commits
-
-
Michael Yang authored
-
Bruce MacDonald authored
* prevent waiting on exited command * close llama runner once
-
- 10 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* check free memory not total * wait for subprocess to exit
-
- 06 Oct, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
- this makes it easier to see that the subprocess is associated with ollama
-
- 05 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 04 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 03 Oct, 2023 1 commit
-
-
Michael Yang authored
-
- 02 Oct, 2023 2 commits
-
-
Bruce MacDonald authored
-
Bruce MacDonald authored
* include seed in params for llama.cpp server and remove empty filter for temp * relay default predict options to llama.cpp - reorganize options to match predict request for readability * omit empty stop --------- Co-authored-by:hallh <hallh@users.noreply.github.com>
-
- 29 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 28 Sep, 2023 1 commit
-
-
Michael Yang authored
-
- 25 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
--------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 21 Sep, 2023 1 commit
-
-
Michael Yang authored
-