"vscode:/vscode.git/clone" did not exist on "ee041cec6af58ccd8a1135a1eb3f8039220ab043"
- 06 Apr, 2024 1 commit
-
-
Michael Yang authored
-
- 01 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 25 Feb, 2024 1 commit
-
-
Ikko Eltociear Ashimine authored
specfied -> specified
-
- 13 Feb, 2024 1 commit
-
-
bnorick authored
-
- 26 Jan, 2024 1 commit
-
-
Patrick Devine authored
-
- 25 Jan, 2024 1 commit
-
-
Patrick Devine authored
-
- 18 Jan, 2024 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
prefer using req.Model and fallback to req.Name
-
- 05 Jan, 2024 1 commit
-
-
Patrick Devine authored
-
- 27 Dec, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 22 Dec, 2023 1 commit
-
-
K0IN authored
-
- 18 Dec, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 12 Dec, 2023 1 commit
-
-
Patrick Devine authored
-
- 11 Dec, 2023 1 commit
-
-
Patrick Devine authored
--------- Co-authored-by:Matt Apperson <mattapperson@Matts-MacBook-Pro.local>
-
- 09 Dec, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 05 Dec, 2023 3 commits
-
-
Michael Yang authored
-
Bruce MacDonald authored
-
Jeffrey Morgan authored
This reverts commit 7a0899d6.
-
- 04 Dec, 2023 1 commit
-
-
Bruce MacDonald authored
- update chat docs - add messages chat endpoint - remove deprecated context and template generate parameters from docs - context and template are still supported for the time being and will continue to work as expected - add partial response to chat history
-
- 29 Nov, 2023 1 commit
-
-
Patrick Devine authored
-
- 15 Nov, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 10 Nov, 2023 1 commit
-
-
Jeffrey Morgan authored
* add `"format": "json"` as an API parameter --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-
- 09 Nov, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 08 Nov, 2023 1 commit
-
-
Bruce MacDonald authored
- add the optional `raw` generate request parameter to bypass prompt formatting and response context -add raw request to docs
-
- 03 Nov, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 02 Nov, 2023 1 commit
-
-
Michael Yang authored
-
- 19 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
- only reload the running llm if the model has changed, or the options for loading the running model have changed - rename loaded llm to runner to differentiate from loaded model image - remove logic which keeps the first system prompt in the generation context
-
- 13 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
- remove new lines from llama.cpp error messages relayed to client - check api option types and return error on wrong type - change num layers from 95% VRAM to 92% VRAM
-
- 12 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 11 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* update streaming request accept header * add optional stream param to request bodies
-
- 05 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 02 Oct, 2023 1 commit
-
-
Bruce MacDonald authored
* include seed in params for llama.cpp server and remove empty filter for temp * relay default predict options to llama.cpp - reorganize options to match predict request for readability * omit empty stop --------- Co-authored-by:hallh <hallh@users.noreply.github.com>
-
- 29 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 28 Sep, 2023 1 commit
-
-
Michael Yang authored
-
- 12 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* linux gpu support * handle multiple gpus * add cuda docker image (#488) --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 06 Sep, 2023 1 commit
-
-
Patrick Devine authored
-