"vscode:/vscode.git/clone" did not exist on "51b53ddb6c3aa77426c7d5cc0b543b79628053c4"
- 02 Sep, 2024 1 commit
-
-
SnoopyTlion authored
-
- 01 Sep, 2024 2 commits
-
-
Vimal Kumar authored
-
rayfiyo authored
-
- 31 Aug, 2024 1 commit
-
-
Daniel Hiltgen authored
This caused missing internal files
-
- 30 Aug, 2024 3 commits
-
-
Michael Yang authored
remove any unneeded build artifacts
-
Michael Yang authored
passthrough OLLAMA_HOST path to client
-
Michael Yang authored
update templates to use messages
-
- 29 Aug, 2024 3 commits
-
-
Michael Yang authored
-
Bryan Honof authored
-
Patrick Devine authored
-
- 28 Aug, 2024 8 commits
-
-
Michael Yang authored
fix(test): do not clobber models directory
-
Michael Yang authored
fix: validate modelpath
-
Michael Yang authored
-
Patrick Devine authored
-
Michael Yang authored
detect chat template from configs that contain lists
-
Michael Yang authored
-
Michael Yang authored
-
Patrick Devine authored
-
- 27 Aug, 2024 12 commits
-
-
Daniel Hiltgen authored
-
Michael Yang authored
-
Michael Yang authored
-
Patrick Devine authored
-
Patrick Devine authored
-
Daniel Hiltgen authored
-
Sean Khatiri authored
-
Patrick Devine authored
-
Michael Yang authored
-
Michael Yang authored
-
Patrick Devine authored
-
Jeffrey Morgan authored
-
- 25 Aug, 2024 1 commit
-
-
Daniel Hiltgen authored
The numa flag may be having a performance impact on multi-socket systems with GPU loads
-
- 23 Aug, 2024 7 commits
-
-
Daniel Hiltgen authored
The recent cuda variant changes uncovered a bug in ByLibrary which failed to group by common variant for GPU types.
-
Michael Yang authored
update faq
-
Michael Yang authored
-
Michael Yang authored
-
Patrick Devine authored
-
Daniel Hiltgen authored
During rebasing, the ordering was inverted causing the cuda version selection logic to break, with driver version being evaluated as zero incorrectly causing a downgrade to v11.
-
Daniel Hiltgen authored
Define changed recently and this slipped through the cracks with the old name.
-
- 22 Aug, 2024 1 commit
-
-
Daniel Hiltgen authored
* Fix embeddings memory corruption The patch was leading to a buffer overrun corruption. Once removed though, parallism in server.cpp lead to hitting an assert due to slot/seq IDs being >= token count. To work around this, only use slot 0 for embeddings. * Fix embed integration test assumption The token eval count has changed with recent llama.cpp bumps (0.3.5+)
-
- 21 Aug, 2024 1 commit
-
-
Michael Yang authored
convert: update llama conversion for llama3.1
-