- 25 Feb, 2025 7 commits
-
-
José Pekkarinen authored
centos-7 images have been deprecated upstream and replaced with almalinux-8 images instead, requiring some small extra work. Signed-off-by:José Pekkarinen <jose.pekkarinen@foxhound.fi>
-
Chuanhui Liu authored
-
Michael Yang authored
this was accidentally removed when moving fs/ggml from its previous location
-
Pavol Rusnak authored
CUDA 12.x still supports Compute Capability 5.0, 5.2 and 5.3, so let's build for these architectures as well
-
frob authored
Co-authored-by:Richard Lyons <frob@cloudstaff.com>
-
Blake Mizerany authored
This commit copies (without history) the bmizerany/ollama-go repository with the intention of integrating it into the ollama as a replacement for the pushing, and pulling of models, and management of the cache they are pushed and pulled from. New homes for these packages will be determined as they are integrated and we have a better understanding of proper package boundaries.
-
Parth Sareen authored
-
- 24 Feb, 2025 3 commits
-
-
Parth Sareen authored
* envconfig: allow setting context length through env var
-
Blake Mizerany authored
-
Jeffrey Morgan authored
-
- 22 Feb, 2025 2 commits
-
-
Jeffrey Morgan authored
-
Blake Mizerany authored
The route assembly in Handler lacked clear organization making it difficult scan for routes and their relationships to each other. This commit aims to fix that by reordering the assembly of routes to group them by category and purpose. Also, be more specific about what "config" refers to (it is about CORS if you were wondering... I was.)
-
- 21 Feb, 2025 3 commits
-
-
Jesse Gross authored
There are two benefits to doing this: - Provide a library function that models can use, reducing code for each model implementation - Enables a single place to drop in optimized implementations of attention based on the backend or other factors. One is provided for GGML. On CUDA this improves token generation rate by about 3%. It does not have a significant effect on Metal. Co-authored-by:Daniel Hiltgen <daniel@ollama.com>
-
Michael Yang authored
-
Junyan Qin (Chin) authored
-
- 20 Feb, 2025 9 commits
-
-
Jesse Gross authored
Currently Rows is called as the last step in a model computation to get the values for the output tokens. However, if we move it earlier in the process then we can trim out computations that never get used. This is similar to how models are defined in llama.cpp. Changing the model definition in this way improves token generation performance by approximately 8%.
-
Jesse Gross authored
We don't need to create and destroy the GGML scheduler for every context. This introduces extra CPU overhead for every forward pass and extra memory for contexts that don't actually get scheduled (for example, KV caches). We can instead just have one scheduler for the backend and reset it each time we call Compute. This improves token generation performance by 1-2% and removes scheduler create/destroy from profile traces.
-
Jesse Gross authored
Currently the following parameters are in the runner but not used: - numGPULayers - mainGPU - threads - tensorSplit This passes them through to the backend, which is where they would actually get used. However, the GGML backend does not yet do anything with them.
-
Bruce MacDonald authored
Added unit tests to verify error handling behavior in the Client.stream and Client.do methods. Tests cover various error scenarios including: - Error responses with status codes >= 400 - Error messages with successful status codes - Empty error messages - Successful responses
-
Michael Yang authored
clang outputs are faster. we were previously building with clang via gcc wrapper in cgo but this was missed during the build updates so there was a drop in performance
-
frob authored
-
danielekp authored
-
Lucas Hahn authored
-
Michael Yang authored
-
- 19 Feb, 2025 5 commits
-
-
Michael Yang authored
build: remove backend build for sapphirerapids
-
yuiseki authored
-
zyxucp authored
-
maninhill authored
-
Jeffrey Morgan authored
-
- 18 Feb, 2025 9 commits
-
-
Michael Yang authored
cmd: fix flickering in progress bar
-
Jeremy Schlatter authored
-
Michael Yang authored
sapphire rapids has amx support but it ends up having a negative performance impact. emerald rapids also has amx support with a positive performance impact however there's no reasonable way in ggml to differentiate between the two. the impact is small (~6%) so disable amx entirely for simplicity
-
Michael Yang authored
-
Michael Yang authored
set owner and group when building the linux tarball so extracted files are consistent. this is the behaviour of release tarballs in version 0.5.7 and lower
-
benhaotang authored
-
L. Jiang authored
-
innightwolfsleep authored
-
Jeremy Schlatter authored
-
- 17 Feb, 2025 2 commits
-
-
Jeremy Schlatter authored
The previous commit fixed flickering in the progress bar itself. Cursor flickering is harder to address. Cursor flickering could be fixed by hiding the cursor altogether while the progress bar is displayed. The downside of this is that if the program is killed in such a way that it can't clean up its state, it would leave the cursor invisible. Instead, this commit introduces an output buffer. All of the escape codes and content for a single progress update are written to a buffer, which is then flushed to the terminal all at once. This significantly decreases the time during which the terminal has seen the cursor-hiding code but has not yet seen the cursor-showing code, thus minimizing (but not 100% eliminating) cursor flickering. For more context, see: https://gitlab.gnome.org/GNOME/vte/-/issues/2837#note_2269501
-
Jeremy Schlatter authored
Previous code cleared the display before writing new content, creating a window where the terminal could (and in some cases did) render empty lines. Instead, we now write new content over the old content, only clearing the trailing end of lines for cases where the new line is shorter. Fixes #1664
-