- 11 Mar, 2025 16 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Patrick Devine authored
-
Michael Yang authored
-
Michael Yang authored
-
Jesse Gross authored
-
Patrick Devine authored
-
Michael Yang authored
-
Patrick Devine authored
-
Jesse Gross authored
-
Michael Yang authored
-
Patrick Devine authored
-
Daniel Hiltgen authored
-
- 10 Mar, 2025 10 commits
-
-
Michael Yang authored
fix: pad tensor item if ge zero
-
Michael Yang authored
this produces a nicer output since both positive and negative values produces the same width
-
Vincent Koc authored
-
Parth Sareen authored
-
Michael Yang authored
Better WantedBy declaration
-
frob authored
-
Xiaowei Zhu authored
-
Sam authored
-
Jeffrey Morgan authored
-
Jesse Gross authored
The encoder cache needs to know the position of images in the input stream so that it knows when to delete them. Previously images didn't have a position, so we implied one by breaking batches before an image and then assuming the image was in the first position. However, multimodal objects are now given explicit positions in the input stream, so we can use that instead. Breaking batches was also a way to simulate a cross attention mask for mllama. However, given that it only supports a single sequence and a single image, this mask doesn't serve any real purpose. Removing the batch break does not appear to affect the quality of the output. Most of this is simply moving the input data structures to a new package to avoid import cycles.
-
- 09 Mar, 2025 1 commit
-
-
Jesse Gross authored
It's ok to fail on startup but we shouldn't panic during runtime based on user input. Downgrade the panic to a warning.
-
- 08 Mar, 2025 5 commits
-
-
Jesse Gross authored
Similar to the llama engine, quantizing the KV cache requires flash attention to be enabled through the Ollama server.
-
Jesse Gross authored
-
Jesse Gross authored
Backends can impose additional alignment requirements on buffer sizes. We should ensure that we meet these or allocations can fail.
-
Jesse Gross authored
Models can disable causality for all or part of their processing while continuing to store data in the KV cache.
-
Jesse Gross authored
Debug logging of every token has previously caused test timeouts on slower machines.
-
- 07 Mar, 2025 8 commits
-
-
Jesse Gross authored
-
Michael Yang authored
this ensures the tensor is created on the right buffer type for backends such as cpu
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
temporary until tensor loading can accurately account for vision models
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-