- 08 May, 2025 1 commit
-
-
Michael Yang authored
-
- 09 Dec, 2024 1 commit
-
-
Jesse Gross authored
New lines can be an important part of a user's prompt and trimming it can alter the results. We previously only trimmed prompts with images but refactoring brought this behavior to all prompts, where it became more noticable. The /generate endpoint adds less whitespace and therefore doesn't need to trim it out - this brings the same behavior to /chat. Thanks to @gabe-l-hart for spotting the issue! Fixes #7795
-
- 17 Nov, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 30 Oct, 2024 1 commit
-
-
Jesse Gross authored
-Update mllama to take the cross attention state as embeddings in a batch, more similar to how Llava handles it. This improves integration with the input cache. -Pass locations in a prompt for embeddings using tags similar to Llava. -Abstract interface to vision models so the main runner accesses Clip and Mllama similarly Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 18 Oct, 2024 1 commit
-
-
Patrick Devine authored
Co-authored-by:
jmorganca <jmorganca@gmail.com> Co-authored-by:
Michael Yang <mxyng@pm.me> Co-authored-by:
Jesse Gross <jesse@ollama.com>
-
- 02 Aug, 2024 1 commit
-
-
Michael Yang authored
-
- 16 Jul, 2024 1 commit
-
-
Michael Yang authored
-
- 15 Jul, 2024 1 commit
-
-
Michael Yang authored
-
- 13 Jul, 2024 1 commit
-
-
Michael Yang authored
* fix system prompt * execute template when hitting previous roles * fix tests --------- Co-authored-by:jmorganca <jmorganca@gmail.com>
-
- 11 Jul, 2024 1 commit
-
-
Michael Yang authored
-
- 05 Jul, 2024 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 01 Jul, 2024 1 commit
-
-
Michael Yang authored
-
- 26 Mar, 2024 1 commit
-
-
Patrick Devine authored
-
- 29 Feb, 2024 1 commit
-
-
Michael Yang authored
instead of appending image tags, prepend them - this generally produces better results
-
- 16 Feb, 2024 1 commit
-
-
Bruce MacDonald authored
-
- 12 Feb, 2024 1 commit
-
-
Jeffrey Morgan authored
-