- 01 Sep, 2023 1 commit
-
-
Quinn Slack authored
The `html/template` package automatically HTML-escapes interpolated strings in templates. This behavior is undesirable because it causes prompts like `<h1>hello` to be escaped to `<h1>hello` before being passed to the LLM. The included test case passes, but before the code change, it failed: ``` --- FAIL: TestModelPrompt images_test.go:21: got "a<h1>b", want "a<h1>b" ```
-
- 31 Aug, 2023 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 30 Aug, 2023 2 commits
-
-
Bruce MacDonald authored
* remove c code * pack llama.cpp * use request context for llama_cpp * let llama_cpp decide the number of threads to use * stop llama runner when app stops * remove sample count and duration metrics * use go generate to get libraries * tmp dir for running llm
-
Quinn Slack authored
The `stop` option to the generate API is a list of sequences that should cause generation to stop. Although these are commonly called "stop tokens", they do not necessarily correspond to LLM tokens (per the LLM's tokenizer). For example, if the caller sends a generate request with `"stop":["\n"]`, then generation should stop on any token containing `\n` (and trim `\n` from the output), not just if the token exactly matches `\n`. If `stop` were interpreted strictly as LLM tokens, then it would require callers of the generate API to know the LLM's tokenizer and enumerate many tokens in the `stop` list. Fixes https://github.com/jmorganca/ollama/issues/295.
-
- 29 Aug, 2023 1 commit
-
-
Patrick Devine authored
-
- 28 Aug, 2023 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 26 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 22 Aug, 2023 6 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Jeffrey Morgan authored
-
Ryan Baker authored
-
- 18 Aug, 2023 2 commits
-
-
Michael Yang authored
The token printed for authorized requests has a lifetime of 1h. If an upload exceeds 1h, a chunk push will fail since the token is created on a "start upload" request. This replaces the Pipe with SectionReader which is simpler and implements Seek, a requirement for makeRequestWithRetry. This is slightly worse than using a Pipe since the progress update is directly tied to the chunk size instead of controlled separately.
-
Michael Yang authored
-
- 17 Aug, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 16 Aug, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 15 Aug, 2023 3 commits
-
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 14 Aug, 2023 5 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
- re-use previously evaluated embeddings when possible - change embeddings digest identifier to be based on model name and embedded file path
-
- 11 Aug, 2023 1 commit
-
-
Patrick Devine authored
-
- 10 Aug, 2023 3 commits
-
-
Patrick Devine authored
-
Michael Yang authored
-
Michael Yang authored
-
- 09 Aug, 2023 3 commits
-
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 08 Aug, 2023 2 commits
-
-
Bruce MacDonald authored
- defer closing llm on embedding - do not override licenses - remove debugging print line - reformat model file docs
-
Bruce MacDonald authored
-