- 05 Sep, 2023 8 commits
-
-
Michael Yang authored
-
Michael Yang authored
fix inherit params
-
Michael Yang authored
remove marshalPrompt which is no longer needed
-
Michael Yang authored
parameters are not inherited because they are processed differently from other layer. fix this by explicitly merging the inherited params into the new params. parameter values defined in the new modelfile will override those defined in the inherited modelfile. array lists are replaced instead of appended
-
Michael Yang authored
fix num_keep
-
Michael Yang authored
-
Jeffrey Morgan authored
-
Michael Yang authored
fix not forwarding last token
-
- 03 Sep, 2023 3 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 02 Sep, 2023 1 commit
-
-
Michael Yang authored
do not HTML-escape prompt
-
- 01 Sep, 2023 4 commits
-
-
Quinn Slack authored
The `html/template` package automatically HTML-escapes interpolated strings in templates. This behavior is undesirable because it causes prompts like `<h1>hello` to be escaped to `<h1>hello` before being passed to the LLM. The included test case passes, but before the code change, it failed: ``` --- FAIL: TestModelPrompt images_test.go:21: got "a<h1>b", want "a<h1>b" ``` -
Michael Yang authored
* update readme * readme: more run examples
-
Michael Yang authored
update readme
-
Michael Yang authored
-
- 31 Aug, 2023 9 commits
-
-
Matt Williams authored
Create a sentiments example
-
Jeffrey Morgan authored
-
Michael Yang authored
windows: fix filepath bugs
-
Michael Yang authored
fix spelling errors in example prompts
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Philip Nuzhnyi authored
-
- 30 Aug, 2023 6 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
* remove c code * pack llama.cpp * use request context for llama_cpp * let llama_cpp decide the number of threads to use * stop llama runner when app stops * remove sample count and duration metrics * use go generate to get libraries * tmp dir for running llm
-
Quinn Slack authored
The `stop` option to the generate API is a list of sequences that should cause generation to stop. Although these are commonly called "stop tokens", they do not necessarily correspond to LLM tokens (per the LLM's tokenizer). For example, if the caller sends a generate request with `"stop":["\n"]`, then generation should stop on any token containing `\n` (and trim `\n` from the output), not just if the token exactly matches `\n`. If `stop` were interpreted strictly as LLM tokens, then it would require callers of the generate API to know the LLM's tokenizer and enumerate many tokens in the `stop` list. Fixes https://github.com/jmorganca/ollama/issues/295.
-
Michael Yang authored
update upload chunks
-
- 29 Aug, 2023 2 commits
-
-
Michael Yang authored
allow F16 to use metal
-
Patrick Devine authored
-
- 28 Aug, 2023 4 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
Michael Yang authored
-
- 27 Aug, 2023 2 commits
-
-
Jeffrey Morgan authored
-
Michael Yang authored
update README.md
-
- 26 Aug, 2023 1 commit
-
-
Michael Yang authored
add 34b to mem check
-