"src/diffusers/quantizers/quanto/__init__.py" did not exist on "9a1810f0de807f936ac3cf344d6e1e2851af723a"
- 20 Sep, 2023 6 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 18 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* subprocess improvements - increase start-up timeout - when runner fails to start fail rather than timing out - try runners in order rather than choosing 1 runner - embed metal runner in metal dir rather than gpu - refactor logging and error messages * Update llama.go * Update llama.go * simplify by using glob
-
- 14 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
* enable packaging multiple cuda versions * use nvcc cuda version if available --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 13 Sep, 2023 1 commit
-
-
Michael Yang authored
-
- 12 Sep, 2023 4 commits
-
-
Michael Yang authored
-
Bruce MacDonald authored
-
Michael Yang authored
get model and file type from bin file
-
Bruce MacDonald authored
* linux gpu support * handle multiple gpus * add cuda docker image (#488) --------- Co-authored-by:Michael Yang <mxyng@pm.me>
-
- 07 Sep, 2023 1 commit
-
-
Bruce MacDonald authored
-
- 06 Sep, 2023 3 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
- 05 Sep, 2023 3 commits
-
-
Bruce MacDonald authored
-
Michael Yang authored
-
Jeffrey Morgan authored
-
- 03 Sep, 2023 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
- 30 Aug, 2023 3 commits
-
-
Bruce MacDonald authored
-
Bruce MacDonald authored
* remove c code * pack llama.cpp * use request context for llama_cpp * let llama_cpp decide the number of threads to use * stop llama runner when app stops * remove sample count and duration metrics * use go generate to get libraries * tmp dir for running llm
-
Quinn Slack authored
The `stop` option to the generate API is a list of sequences that should cause generation to stop. Although these are commonly called "stop tokens", they do not necessarily correspond to LLM tokens (per the LLM's tokenizer). For example, if the caller sends a generate request with `"stop":["\n"]`, then generation should stop on any token containing `\n` (and trim `\n` from the output), not just if the token exactly matches `\n`. If `stop` were interpreted strictly as LLM tokens, then it would require callers of the generate API to know the LLM's tokenizer and enumerate many tokens in the `stop` list. Fixes https://github.com/jmorganca/ollama/issues/295.
-
- 26 Aug, 2023 3 commits
-
-
Michael Yang authored
warning F16 uses significantly more memory than quantized model so the standard requires don't apply.
-
Michael Yang authored
-
Jeffrey Morgan authored
-
- 25 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 24 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 18 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 17 Aug, 2023 1 commit
-
-
Michael Yang authored
-
- 14 Aug, 2023 4 commits
-
-
Michael Yang authored
-
Michael Yang authored
-
Bruce MacDonald authored
-
Bruce MacDonald authored
-
- 13 Aug, 2023 1 commit
-
-
Jeffrey Morgan authored
-
- 11 Aug, 2023 1 commit
-
-
Michael Yang authored
remove used Unknown FileType
-
- 10 Aug, 2023 2 commits
-
-
Michael Yang authored
-
Michael Yang authored
-