- 23 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 21 May, 2024 1 commit
-
-
Patrick Devine authored
-
- 20 May, 2024 1 commit
-
-
alwqx authored
-
- 09 May, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 01 Apr, 2024 1 commit
-
-
Daniel Hiltgen authored
We may have users that run into problems with our current payload model, so this gives us an escape valve.
-
- 21 Mar, 2024 1 commit
-
-
Bruce MacDonald authored
-
- 15 Mar, 2024 1 commit
-
-
Daniel Hiltgen authored
-
- 11 Mar, 2024 1 commit
-
-
Jeffrey Morgan authored
-
- 07 Mar, 2024 2 commits
-
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This refines where we extract the LLM libraries to by adding a new OLLAMA_HOME env var, that defaults to `~/.ollama` The logic was already idempotenent, so this should speed up startups after the first time a new release is deployed. It also cleans up after itself. We now build only a single ROCm version (latest major) on both windows and linux. Given the large size of ROCms tensor files, we split the dependency out. It's bundled into the installer on windows, and a separate download on windows. The linux install script is now smart and detects the presence of AMD GPUs and looks to see if rocm v6 is already present, and if not, then downloads our dependency tar file. For Linux discovery, we now use sysfs and check each GPU against what ROCm supports so we can degrade to CPU gracefully instead of having llama.cpp+rocm assert/crash on us. For Windows, we now use go's windows dynamic library loading logic to access the amdhip64.dll APIs to query the GPU information.
-
- 15 Feb, 2024 1 commit
-
-
Daniel Hiltgen authored
This focuses on Windows first, but coudl be used for Mac and possibly linux in the future.
-
- 29 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
Some users are new to containers and unsure where the server logs go
-
- 11 Jan, 2024 1 commit
-
-
Daniel Hiltgen authored
This reduces the built-in linux version to not use any vector extensions which enables the resulting builds to run under Rosetta on MacOS in Docker. Then at runtime it checks for the actual CPU vector extensions and loads the best CPU library available
-
- 22 Dec, 2023 1 commit
-
-
Matt Williams authored
* Clean up documentation Will probably need to update with PRs for new release. Signed-off-by:
Matt Williams <m@technovangelist.com> * Correcting to fit in 0.1.15 changes Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * addressing comments Signed-off-by:
Matt Williams <m@technovangelist.com> * more api cleanup Signed-off-by:
Matt Williams <m@technovangelist.com> * its llava not llama Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Updated hosting to server and documented all env vars Signed-off-by:
Matt Williams <m@technovangelist.com> * remove last of the cli descriptions Signed-off-by:
Matt Williams <m@technovangelist.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update further per conversation with jeff earlier today Signed-off-by:
Matt Williams <m@technovangelist.com> * cleanup the doc readme Signed-off-by:
Matt Williams <m@technovangelist.com> * move upgrade to faq Signed-off-by:
Matt Williams <m@technovangelist.com> * first change Signed-off-by:
Matt Williams <m@technovangelist.com> * updated Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * examples in parent Signed-off-by:
Matt Williams <m@technovangelist.com> * add exapmle for create model. Signed-off-by:
Matt Williams <m@technovangelist.com> * update faq Signed-off-by:
Matt Williams <m@technovangelist.com> * update create model api Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/api.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * update the readme in docs Signed-off-by:
Matt Williams <m@technovangelist.com> * update a few more things Signed-off-by:
Matt Williams <m@technovangelist.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/faq.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update README.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/modelfile.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> * Update docs/troubleshooting.md Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com> --------- Signed-off-by:
Matt Williams <m@technovangelist.com> Co-authored-by:
Jeffrey Morgan <jmorganca@gmail.com>
-