- 22 Dec, 2023 4 commits
-
-
Daniel Hiltgen authored
Add cgo implementation for llama.cpp
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
The default thread count logic was broken and resulted in 2x the number of threads as it should on a hyperthreading CPU resulting in thrashing and poor performance.
-
- 21 Dec, 2023 3 commits
-
-
Bruce MacDonald authored
-
Daniel Hiltgen authored
The windows native setup still needs some more work, but this gets it building again and if you set the PATH properly, you can run the resulting exe on a cuda system.
-
Patrick Devine authored
-
- 20 Dec, 2023 2 commits
-
-
Daniel Hiltgen authored
This switches the default llama.cpp to be CPU based, and builds the GPU variants as dynamically loaded libraries which we can select at runtime. This also bumps the ROCm library to version 6 given 5.7 builds don't work on the latest ROCm library that just shipped.
-
Jeffrey Morgan authored
-
- 19 Dec, 2023 23 commits
-
-
Daniel Hiltgen authored
-
Michael Yang authored
fix(test): use real version string for comparison
-
Michael Yang authored
-
Daniel Hiltgen authored
-
Erick Ghaumez authored
* Add Langchain Dart * Update README.md --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This allows the CPU only builds to work on systems with Radeon cards
-
Daniel Hiltgen authored
If someone checks out the ollama repo and doesn't install the CUDA library, this will ensure they can build a CPU only version
-
Daniel Hiltgen authored
A simple test case that verifies llava:7b can read text in an image
-
Daniel Hiltgen authored
-
Daniel Hiltgen authored
This changes the container-based linux build to use an older Ubuntu distro to improve our compatibility matrix for older user machines
-
Daniel Hiltgen authored
-
65a authored
The build tags rocm or cuda must be specified to both go generate and go build. ROCm builds should have both ROCM_PATH set (and the ROCM SDK present) as well as CLBlast installed (for GGML) and CLBlast_DIR set in the environment to the CLBlast cmake directory (likely /usr/lib/cmake/CLBlast). Build tags are also used to switch VRAM detection between cuda and rocm implementations, using added "accelerator_foo.go" files which contain architecture specific functions and variables. accelerator_none is used when no tags are set, and a helper function addRunner will ignore it if it is the chosen accelerator. Fix go generate commands, thanks @deadmeu for testing.
-
Daniel Hiltgen authored
Run the server.cpp directly inside the Go runtime via cgo while retaining the LLM Go abstractions.
-
Bruce MacDonald authored
-
Bruce MacDonald authored
- remove ggml runner - automatically pull gguf models when ggml detected - tell users to update to gguf in the case automatic pull fails Co-Authored-By:Jeffrey Morgan <jmorganca@gmail.com>
-
Matt Williams authored
Added cmdh to community section in README
-
Matt Williams authored
Lets get rid of these old modelfile examples
-
pgibler authored
-
Matt Williams authored
Signed-off-by:Matt Williams <m@technovangelist.com>
-
- 18 Dec, 2023 5 commits
-
-
Jeffrey Morgan authored
-
Jeffrey Morgan authored
-
Bruce MacDonald authored
-
Patrick Devine authored
-
Jeffrey Morgan authored
-
- 15 Dec, 2023 3 commits
-
-
Patrick Devine authored
-
Augustinas Malinauskas authored
* feat(docs): README with Enchanted iOS app * Update README.md --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-
Ian Purton authored
* Added Bionic GPT as a front end. * Update README.md --------- Co-authored-by:Bruce MacDonald <brucewmacdonald@gmail.com>
-