"server/requirements_cuda.txt" did not exist on "5485c142e87a182f2eee713a6b056ee38bd901f5"
Unverified Commit a2e95308 authored by Hailey Schoelkopf's avatar Hailey Schoelkopf Committed by GitHub
Browse files

ggml -> gguf in readme

parent 35bdecd3
......@@ -141,11 +141,11 @@ python main.py \
--tasks hellaswag
```
GGML quantized models can be loaded by using `llama-cpp-python` server:
GGUF or GGML quantized models can be loaded by using `llama-cpp-python` server:
```bash
python main.py \
--model ggml \
--model gguf \
--model_args base_url=http://localhost:8000 \
--tasks hellaswag
```
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment