- 27 Feb, 2024 1 commit
-
-
Dylan Hawk authored
-
- 26 Feb, 2024 1 commit
-
-
Jared Moore authored
-
- 17 Feb, 2024 1 commit
-
-
jvmncs authored
how to serve the loras (mimicking the [multilora inference example](https://github.com/vllm-project/vllm/blob/main/examples/multilora_inference.py)): ```terminal $ export LORA_PATH=~/.cache/huggingface/hub/models--yard1--llama-2-7b-sql-lora-test/ $ python -m vllm.entrypoints.api_server \ --model meta-llama/Llama-2-7b-hf \ --enable-lora \ --lora-modules sql-lora=$LORA_PATH sql-lora2=$LORA_PATH ``` the above server will list 3 separate values if the user queries `/models`: one for the base served model, and one each for the specified lora modules. in this case sql-lora and sql-lora2 point to the same underlying lora, but this need not be the case. lora config values take the same values they do in EngineArgs no work has been done here to scope client permissions to specific models
-
- 25 Jan, 2024 1 commit
-
-
Simon Mo authored
-
- 19 Jan, 2024 1 commit
-
-
Simon Mo authored
-
- 17 Jan, 2024 1 commit
-
-
FlorianJoncour authored
-