@@ -18,7 +18,7 @@ Easy, fast, and cheap LLM serving for everyone
...
@@ -18,7 +18,7 @@ Easy, fast, and cheap LLM serving for everyone
*Latest News* 🔥
*Latest News* 🔥
-[2023/06] We officially released vLLM! vLLM has powered [LMSYS Vicuna and Chatbot Arena](https://chat.lmsys.org) since mid April. Check out our [blog post]().
-[2023/06] We officially released vLLM! vLLM has powered [LMSYS Vicuna and Chatbot Arena](https://chat.lmsys.org) since mid April. Check out our [blog post](https://vllm.ai).
---
---
...
@@ -62,7 +62,7 @@ Visit our [documentation](https://vllm.readthedocs.io/en/latest/) to get started
...
@@ -62,7 +62,7 @@ Visit our [documentation](https://vllm.readthedocs.io/en/latest/) to get started
## Performance
## Performance
vLLM outperforms HuggingFace Transformers (HF) by up to 24x and Text Generation Inference (TGI) by up to 3.5x, in terms of throughput.
vLLM outperforms HuggingFace Transformers (HF) by up to 24x and Text Generation Inference (TGI) by up to 3.5x, in terms of throughput.
For details, check out our [blog post]().
For details, check out our [blog post](https://vllm.ai).
<palign="center">
<palign="center">
<picture>
<picture>
...
@@ -70,7 +70,7 @@ For details, check out our [blog post]().
...
@@ -70,7 +70,7 @@ For details, check out our [blog post]().