@@ -18,7 +18,7 @@ Easy, fast, and cheap LLM serving for everyone
*Latest News* 🔥
-[2023/06] We officially released vLLM! vLLM has powered [LMSYS Vicuna and Chatbot Arena](https://chat.lmsys.org) since mid April. Check out our [blog post]().
-[2023/06] We officially released vLLM! vLLM has powered [LMSYS Vicuna and Chatbot Arena](https://chat.lmsys.org) since mid April. Check out our [blog post](https://vllm.ai).
---
...
...
@@ -62,7 +62,7 @@ Visit our [documentation](https://vllm.readthedocs.io/en/latest/) to get started
## Performance
vLLM outperforms HuggingFace Transformers (HF) by up to 24x and Text Generation Inference (TGI) by up to 3.5x, in terms of throughput.
For details, check out our [blog post]().
For details, check out our [blog post](https://vllm.ai).
<palign="center">
<picture>
...
...
@@ -70,7 +70,7 @@ For details, check out our [blog post]().