- 24 Nov, 2023 1 commit
-
-
Yanming W authored
-
- 22 Nov, 2023 4 commits
- 21 Nov, 2023 6 commits
-
-
Zhuohan Li authored
-
Woosuk Kwon authored
-
boydfd authored
Co-authored-by:ran_lin <rlin@thoughtworks.com>
-
ljss authored
-
Zhuofan authored
-
陈序 authored
-
- 20 Nov, 2023 5 commits
-
-
Woosuk Kwon authored
-
Zhuohan Li authored
-
Simon Mo authored
-
Wen Sun authored
-
Simon Mo authored
-
- 19 Nov, 2023 6 commits
-
-
Woosuk Kwon authored
-
Woosuk Kwon authored
-
Woosuk Kwon authored
-
ljss authored
-
Woosuk Kwon authored
-
twaka authored
-
- 18 Nov, 2023 3 commits
-
-
liuyhwangyh authored
-
Woosuk Kwon authored
-
Roy authored
-
- 17 Nov, 2023 3 commits
-
-
Zhuofan authored
-
Zhuohan Li authored
-
Zhuohan Li authored
-
- 16 Nov, 2023 8 commits
-
-
Iskren Ivov Chernev authored
-
Zhuohan Li authored
-
maximzubkov authored
-
Simon Mo authored
-
twaka authored
-
Aaron Pham authored
Signed-off-by:Aaron <29749331+aarnphm@users.noreply.github.com>
-
Megha Agarwal authored
-
Zhuohan Li authored
TP/quantization/weight loading refactor part 2 - Refactor quantized linear logic and extend quantization support to all models (#1622) Refactor the tensor parallelism, quantization, and weight-loading codes. Summary of the new features enabled by this PR: - **All models** are able to be quantized with AWQ and SqueezeLLM, and [soon GPTQ](https://github.com/vllm-project/vllm/pull/1580). - Model loading code became much simpler. - Support model parallelism for all MQA/GQA models when the number of key/value heads is smaller than the tensor parallel size.
-
- 14 Nov, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 13 Nov, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 12 Nov, 2023 1 commit
-
-
lirui authored
-
- 11 Nov, 2023 1 commit
-
-
Dominik Schwabe authored
-