- 20 Nov, 2023 2 commits
-
-
Woosuk Kwon authored
-
Simon Mo authored
-
- 19 Nov, 2023 1 commit
-
-
twaka authored
-
- 18 Nov, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 16 Nov, 2023 2 commits
-
-
twaka authored
-
Zhuohan Li authored
TP/quantization/weight loading refactor part 2 - Refactor quantized linear logic and extend quantization support to all models (#1622) Refactor the tensor parallelism, quantization, and weight-loading codes. Summary of the new features enabled by this PR: - **All models** are able to be quantized with AWQ and SqueezeLLM, and [soon GPTQ](https://github.com/vllm-project/vllm/pull/1580). - Model loading code became much simpler. - Support model parallelism for all MQA/GQA models when the number of key/value heads is smaller than the tensor parallel size.
-
- 12 Oct, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 16 Sep, 2023 1 commit
-
-
Woosuk Kwon authored
Co-authored-by:
Robert Irvine <robert@seamlessml.com> Co-authored-by:
root <rirv938@gmail.com> Co-authored-by:
Casper <casperbh.96@gmail.com> Co-authored-by:
julian-q <julianhquevedo@gmail.com>
-
- 13 Sep, 2023 1 commit
-
-
Jasmond L authored
Co-authored-by:
Jasmond Loh <Jasmond.Loh@hotmail.com> Co-authored-by:
Zhuohan Li <zhuohan123@gmail.com>
-
- 07 Sep, 2023 1 commit
-
-
Zhuohan Li authored
-
- 30 Aug, 2023 1 commit
-
-
JFDuan authored
-
- 17 Aug, 2023 1 commit
-
-
Xinyu Yang authored
-
- 08 Jul, 2023 1 commit
-
-
Fazlul Shahriar authored
-
- 03 Jul, 2023 1 commit
-
-
Zhuohan Li authored
-
- 30 Jun, 2023 1 commit
-
-
Zhuohan Li authored
-
- 17 Jun, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 15 May, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 09 May, 2023 1 commit
-
-
Woosuk Kwon authored
-
- 03 May, 2023 2 commits
-
-
Woosuk Kwon authored
-
Zhuohan Li authored
-
- 29 Mar, 2023 1 commit
-
-
Zhuohan Li authored
-
- 21 Mar, 2023 1 commit
-
-
Zhuohan Li authored
-
- 12 Mar, 2023 1 commit
-
-
Woosuk Kwon authored
-