Unverified Commit ab127632 authored by Haotian (Ken) Tang's avatar Haotian (Ken) Tang Committed by GitHub
Browse files

[Minor] Update README.md

parent 3fce6906
# AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration [[Paper](https://arxiv.org/abs/2306.00978)]
**Efficient and accurate** low-bit weight quantization (INT3/4) for LLMs, supporting **instruction-tuned** models and **multi-modal** LMs.
**Efficient and accurate** low-bit weight quantization (INT3/4) for LLMs, supporting **instruction-tuned** models and **multi-modal** LMs, democratizing LLMs on **edge devices**.
![overview](figures/overview.png)
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment