Unverified Commit abf44cce authored by Casper's avatar Casper Committed by GitHub
Browse files

Bump to v0.1.6 (#139)

parent 2bfa234f
...@@ -19,7 +19,7 @@ ...@@ -19,7 +19,7 @@
AutoAWQ is an easy-to-use package for 4-bit quantized models. AutoAWQ speeds up models by 2x while reducing memory requirements by 3x compared to FP16. AutoAWQ implements the Activation-aware Weight Quantization (AWQ) algorithm for quantizing LLMs. AutoAWQ was created and improved upon from the [original work](https://github.com/mit-han-lab/llm-awq) from MIT. AutoAWQ is an easy-to-use package for 4-bit quantized models. AutoAWQ speeds up models by 2x while reducing memory requirements by 3x compared to FP16. AutoAWQ implements the Activation-aware Weight Quantization (AWQ) algorithm for quantizing LLMs. AutoAWQ was created and improved upon from the [original work](https://github.com/mit-han-lab/llm-awq) from MIT.
*Latest News* 🔥 *Latest News* 🔥
- [2023/11] AutoAWQ has been merged into 🤗 transformers. Example found in: [examples/basic_transformers](examples/basic_transformers.py). - [2023/11] AutoAWQ has been merged into 🤗 transformers. Now includes CUDA 12.1 wheels.
- [2023/10] Mistral (Fused Modules), Bigcode, Turing support, Memory Bug Fix (Saves 2GB VRAM) - [2023/10] Mistral (Fused Modules), Bigcode, Turing support, Memory Bug Fix (Saves 2GB VRAM)
- [2023/09] 1.6x-2.5x speed boost on fused models (now including MPT and Falcon). - [2023/09] 1.6x-2.5x speed boost on fused models (now including MPT and Falcon).
- [2023/09] Multi-GPU support, bug fixes, and better benchmark scripts available - [2023/09] Multi-GPU support, bug fixes, and better benchmark scripts available
......
__version__ = "0.1.5" __version__ = "0.1.6"
from awq.models.auto import AutoAWQForCausalLM from awq.models.auto import AutoAWQForCausalLM
\ No newline at end of file
...@@ -14,7 +14,7 @@ except Exception as ex: ...@@ -14,7 +14,7 @@ except Exception as ex:
raise RuntimeError("Your system must have an Nvidia GPU for installing AutoAWQ") raise RuntimeError("Your system must have an Nvidia GPU for installing AutoAWQ")
common_setup_kwargs = { common_setup_kwargs = {
"version": f"0.1.5+cu{CUDA_VERSION}", "version": f"0.1.6+cu{CUDA_VERSION}",
"name": "autoawq", "name": "autoawq",
"author": "Casper Hansen", "author": "Casper Hansen",
"license": "MIT", "license": "MIT",
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment