Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ This repo contains AWQ model files for [PMC_LLaMA_13B](https://huggingface.co/ax
|
|
13 |
|
14 |
### About AWQ
|
15 |
|
16 |
-
[AWQ](https://arxiv.org/abs/2306.00978)
|
17 |
|
18 |
Example of usage with vLLM library:
|
19 |
|
|
|
13 |
|
14 |
### About AWQ
|
15 |
|
16 |
+
[Activation-aware Weight Quantization (AWQ)](https://arxiv.org/abs/2306.00978) selectively preserves a subset of crucial weights for LLM performance instead of quantizing all weights in a model. This targeted approach minimizes quantization loss, allowing models to operate in 4-bit precision without compromising performance.
|
17 |
|
18 |
Example of usage with vLLM library:
|
19 |
|