File size: 4,879 Bytes
297524a 3c5ebdf 297524a 3c5ebdf 297524a e820cb6 297524a 3eb54e6 e820cb6 14ae6c6 e820cb6 3eb54e6 5620919 3eb54e6 14ae6c6 d7d5e8f 4ee1d6b d7d5e8f 76358a7 d7d5e8f 233e52c d7d5e8f 233e52c d7d5e8f 0d03eac 233e52c 3eb54e6 e820cb6 059662f 5620919 059662f 5620919 e60bc24 3eb54e6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
---
language:
- en
library_name: transformers
license: llama2
quantized_by: mradermacher
---
## About
weighted/imatrix wants of https://huggingface.co/sophosympatheia/Aurora-Nights-103B-v1.0
the weights were calculated using 164k semi-random english-only tokens.
<!-- provided-files -->
## Usage
If you are unsure how to use GGUF files, refer to one of [TheBloke's
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
more details, including on how to concatenate multi-part files.
## Provided Quants
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
| Link | Type | Size/GB | Notes |
|:-----|:-----|--------:|:------|
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ1_S.gguf) | i1-IQ1_S | 22.1 | for the desperate |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 27.7 | |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ2_XS.gguf) | i1-IQ2_XS | 30.8 | |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ2_M.gguf) | i1-IQ2_M | 35.1 | |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q2_K.gguf) | i1-Q2_K | 38.3 | IQ3_XXS probably better |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 40.6 | fast, lower quality |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_XS.gguf) | i1-Q3_K_XS | 42.4 | |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ3_XS.gguf) | i1-IQ3_XS | 42.5 | |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_S.gguf) | i1-Q3_K_S | 44.9 | IQ3_XS probably better |
| [GGUF](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-IQ3_S.gguf) | i1-IQ3_S | 45.0 | fast, beats Q3_K* |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_M.gguf.split-aa) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_M.gguf.split-ab) | i1-Q3_K_M | 50.0 | IQ3_S probably better |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_L.gguf.split-aa) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q3_K_L.gguf.split-ab) | i1-Q3_K_L | 54.5 | IQ3_M probably better |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q4_K_S.gguf.split-aa) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q4_K_S.gguf.split-ab) | i1-Q4_K_S | 59.0 | optimal size/speed/quality |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q4_K_M.gguf.split-aa) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q4_K_M.gguf.split-ab) | i1-Q4_K_M | 62.3 | fast, medium quality |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q5_K_S.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q5_K_S.gguf.part2of2) | i1-Q5_K_S | 71.4 | |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q5_K_M.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q5_K_M.gguf.part2of2) | i1-Q5_K_M | 73.3 | |
| [PART 1](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q6_K.gguf.part1of2) [PART 2](https://huggingface.co/mradermacher/Aurora-Nights-103B-v1.0-i1-GGUF/resolve/main/Aurora-Nights-103B-v1.0.i1-Q6_K.gguf.part2of2) | i1-Q6_K | 85.1 | practically like static Q6_K |
Here is a handy graph by ikawrakow comparing some lower-quality quant
types (lower is better):
![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png)
And here are Artefact2's thoughts on the matter:
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
<!-- end -->
|