Llamacpp quants
Browse files- .gitattributes +12 -0
- README.md +75 -0
- mistral-orpo-alpha-Q2_K.gguf +3 -0
- mistral-orpo-alpha-Q3_K_L.gguf +3 -0
- mistral-orpo-alpha-Q3_K_M.gguf +3 -0
- mistral-orpo-alpha-Q3_K_S.gguf +3 -0
- mistral-orpo-alpha-Q4_0.gguf +3 -0
- mistral-orpo-alpha-Q4_K_M.gguf +3 -0
- mistral-orpo-alpha-Q4_K_S.gguf +3 -0
- mistral-orpo-alpha-Q5_0.gguf +3 -0
- mistral-orpo-alpha-Q5_K_M.gguf +3 -0
- mistral-orpo-alpha-Q5_K_S.gguf +3 -0
- mistral-orpo-alpha-Q6_K.gguf +3 -0
- mistral-orpo-alpha-Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
mistral-orpo-alpha-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
mistral-orpo-alpha-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
mistral-orpo-alpha-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
mistral-orpo-alpha-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
mistral-orpo-alpha-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
mistral-orpo-alpha-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
mistral-orpo-alpha-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
mistral-orpo-alpha-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
mistral-orpo-alpha-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
45 |
+
mistral-orpo-alpha-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
46 |
+
mistral-orpo-alpha-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
47 |
+
mistral-orpo-alpha-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,75 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
license: mit
|
5 |
+
base_model:
|
6 |
+
- mistralai/Mistral-7B-v0.1
|
7 |
+
datasets:
|
8 |
+
- HuggingFaceH4/ultrafeedback_binarized
|
9 |
+
pipeline_tag: text-generation
|
10 |
+
model-index:
|
11 |
+
- name: Mistral-ORPO-⍺
|
12 |
+
results:
|
13 |
+
- task:
|
14 |
+
type: text-generation
|
15 |
+
dataset:
|
16 |
+
name: AlpacaEval 1
|
17 |
+
type: AlpacaEval
|
18 |
+
metrics:
|
19 |
+
- type: AlpacaEval 1.0
|
20 |
+
value: 87.92%
|
21 |
+
name: Win Rate
|
22 |
+
source:
|
23 |
+
url: https://github.com/tatsu-lab/alpaca_eval
|
24 |
+
name: self-reported
|
25 |
+
- task:
|
26 |
+
type: text-generation
|
27 |
+
dataset:
|
28 |
+
name: AlpacaEval 2
|
29 |
+
type: AlpacaEval
|
30 |
+
metrics:
|
31 |
+
- type: AlpacaEval 2.0
|
32 |
+
value: 11.33%
|
33 |
+
name: Win Rate
|
34 |
+
source:
|
35 |
+
url: https://github.com/tatsu-lab/alpaca_eval
|
36 |
+
name: self-reported
|
37 |
+
- task:
|
38 |
+
type: text-generation
|
39 |
+
dataset:
|
40 |
+
name: MT-Bench
|
41 |
+
type: MT-Bench
|
42 |
+
metrics:
|
43 |
+
- type: MT-Bench
|
44 |
+
value: 7.23
|
45 |
+
name: Score
|
46 |
+
source:
|
47 |
+
url: https://github.com/lm-sys/FastChat/blob/main/fastchat/llm_judge/
|
48 |
+
name: self-reported
|
49 |
+
quantized_by: bartowski
|
50 |
+
---
|
51 |
+
|
52 |
+
## Llamacpp Quantizations of mistral-orpo-alpha
|
53 |
+
|
54 |
+
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b2440">b2440</a> for quantization.
|
55 |
+
|
56 |
+
Original model: https://huggingface.co/kaist-ai/mistral-orpo-alpha
|
57 |
+
|
58 |
+
Download a file (not the whole branch) from below:
|
59 |
+
|
60 |
+
| Filename | Quant type | File Size | Description |
|
61 |
+
| -------- | ---------- | --------- | ----------- |
|
62 |
+
| [mistral-orpo-alpha-Q8_0.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q8_0.gguf) | Q8_0 | 7.69GB | Extremely high quality, generally unneeded but max available quant. |
|
63 |
+
| [mistral-orpo-alpha-Q6_K.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q6_K.gguf) | Q6_K | 5.94GB | Very high quality, near perfect, *recommended*. |
|
64 |
+
| [mistral-orpo-alpha-Q5_K_M.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q5_K_M.gguf) | Q5_K_M | 5.13GB | High quality, very usable. |
|
65 |
+
| [mistral-orpo-alpha-Q5_K_S.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q5_K_S.gguf) | Q5_K_S | 4.99GB | High quality, very usable. |
|
66 |
+
| [mistral-orpo-alpha-Q5_0.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q5_0.gguf) | Q5_0 | 4.99GB | High quality, older format, generally not recommended. |
|
67 |
+
| [mistral-orpo-alpha-Q4_K_M.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q4_K_M.gguf) | Q4_K_M | 4.36GB | Good quality, similar to 4.25 bpw. |
|
68 |
+
| [mistral-orpo-alpha-Q4_K_S.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q4_K_S.gguf) | Q4_K_S | 4.14GB | Slightly lower quality with small space savings. |
|
69 |
+
| [mistral-orpo-alpha-Q4_0.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q4_0.gguf) | Q4_0 | 4.10GB | Decent quality, older format, generally not recommended. |
|
70 |
+
| [mistral-orpo-alpha-Q3_K_L.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q3_K_L.gguf) | Q3_K_L | 3.82GB | Lower quality but usable, good for low RAM availability. |
|
71 |
+
| [mistral-orpo-alpha-Q3_K_M.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q3_K_M.gguf) | Q3_K_M | 3.51GB | Even lower quality. |
|
72 |
+
| [mistral-orpo-alpha-Q3_K_S.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q3_K_S.gguf) | Q3_K_S | 3.16GB | Low quality, not recommended. |
|
73 |
+
| [mistral-orpo-alpha-Q2_K.gguf](https://huggingface.co/bartowski/mistral-orpo-alpha-GGUF/blob/main/mistral-orpo-alpha-Q2_K.gguf) | Q2_K | 2.71GB | Extremely low quality, *not* recommended.
|
74 |
+
|
75 |
+
Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
|
mistral-orpo-alpha-Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a70b8513a62f0bb551c4b70a41e29cddaa530da8da97159a885c62a4530840e7
|
3 |
+
size 2719242400
|
mistral-orpo-alpha-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:502553535511669615c716b4328a5422db39a7fbcb3ed243eca3e08f2fa73dae
|
3 |
+
size 3822024864
|
mistral-orpo-alpha-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:412fc2b8634505e9957455ac6c12a30dc3a7a3dca91626cb0f23a739d1e17ecc
|
3 |
+
size 3518986400
|
mistral-orpo-alpha-Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46e6a1bdad8fc42a933c9504861bfc34799f4db44f10fa5da4592a7a06ef4665
|
3 |
+
size 3164567712
|
mistral-orpo-alpha-Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:010398f4d0dab68daec40ceebb8ab8f2c4eed47a98a1b75a82fbd8f06ec8d680
|
3 |
+
size 4108916896
|
mistral-orpo-alpha-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce97160c6b334fe0222a27d902527ef1fa4328735160f2df157324709d00774f
|
3 |
+
size 4368439456
|
mistral-orpo-alpha-Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:168f6c12e2391b9239ad4a908b2df769520a74f5bcf30d8d76d0d34d754b3b97
|
3 |
+
size 4140374176
|
mistral-orpo-alpha-Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d098ac2810ee9ee0b046242fe8437546a2be6d9265b7258008ee4c2208fb524b
|
3 |
+
size 4997716128
|
mistral-orpo-alpha-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15372c2d168246199783ee930b8c2da7a354c50c3b77895d1e6b85645d59a6db
|
3 |
+
size 5131409568
|
mistral-orpo-alpha-Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:842b6e41cfd873d73d6848306a327fcf06bc6c3bd4e0f4733b5a08667bcac93e
|
3 |
+
size 4997716128
|
mistral-orpo-alpha-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26bfecab9f5cd5df9edc71e8128023fabc301ae57f66e5d6894a17dc3c79af87
|
3 |
+
size 5942065312
|
mistral-orpo-alpha-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93d558bc25d047b3db58213f1bdfb9ff9571a6908bbfda96004e9c5b202ddaee
|
3 |
+
size 7695857824
|