Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,15 @@ license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/ma
|
|
5 |
---
|
6 |
GGUF of [Replete-AI Llama 3 11.5B Instruct V2](https://huggingface.co/Replete-AI/Llama-3-11.5B-Instruct-v2)
|
7 |
|
8 |
-
Quantized with llama.cpp commit <s>[b2710](https://github.com/ggerganov/llama.cpp/releases/tag/b2710)</s> [b2780](https://github.com/ggerganov/llama.cpp/releases/tag/b2780)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
|
10 |
Original model description below<hr>
|
11 |
|
|
|
5 |
---
|
6 |
GGUF of [Replete-AI Llama 3 11.5B Instruct V2](https://huggingface.co/Replete-AI/Llama-3-11.5B-Instruct-v2)
|
7 |
|
8 |
+
Quantized with llama.cpp commit <s>[b2710](https://github.com/ggerganov/llama.cpp/releases/tag/b2710)</s> <s>[b2780](https://github.com/ggerganov/llama.cpp/releases/tag/b2780)</s> [b2876](https://github.com/ggerganov/llama.cpp/releases/tag/b2876), verified no warnings in llama.cpp
|
9 |
+
|
10 |
+
Simple PPL comparison<br>
|
11 |
+
<code>
|
12 |
+
<i>perplexity.exe -[MODEL] -f wiki.test.raw -b 512 -ngl 99</i>
|
13 |
+
|
14 |
+
SFR-Iterative-DPO-LLaMA-3-11.5B-R-Q6_K.gguf = Final estimate: <b>PPL = 8.4438 +/- 0.06271</b><br>
|
15 |
+
Meta-Llama-3-8B-Instruct-Q6_K = Final estimate: <b>PPL = 8.4727 +/- 0.06308</b>
|
16 |
+
</code>
|
17 |
|
18 |
Original model description below<hr>
|
19 |
|