IHaveNoClueAndIMustPost commited on
Commit
9c8a059
·
verified ·
1 Parent(s): 91a5fcc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -5,7 +5,15 @@ license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/ma
5
  ---
6
  GGUF of [Replete-AI Llama 3 11.5B Instruct V2](https://huggingface.co/Replete-AI/Llama-3-11.5B-Instruct-v2)
7
 
8
- Quantized with llama.cpp commit <s>[b2710](https://github.com/ggerganov/llama.cpp/releases/tag/b2710)</s> [b2780](https://github.com/ggerganov/llama.cpp/releases/tag/b2780)
 
 
 
 
 
 
 
 
9
 
10
  Original model description below<hr>
11
 
 
5
  ---
6
  GGUF of [Replete-AI Llama 3 11.5B Instruct V2](https://huggingface.co/Replete-AI/Llama-3-11.5B-Instruct-v2)
7
 
8
+ Quantized with llama.cpp commit <s>[b2710](https://github.com/ggerganov/llama.cpp/releases/tag/b2710)</s> <s>[b2780](https://github.com/ggerganov/llama.cpp/releases/tag/b2780)</s> [b2876](https://github.com/ggerganov/llama.cpp/releases/tag/b2876), verified no warnings in llama.cpp
9
+
10
+ Simple PPL comparison<br>
11
+ <code>
12
+ <i>perplexity.exe -[MODEL] -f wiki.test.raw -b 512 -ngl 99</i>
13
+
14
+ SFR-Iterative-DPO-LLaMA-3-11.5B-R-Q6_K.gguf = Final estimate: <b>PPL = 8.4438 +/- 0.06271</b><br>
15
+ Meta-Llama-3-8B-Instruct-Q6_K = Final estimate: <b>PPL = 8.4727 +/- 0.06308</b>
16
+ </code>
17
 
18
  Original model description below<hr>
19