Update README.md
Browse files
README.md
CHANGED
@@ -49,6 +49,9 @@ We want the good ending, not the bad one.
|
|
49 |
|
50 |
## Metrics
|
51 |
|
|
|
|
|
|
|
52 |
```sh
|
53 |
perplexity -m lmstudio-community/Meta-Llama-3-8B-Instruct-Q6_K.gguf -b 32 -c 512 -f wiki.test.raw
|
54 |
# Final estimate: PPL = 7.5588 +/- 0.05599
|
@@ -58,7 +61,15 @@ perplexity -m cognitivecomputations/dolphin-2.9-llama3-8b-q5_K_M.gguf -b 32 -c 5
|
|
58 |
# Final estimate: PPL = 9.9277 +/- 0.08261
|
59 |
```
|
60 |
|
61 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
|
63 |
---
|
64 |
license: llama3
|
|
|
49 |
|
50 |
## Metrics
|
51 |
|
52 |
+
|
53 |
+
Meausing the gguf, there is a difference in perplexity
|
54 |
+
|
55 |
```sh
|
56 |
perplexity -m lmstudio-community/Meta-Llama-3-8B-Instruct-Q6_K.gguf -b 32 -c 512 -f wiki.test.raw
|
57 |
# Final estimate: PPL = 7.5588 +/- 0.05599
|
|
|
61 |
# Final estimate: PPL = 9.9277 +/- 0.08261
|
62 |
```
|
63 |
|
64 |
+
Measuring it in the original huggingface format, the increase is much small
|
65 |
+
(as a %, the absolute values are not comparable with the ones above as there are differences in measurement)
|
66 |
+
|
67 |
+
model | perplexity|
|
68 |
+
|-|-|
|
69 |
+
|base |295.462970|
|
70 |
+
|orthogonalized |309.856348|
|
71 |
+
|
72 |
+
So yes this model edit does increase the perplexity :(. But more investigation is needed.
|
73 |
|
74 |
---
|
75 |
license: llama3
|