Update README.md
Browse files
README.md
CHANGED
@@ -8,9 +8,11 @@ But it has a little thing which distinguishes it :
|
|
8 |
|
9 |
It uses Gryphe's MergeMonster as a tool to trim out the GPTisms, Yisms, and Llamaisms, and give a more natural output.
|
10 |
|
11 |
-
|
|
|
|
|
12 |
|
13 |
-
That make me wonder about the future, when we'll get Miqu models properly finetuned with the best datatsets AND with the Mistralisms trimmed out.
|
14 |
|
15 |
---
|
16 |
|
@@ -20,4 +22,8 @@ Q8_0, Q5_K_S, Q4_K_M, Q4_K_S, Q3_K_M, Q2.
|
|
20 |
|
21 |
To come in the week :
|
22 |
|
23 |
-
IQ3_XXS, IQ2_XS
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
It uses Gryphe's MergeMonster as a tool to trim out the GPTisms, Yisms, and Llamaisms, and give a more natural output.
|
10 |
|
11 |
+
The clearing of any problematic gptism, llamaism, or yiism which was specified to MergeMonster is noticeable
|
12 |
+
And it's like the model is freed of these sequences which represent some form of "EOS chains of tokens" in many models, this in the sense that they conclude many outputs, this ofc in an unwanted way
|
13 |
+
It's quite a step in the right direction which should become the standard practice.
|
14 |
|
15 |
+
That make me wonder about the future, when we'll get Miqu 70b models properly finetuned with the best datatsets AND with the Mistralisms trimmed out as well.
|
16 |
|
17 |
---
|
18 |
|
|
|
22 |
|
23 |
To come in the week :
|
24 |
|
25 |
+
IQ3_XXS, IQ2_XS
|
26 |
+
|
27 |
+
---
|
28 |
+
|
29 |
+
The merge parameters and logs are in the repo : https://huggingface.co/TeeZee/Kyllene-34B-v1.1/tree/main
|