Quants with iMatrix for : https://huggingface.co/TeeZee/Kyllene-34B-v1.1
TeeZee's Kyllene model is one of the best Yi_34b merge around with those of BruceTheMoose.
But it has a little thing which distinguishes it :
It uses Gryphe's MergeMonster as a tool to trim out the GPTisms, Yisms, and Llamaisms, and give a more natural output.
The clearing of any problematic gptism, llamaism, or yiism which was specified to MergeMonster is noticeable And it's like the model is freed of these sequences which represent some form of "EOS chains of tokens" in many models, this in the sense that they conclude many outputs, this ofc in an unwanted way It's quite a step in the right direction which should become the standard practice.
That make me wonder about the future, when we'll get Miqu 70b models properly finetuned with the best datatsets AND with the Mistralisms trimmed out as well.
Available quants :
Q8_0, Q5_K_S, Q4_K_M, Q4_K_S, Q3_K_M, Q2_K
To come in the week :
IQ3_XXS, Q2_K_S, IQ2_XS, IQ2_XXS
The merge parameters and logs are in the repo : https://huggingface.co/TeeZee/Kyllene-34B-v1.1/tree/main