Originally these were my personal GGUF-IQ-Imatrix quants of openlynn/Llama-3-Soliloquy-8B-v2.
Read the original model page for details.

Author:
"Soliloquy-L3 is a highly capable roleplaying model designed for immersive, dynamic experiences. Trained on over 250 million tokens of roleplaying data, Soliloquy-L3 has a vast knowledge base, rich literary expression, and support for up to 24k context length. It outperforms existing ~13B models, delivering enhanced roleplaying capabilities."

Note:
Took me a bit to get into it as I've been busy with life things but this model has performed amazingly well so far. Even the formatting is more stable than others when it comes to asterisks. Not perfect, but close.

SillyTavern:
Use the Llama-3 presets (simple) or Virt's amazing roleplay presets here (recommended) with the Simple samplers. If you have questions, please do ask.

Support:
My upload speeds have been cooked and unstable lately.
Realistically I'd need to move to get a better provider.
If you want and you are able to...
You can support my various endeavors here (Ko-fi).
I apologize for disrupting your experience.

image/png

Downloads last month
831
GGUF
Model size
8.03B params
Architecture
llama

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference API
Unable to determine this model's library. Check the docs .

Collection including Lewdiculous/Llama-3-Soliloquy-8B-v2-GGUF-IQ-Imatrix