--- license: other license_name: mistral-ai-research-licence license_link: https://mistral.ai/licenses/MRL-0.1.md base_model: [] library_name: transformers tags: - mergekit - lumikabra-123B --- # lumikabra-123B v0.3
Lumikabra
This is lumikabra. It's based on [Mistral-Large-Instruct-2407 ](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407), merged with Magnum-v2-123B, Luminum-v0.1-123B and Tess-3-Mistral-Large-2-123B. I shamelessly took this idea from [FluffyKaeloky](https://huggingface.co/FluffyKaeloky/Luminum-v0.1-123B). Like him, i always had my troubles with each of the current large mistral based models. Either it gets repetitive, shows too many GPTisms, is too horny or too unhorny. RP and storytelling is always a matter of taste, and i found myself swiping too often for new answers or even fixing them when I missed a little spice or cleverness. Luminum was a great improvement, mixing a lot of desired traits, but I still missed some spice, another sauce. So i took Luminum, added magnum again and also Tess for knowledge and structure. This is the third iteration. For my taste this is the best one, with a little bit more of Tess in the mix. Sadly all three models seem to repeat stuff in the long run. Let's say the character gives you a first kiss... It can happen that two or three scenes later the character tries again to give you your first kiss. ## Quants - [exl2-8.0](https://huggingface.co/schnapper79/lumikabra-123B_v0.3-exl2-8.0bpw) ## Merge Details ### Merge Method This model was merged using [mergekit](https://github.com/cg123/mergekit) with the della_linear merge method using mistralai_Mistral-Large-Instruct-2407 as a base. ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: /workspace/text-generation-webui/models/anthracite-org_magnum-v2-123b parameters: weight: 0.34 density: 0.5 - model: /workspace/text-generation-webui/models/FluffyKaeloky_Luminum-v0.1-123B parameters: weight: 0.24 density: 0.8 - model: /workspace/text-generation-webui/models/migtissera_Tess-3-Mistral-Large-2-123B parameters: weight: 0.34 density: 0.9 merge_method: della_linear base_model: /workspace/text-generation-webui/models/mistralai_Mistral-Large-Instruct-2407 parameters: epsilon: 0.05 lambda: 1 int8_mask: true dtype: bfloat16 ```