---
license: cc-by-nc-4.0
language:
- en
---
![Lyra](https://huggingface.co/Sao10K/MN-12B-Lyra-v4/resolve/main/lyra.png)
# MN-12B-Lyra-v4 - EXL2 8bpw max
This is a 8bpw EXL2 quant of [Sao10K/MN-12B-Lyra-v4](https://huggingface.co/Sao10K/MN-12B-Lyra-v4)
This quant was made using exllamav2-0.2.0 with default dataset. I used a slightly modified quantization script to force use of highest bpw methods for all layers in the model (which is usually "1:8b_128g s4") to ensure max quality.
I also added a small fix in config file to set max default context at 128k as original Mistral-Nemo should have.
I tested this quant shortly in some random RPs (including ones over 8k context) and it seems to work fine.
## Prompt Templates
Uses ChatML or modified mistral format like below. I tested it with ChatML.
### Original readme below
---
Mistral-NeMo-12B-Lyra-v4, a variation of [Lyra-v4a1](https://huggingface.co/Sao10K/MN-12B-Lyra-v4a1), layered over [Lyra-v3](https://huggingface.co/Sao10K/MN-12B-Lyra-v3), which was built on top of [Lyra-v2a2](https://huggingface.co/Sao10K/MN-12B-Lyra-v2a2), which itself was built upon [Lyra-v2a1](https://huggingface.co/Sao10K/MN-12B-Lyra-v2a1).
# Model Versioning
```
[See Previous Models]
|
Lyra-v4a1
|
------------> Lyra-v4 [Seperate RL Step targeting Instruct and Coherency over Base Nemo instead of SFT First, Result is Merged with Lyra-v4a1, fixes most quant-based issues. Somehow.]
```
# This uses ChatML, or any of its variants which were included in previous versions.
```
<|im_start|>system
This is the system prompt.<|im_end|>
<|im_start|>user
Instructions placed here.<|im_end|>
<|im_start|>assistant
The model's response will be here.<|im_end|>
--------------------------------------------------
[INST]system
This is another system prompt.[/INST]
[INST]user
Your instructions placed here.[/INST]
[INST]assistant
The model's response will be here.[/INST]
```
# Recommended Samplers:
```
Temperature: 0.6 - 1 # Make sure min_p is set before Temperature in Sampler Orders
min_p: 0.1 - 0.2 # Crucial for NeMo
```
# Recommended Stopping Strings:
```
<|im_end|>
[/INST]
```
# Notes
\- I think I fixed the extra token stuff some users seem to be facing, while retaining everything else? It's some error alright.
\- If you're using XML tags, you may see weird malformed stopping strings. Just add them to your current list. and move on.
\- Its pretty nice, imo. I've been messing around with it a lot.
\- Make sure the ChatML template is correct, I think there's some issues with the one used in SillyTavern which might cause improper replies?