|
--- |
|
license: cc-by-nc-4.0 |
|
language: |
|
- en |
|
--- |
|
|
|
 |
|
|
|
# MN-12B-Lyra-v4 - EXL2 8bpw max |
|
|
|
This is a 8bpw EXL2 quant of [Sao10K/MN-12B-Lyra-v4](https://huggingface.co/Sao10K/MN-12B-Lyra-v4) |
|
|
|
This quant was made using exllamav2-0.2.0 with default dataset. I used a slightly modified quantization script to force use of highest bpw methods for all layers in the model (which is usually "1:8b_128g s4") to ensure max quality. |
|
|
|
I also added a small fix in config file to set max default context at 128k as original Mistral-Nemo should have. |
|
|
|
I tested this quant shortly in some random RPs (including ones over 8k context) and it seems to work fine. |
|
|
|
## Prompt Templates |
|
|
|
Uses ChatML or modified mistral format like below. I tested it with ChatML. |
|
|
|
### Original readme below |
|
|
|
--- |
|
|
|
Mistral-NeMo-12B-Lyra-v4, a variation of [Lyra-v4a1](https://huggingface.co/Sao10K/MN-12B-Lyra-v4a1), layered over [Lyra-v3](https://huggingface.co/Sao10K/MN-12B-Lyra-v3), which was built on top of [Lyra-v2a2](https://huggingface.co/Sao10K/MN-12B-Lyra-v2a2), which itself was built upon [Lyra-v2a1](https://huggingface.co/Sao10K/MN-12B-Lyra-v2a1). |
|
|
|
# Model Versioning |
|
``` |
|
[See Previous Models] |
|
| |
|
Lyra-v4a1 |
|
| |
|
------------> Lyra-v4 [Seperate RL Step targeting Instruct and Coherency over Base Nemo instead of SFT First, Result is Merged with Lyra-v4a1, fixes most quant-based issues. Somehow.] |
|
``` |
|
|
|
# This uses ChatML, or any of its variants which were included in previous versions. |
|
|
|
``` |
|
|
|
<|im_start|>system |
|
This is the system prompt.<|im_end|> |
|
<|im_start|>user |
|
Instructions placed here.<|im_end|> |
|
<|im_start|>assistant |
|
The model's response will be here.<|im_end|> |
|
-------------------------------------------------- |
|
[INST]system |
|
This is another system prompt.[/INST] |
|
[INST]user |
|
Your instructions placed here.[/INST] |
|
[INST]assistant |
|
The model's response will be here.[/INST] |
|
``` |
|
|
|
# Recommended Samplers: |
|
|
|
``` |
|
Temperature: 0.6 - 1 # Make sure min_p is set before Temperature in Sampler Orders |
|
min_p: 0.1 - 0.2 # Crucial for NeMo |
|
``` |
|
|
|
# Recommended Stopping Strings: |
|
|
|
``` |
|
<|im_end|> |
|
</s> |
|
[/INST] |
|
``` |
|
|
|
# Notes |
|
|
|
\- I think I fixed the extra token stuff some users seem to be facing, while retaining everything else? It's some error alright. |
|
<br>\- If you're using XML tags, you may see weird malformed stopping strings. Just add them to your current list. and move on. |
|
<br>\- Its pretty nice, imo. I've been messing around with it a lot. |
|
<br>\- Make sure the ChatML template is correct, I think there's some issues with the one used in SillyTavern which might cause improper replies? |