|
--- |
|
base_model: [] |
|
library_name: transformers |
|
tags: |
|
- mergekit |
|
- merge |
|
|
|
--- |
|
# Prismatic 12b v0.1 Experimental 11/15 |
|
|
|
## This is a fix for ChatML format, since before it did not have an EOS token |
|
*The sparkling courage I longed for, what I got is small... My tears are surely the prism of tomorrow... Say "Hello!" to the ideal future, let's go see them~* |
|
|
|
Listen to the song on youtube: https://www.youtube.com/watch?v=v3I6EVlyPx4 |
|
|
|
One off merge for a friend, though it came out rather good, I like it, so try it? |
|
|
|
mistralai/Mistral-Nemo-Base-2407 |
|
inflatebot/MN-12b-Mag-Mell-R1 |
|
nbeerbower/Mistral-Nemo-Prism-12B-v5 |
|
|
|
License for this model Apache 2.0 |
|
|
|
|
|
Format: Mistral Tekken or ChatML |
|
|
|
Thank you to AuriAetherwiing for helping me merge the models and for providing compute (A40). |
|
|
|
Details |
|
|
|
|
|
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). |
|
|
|
## Merge Details |
|
### Merge Method |
|
|
|
This model was merged using the ties merge method using mistralai_Mistral-Nemo-Base-2407 as a base. |
|
|
|
### Models Merged |
|
|
|
Models Merged |
|
The following models were included in the merge: |
|
|
|
/inflatebot_MN-12B-Mag-Mell-R1 |
|
/nbeerbower_Mistral-Nemo-Prism-12B-v5 |
|
|
|
#### Configuration |
|
The following YAML configuration was used to produce this model: |
|
|
|
models: |
|
- model: /inflatebot_MN-12B-Mag-Mell-R1 |
|
parameters: |
|
weight: 0.3 |
|
density: 0.5 |
|
- model: /nbeerbower_Mistral-Nemo-Prism-12B-v5 |
|
parameters: |
|
weight: 0.4 |
|
density: 0.75 |
|
base_model: /mistralai_Mistral-Nemo-Base-2407 |
|
parameters: |
|
epsilon: 0.05 |
|
normalize: true |
|
lambda: 1 |
|
merge_method: ties |
|
dtype: bfloat16 |
|
|