|
---
|
|
base_model: []
|
|
library_name: transformers
|
|
tags:
|
|
- mergekit
|
|
- merge
|
|
|
|
---
|
|
# Stellar Odyssey 12b v0.0
|
|
*The sparkling courage I longed for, what I got is small... My tears are surely the prism of tomorrow... Say "Hello!" to the ideal future, let's go see them~*
|
|
|
|
Listen to the song on youtube: https://www.youtube.com/watch?v=v3I6EVlyPx4
|
|
|
|
One off merge for a friend, though it came out rather good, I like it, so try it?
|
|
|
|
mistralai/Mistral-Nemo-Base-2407
|
|
inflatebot/MN-12b-Mag-Mell-R1
|
|
nbeerbower/Mistral-Nemo-Prism-12B-v5
|
|
|
|
License for this model Apache 2.0
|
|
|
|
|
|
Format: Mistral Tekken or ChatML
|
|
|
|
Thank you to AuriAetherwiing for helping me merge the models and for providing compute (A40).
|
|
|
|
Details
|
|
|
|
|
|
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
|
## Merge Details
|
|
### Merge Method
|
|
|
|
This model was merged using the della_linear merge method using C:\Users\lg911\Downloads\Mergekit-Fixed\mergekit\mistralai_Mistral-Nemo-Base-2407 as a base.
|
|
|
|
### Models Merged
|
|
|
|
Models Merged
|
|
The following models were included in the merge:
|
|
|
|
/inflatebot_MN-12B-Mag-Mell-R1
|
|
/nbeerbower_Mistral-Nemo-Prism-12B-v5
|
|
|
|
#### Configuration
|
|
The following YAML configuration was used to produce this model:
|
|
|
|
models:
|
|
- model: /inflatebot_MN-12B-Mag-Mell-R1
|
|
parameters:
|
|
weight: 0.3
|
|
density: 0.5
|
|
- model: /nbeerbower_Mistral-Nemo-Prism-12B-v5
|
|
parameters:
|
|
weight: 0.4
|
|
density: 0.75
|
|
base_model: /mistralai_Mistral-Nemo-Base-2407
|
|
parameters:
|
|
epsilon: 0.05
|
|
normalize: true
|
|
lambda: 1
|
|
merge_method: ties
|
|
dtype: bfloat16
|
|
|