File size: 2,042 Bytes
db6a0be 3a7d7a5 3e3bd97 f2a12fa 3e3bd97 230978a 3e3bd97 6b0b6c6 c1c398d 7f0471c 981c056 db6a0be |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
---
base_model:
- inflatebot/thorn-0.5
- inflatebot/thorn-0.35
- inflatebot/thorn-0.55
- inflatebot/thorn-0.45
library_name: transformers
tags:
- mergekit
- merge
---
# merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
![Made with NovelAI](https://huggingface.co/inflatebot/guns-and-roses-r1/resolve/main/guns%20and%20roses.png)
`Quickest draw in the West.`
### NOTE: If you are getting phrase repetition or nonsense outputs with SillyTavern, make sure that "Include names" is disabled under Advanced Formatting. Nemo models tend to exhibit these issues when this is enabled.
A re-application of the Helium-3 process to Mistral Nemo analogues. Experimental (as you can tell by it having a revision number, I'll be playing with this more in coming time.)
Based ultimately on [Magnum-12B-V2](https://huggingface.co/anthracite-org/magnum-12b-v2) and [MN-12B-Rosier-v1](https://huggingface.co/Fizzarolli/MN-12b-Rosier-v1).
Quants available from [Reiterate3680](https://huggingface.co/Reiterate3680/guns-and-roses-r1-GGUF/tree/main)
Special thanks to Fizz and Toasty Pigeon.
Use **ChatML** formatting, as Rosier was trained from base (so no instruct format) and Magnum V2 was trained on ChatML!
### Merge Method
This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [inflatebot/thorn-0.35](https://huggingface.co/inflatebot/thorn-0.35) as a base.
### Models Merged
The following models were included in the merge:
* [inflatebot/thorn-0.5](https://huggingface.co/inflatebot/thorn-0.5)
* [inflatebot/thorn-0.55](https://huggingface.co/inflatebot/thorn-0.55)
* [inflatebot/thorn-0.45](https://huggingface.co/inflatebot/thorn-0.45)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: inflatebot/thorn-0.5
- model: inflatebot/thorn-0.45
- model: inflatebot/thorn-0.55
merge_method: model_stock
base_model: inflatebot/thorn-0.35
dtype: bfloat16
```
|