File size: 1,793 Bytes
39643c4 49c886c 39643c4 49c886c 1a6840a 6ad5f3d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
---
base_model:
- Pyroserenus/Orthrus-12b-v0.8
- nbeerbower/mistral-nemo-gutenberg-12B-v2
- mergekit-community/Deutscher-Pantheon-12B
- IntervitensInc/Mistral-Nemo-Base-2407-chatml
library_name: transformers
tags:
- mergekit
- merge
license: cc-by-nc-4.0
---
# merge
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [IntervitensInc/Mistral-Nemo-Base-2407-chatml](https://huggingface.co/IntervitensInc/Mistral-Nemo-Base-2407-chatml) as a base.
### Models Merged
The following models were included in the merge:
* [Pyroserenus/Orthrus-12b-v0.8](https://huggingface.co/Pyroserenus/Orthrus-12b-v0.8)
* [nbeerbower/mistral-nemo-gutenberg-12B-v2](https://huggingface.co/nbeerbower/mistral-nemo-gutenberg-12B-v2)
* [mergekit-community/Deutscher-Pantheon-12B](https://huggingface.co/mergekit-community/Deutscher-Pantheon-12B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: mergekit-community/Deutscher-Pantheon-12B
parameters:
weight: 0.3
density: 0.5
- model: nbeerbower/mistral-nemo-gutenberg-12B-v2
parameters:
weight: 0.3
density: 0.5
- model: Pyroserenus/Orthrus-12b-v0.8
parameters:
weight: 0.6
density: 0.5
merge_method: dare_ties
base_model: IntervitensInc/Mistral-Nemo-Base-2407-chatml
dtype: bfloat16
name: Ultra-Instruct-12B
```
Not responsible for what you do with it. Use with caution. WARNING: UNCENSORED. SMART.
Use ChatML. Note: It appears it has trouble stopping. If you value extremly long replies, this might be the model for you. |