--- base_model: - unsloth/gemma-2-9b - crestf411/gemma2-9B-sunfall-v0.5.2 - anthracite-org/magnum-v4-9b - unsloth/gemma-2-9b-it - nbeerbower/Gemma2-Gutenberg-Doppel-9B library_name: transformers tags: - mergekit - merge --- # Gemma2-Eclipse-9B The merge script was taken from [MarinaraSpaghetti/NemoMix-Unleashed-12B](https://huggingface.co/MarinaraSpaghetti/NemoMix-Unleashed-12B). All I did was replace the models (and I couldn't find a replacement for one of the models so I just increased the weight on one) and it seems to work pretty well. ChatML prompt format seems to work fine. Gemma 2 format seems to also work. ChatML seems cause it to incorrectly spew this at the end |im_end|> so Gemma 2 format might work better. This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the della_linear merge method using [unsloth/gemma-2-9b](https://huggingface.co/unsloth/gemma-2-9b) as a base. ### Models Merged The following models were included in the merge: * [crestf411/gemma2-9B-sunfall-v0.5.2](https://huggingface.co/crestf411/gemma2-9B-sunfall-v0.5.2) * [anthracite-org/magnum-v4-9b](https://huggingface.co/anthracite-org/magnum-v4-9b) * [unsloth/gemma-2-9b-it](https://huggingface.co/unsloth/gemma-2-9b-it) * [nbeerbower/Gemma2-Gutenberg-Doppel-9B](https://huggingface.co/nbeerbower/Gemma2-Gutenberg-Doppel-9B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: unsloth/gemma-2-9b-it parameters: weight: 0.22 density: 0.5 - model: nbeerbower/Gemma2-Gutenberg-Doppel-9B parameters: weight: 0.2 density: 0.6 - model: crestf411/gemma2-9B-sunfall-v0.5.2 parameters: weight: 0.25 density: 0.7 - model: anthracite-org/magnum-v4-9b parameters: weight: 0.33 density: 0.8 merge_method: della_linear base_model: unsloth/gemma-2-9b parameters: epsilon: 0.05 lambda: 1 dtype: bfloat16 tokenizer_source: base ```