File size: 9,164 Bytes
eda077e 3bd680f eda077e 28fb952 eda077e 6ce03aa eda077e 6ce03aa 28fb952 eda077e 28fb952 eda077e 28fb952 eda077e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 |
---
base_model:
- Sao10K/L3-8B-Niitama-v1
- Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- ArliAI/ArliAI-Llama-3-8B-Formax-v1.0
- nothingiisreal/L3-8B-Celeste-V1.2
- ResplendentAI/Nymph_8B
- TheDrummer/Llama-3SOME-8B-v2
- nothingiisreal/L3-8B-Instruct-Abliterated-DWP
library_name: transformers
tags:
- mergekit
- merge
- not-for-all-audiences
---
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/667eea5cdebd46a5ec4dcc3d/hFHS4nKKh3sbSUq6MlPaK.jpeg)
This did not go in the direction I wanted it to. Why is it always my best models are the ones that happen out of sheer coincidence or stupidity. To start, this model doesn't
have Crepus Horizon in it despite its name. As of writing this, v0.1 Crepus Horzion is way too broken to merge with and I've taken that repo down until an update is out. This
model actually works better without it so, a plus. Still keeping the name though, sounds cool lmao.
Tangent aside, Antheros was part elaborate experiment, part 'throw shit at the wall as see what sticks'. I was aiming for a more narritive heavy RP model with some moist and spice thrown
in, ended up with that and solid intruct following. Took a week and a half, but it's here. Any feedback would be much appreciated.
### Quants
[OG Q8 GGUF](https://huggingface.co/kromeurus/L3-Horizon-Anteros-v0.1-13B-Q8-GGUF) by me.
[GGUFs](https://huggingface.co/backyardai/L3-Horizon-Anteros-v0.1-13B-GGUF) by [BackyardAI](https://huggingface.co/backyardai)
### Details & Recommended Settings
(Still testing; subject to change)
Thrives at more story heavy and narritive RP yet still excels with the basics. Leans vaguely sensual and smooth-tongued with it's description while keeping up human dialogue
and actions. Fully uncensored and a tad dramatic, so have fun with those eRP scenes.
Clings onto whatever instructs you give it and fucking sprints with it (thanks Formax). This model's generation quality heavily depends on the character card given; a shit
card will give you shit outputs and vise versa. Be careful with your choice of words and the examples you give it.
Has a certain tendency to speak for the {user}, but that's easily negated with a few instructs.
Rec. Settings:
```
Template: Model Default
Temperature: 1.3
Min P: 0.1
Repeat Penalty: 1.1
Repeat Penalty Tokens: 256
```
### Models Merged & Merge Theory
The following models were included in the merge:
* [ResplendentAI/Nymph_8B](https://huggingface.co/ResplendentAI/Nymph_8B)
* [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)
* [nothingiisreal/L3-8B-Instruct-Abliterated-DWP](https://huggingface.co/nothingiisreal/L3-8B-Instruct-Abliterated-DWP)
* [Sao10K/L3-8B-Niitama-v1](https://huggingface.co/Sao10K/L3-8B-Niitama-v1)
* [Nitral-AI/Hathor_Tahsin-L3-8B-v0.85](https://huggingface.co/Nitral-AI/Hathor_Tahsin-L3-8B-v0.85)
* [ArliAI/ArliAI-Llama-3-8B-Formax-v1.0](https://huggingface.co/ArliAI/ArliAI-Llama-3-8B-Formax-v1.0)
* [nothingiisreal/L3-8B-Celeste-V1.2](https://huggingface.co/nothingiisreal/L3-8B-Celeste-V1.2)
It looks like a lot, I swear it's not that bad.
General idea was for a more story like RP merge so I picked models that were either for basic RP and story wriing. Originally it was going to only be six models split into two
groups to min max their attributes, then someone on the BackyardAI discord asked if someone could fit in Formax into an RP model and I said 'bet'. Small problem, Formax is a
4k context model while the rest are 8k. As excpected, some frakenmerges straight up failed to quant. But, I managed to finagle it in.
The first merge (anteros.b) is the more story-writing heavy base of Anteros, also the part with majority of the *moist and spice* in it. Given that the models in it are all NSFW and one
trained off of r/DirtyWritingPrompts, who's suprised? Since Instruct DWP is a story writing model, the top end had to be sole RP models to balence it out as well as right at the
end. But, I wanted to keep that 'human' writing style so I capped the end still with Instruct DWP. What came out was a merge that rambled a ton but verbose and narritive driven.
Second merge (anteros.c) is the RP forward merge with Formax at the start. Still keeping that 'human' style with by capping the end with Celeste v1.2 while stuffing the mids with Niitama
and Tahsin; two very competent models in their own right. Took a page out of [@matchaaaaa](https://huggingface.co/matchaaaaa)'s recent models with the splice method where
you take a slice out of two models of the same layers then merge it. I took it a step further, splicing where there was major (4+) layer overlap, streamlining the merge overall.
The resulting merge is consise and almost snippy, but suprisingly consise and coherent though it doesn't pick up nuance well.
As for how I managed to get Formax in, I did some testing and found out that you could get away with merging roughly the first quarter of the layers without much error.
It's the last few layers that cause trouble; trying to put the end layers at the end of the merge will make the model unquantable. But, I don't know how much that effects the
overall context length for the final model, still need more testing for that.
The final merge was a DELLA merge with both anteros.b and anteros.c with multiple gradients. DELLA is a very new merge method as of this model's release, you can read more about it [here](https://arxiv.org/abs/2406.11617).
Wanted to keep anteros.b's narritive affinity and nuance while also keeping anteros.c's coherence; so, I did a high weight of anteros.c with low density at the start then eased
the weight down until the end, where it went up again. Maintained an averege density around ~0.4 and epsilon around 0.05.
### Config
```yaml
models:
slices:
- sources:
- layer_range: [0, 4]
model: ResplendentAI/Nymph_8B
- sources:
- layer_range: [3, 9]
model: TheDrummer/Llama-3SOME-8B-v2
parameters:
scale:
- filter: q_proj
value: 0.9
- filter: down_proj
value: 0.9
- sources:
- layer_range: [5, 12]
model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP
parameters:
scale:
- filter: up_proj
value: 0.9
- sources:
- layer_range: [10, 18]
model: ResplendentAI/Nymph_8B
- sources:
- layer_range: [11, 22]
model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP
- sources:
- layer_range: [17, 19]
model: TheDrummer/Llama-3SOME-8B-v2
- sources:
- layer_range: [19, 24]
model: ResplendentAI/Nymph_8B
- sources:
- layer_range: [22, 26]
model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP
- sources:
- layer_range: [27, 29]
model: TheDrummer/Llama-3SOME-8B-v2
parameters:
scale:
- filter: v_proj
value: 0.9
- sources:
- layer_range: [29, 30]
model: ResplendentAI/Nymph_8B
- sources:
- layer_range: [26, 32]
model: nothingiisreal/L3-8B-Instruct-Abliterated-DWP
parameters:
int8_mask: true
merge_method: passthrough
dtype: bfloat16
name: anteros.b
---
models:
slices:
- sources:
- layer_range: [14, 20]
model: nothingiisreal/L3-8B-Celeste-V1.2
parameters:
int8_mask: true
merge_method: passthrough
dtype: bfloat16
name: celeste14-20.sl
---
models:
slices:
- sources:
- layer_range: [14, 20]
model: Sao10K/L3-8B-Niitama-v1
parameters:
int8_mask: true
merge_method: passthrough
dtype: bfloat16
name: niitama14-20.sl
---
models:
- model: celeste14-20.sl
parameters:
weight: [1, 0.75, 0.625, 0.5, 0.375, 0.25, 0]
- model: niitama14-20.sl
parameters:
weight: [0, 0.25, 0.375, 0.5, 0.625, 0.75, 1]
merge_method: dare_linear
base_model: celeste14-20.sl
dtype: bfloat16
name: celeniit14-20.sl
---
models:
slices:
- sources:
- layer_range: [0, 4]
model: Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- sources:
- layer_range: [1, 5]
model: ArliAI/ArliAI-Llama-3-8B-Formax-v1.0
- sources:
- layer_range: [4, 8]
model: Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- sources:
- layer_range: [5, 9]
model: ArliAI/ArliAI-Llama-3-8B-Formax-v1.0
- sources:
- layer_range: [8, 10]
model: Sao10K/L3-8B-Niitama-v1
- sources:
- layer_range: [6, 14]
model: nothingiisreal/L3-8B-Celeste-V1.2
- sources:
- layer_range: [0, 6]
model: celeniit14-20.sl
- sources:
- layer_range: [20, 23]
model: Sao10K/L3-8B-Niitama-v1
- sources:
- layer_range: [22, 26]
model: Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- sources:
- layer_range: [22, 28]
model: nothingiisreal/L3-8B-Celeste-V1.2
- sources:
- layer_range: [25, 27]
model: Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
- sources:
- layer_range: [28, 30]
model: Sao10K/L3-8B-Niitama-v1
- sources:
- layer_range: [25, 32]
model: nothingiisreal/L3-8B-Celeste-V1.2
parameters:
int8_mask: true
merge_method: passthrough
dtype: bfloat16
name: anteros.c # Himerus Basis.C, available on its own.
----
models:
- model: himerus.c
parameters:
weight: [0.8, 0.32, 0.6]
density: [0.23, 0.55, 0.37]
epsilon: [0.05, 0.06, 0.04]
lambda: 0.07
- model: anteros.b
base_model: anteros.b
parameters:
normalize: false
int8_mask: true
merge_method: della
dtype: bfloat16
``` |