Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model:
|
3 |
+
- localfultonextractor/Erosumika-7B
|
4 |
+
- Nitral-AI/Infinitely-Laydiculous-7B
|
5 |
+
- Kunocchini-7b-128k-test
|
6 |
+
- Endevor/EndlessRP-v3-7B
|
7 |
+
- ChaoticNeutrals/BuRP_7B
|
8 |
+
- daybreak-kunoichi-2dpo-7b
|
9 |
+
library_name: transformers
|
10 |
+
tags:
|
11 |
+
- mergekit
|
12 |
+
- merge
|
13 |
+
---
|
14 |
+
|
15 |
+
My first merge of RP models 7B using mergekit, They are just take .r/ trend RP, half is BuRP_7B. not used any, **Dumb** merge but hopfully lucky merge! ^^'
|
16 |
+
<div style="width: auto; margin-left: auto; margin-right: auto">
|
17 |
+
<img src="https://i.imgur.com/d38LuOG.png" alt="Nekochu" style="width: 250%; min-width: 400px; display: block; margin: auto;">
|
18 |
+
</div>
|
19 |
+
|
20 |
+
Name symbolize by *Confluence* for many unique RP model with *Renegade* mostly come from no-guardrail.
|
21 |
+
|
22 |
+
### Configuration
|
23 |
+
|
24 |
+
The following YAML configuration was used to produce this model:
|
25 |
+
|
26 |
+
```yaml
|
27 |
+
models:
|
28 |
+
- model: ./modela/Erosumika-7B
|
29 |
+
parameters:
|
30 |
+
density: [1, 0.8, 0.6]
|
31 |
+
weight: 0.2
|
32 |
+
- model: ./modela/Infinitely-Laydiculous-7B
|
33 |
+
parameters:
|
34 |
+
density: [0.9, 0.7, 0.5]
|
35 |
+
weight: 0.2
|
36 |
+
- model: ./modela/Kunocchini-7b-128k-test
|
37 |
+
parameters:
|
38 |
+
density: [0.8, 0.6, 0.4]
|
39 |
+
weight: 0.2
|
40 |
+
- model: ./modela/EndlessRP-v3-7B
|
41 |
+
parameters:
|
42 |
+
density: [0.7, 0.5, 0.3]
|
43 |
+
weight: 0.2
|
44 |
+
- model: ./modela/daybreak-kunoichi-2dpo-7b
|
45 |
+
parameters:
|
46 |
+
density: [0.5, 0.3, 0.1]
|
47 |
+
weight: 0.2
|
48 |
+
merge_method: dare_linear
|
49 |
+
base_model: ./modela/Mistral-7B-v0.1
|
50 |
+
parameters:
|
51 |
+
normalize: true
|
52 |
+
int8_mask: true
|
53 |
+
dtype: bfloat16
|
54 |
+
name: intermediate-model
|
55 |
+
---
|
56 |
+
slices:
|
57 |
+
- sources:
|
58 |
+
- model: intermediate-model
|
59 |
+
layer_range: [0, 32]
|
60 |
+
- model: ./modela/BuRP_7B
|
61 |
+
layer_range: [0, 32]
|
62 |
+
merge_method: slerp
|
63 |
+
base_model: intermediate-model
|
64 |
+
parameters:
|
65 |
+
t:
|
66 |
+
- filter: self_attn
|
67 |
+
value: [0, 0.5, 0.3, 0.7, 1]
|
68 |
+
- filter: mlp
|
69 |
+
value: [1, 0.5, 0.7, 0.3, 0]
|
70 |
+
- value: 0.5 # fallback for rest of tensors
|
71 |
+
dtype: bfloat16
|
72 |
+
name: gradient-slerp
|
73 |
+
```
|
74 |
+
|
75 |
+
```mergekit-mega config.yml ./output-model-directory --cuda --allow-crimes --lazy-unpickle```
|
76 |
+
|
77 |
+
### Models Merged
|
78 |
+
|
79 |
+
The following models were included in the merge:
|
80 |
+
- [localfultonextractor/Erosumika-7B](https://huggingface.co/localfultonextractor/Erosumika-7B)
|
81 |
+
- [Nitral-AI/Infinitely-Laydiculous-7B](https://huggingface.co/Nitral-AI/Infinitely-Laydiculous-7B)
|
82 |
+
- [Kunocchini-7b-128k-test](https://huggingface.co/Nitral-AI/Kunocchini-7b-128k-test)
|
83 |
+
- [Endevor/EndlessRP-v3-7B](https://huggingface.co/Endevor/EndlessRP-v3-7B)
|
84 |
+
- [ChaoticNeutrals/BuRP_7B](https://huggingface.co/ChaoticNeutrals/BuRP_7B)
|
85 |
+
- [daybreak-kunoichi-2dpo-7b](https://huggingface.co/crestf411/daybreak-kunoichi-2dpo-7b)
|