invisietch commited on
Commit
e674bc1
1 Parent(s): de3ef8e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +138 -3
README.md CHANGED
@@ -1,3 +1,138 @@
1
- ---
2
- license: llama3
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+ - not-for-all-audiences
8
+ license: llama3
9
+ language:
10
+ - en
11
+ ---
12
+
13
+ <div align="center">
14
+ <b style="font-size: 36px;">EtherealRainbow-v0.2-8B (GGUF)</b>
15
+
16
+ <img src="https://huggingface.co/invisietch/EtherealRainbow-v0.2-8B/resolve/main/ethrheader.png" style="width:60%">
17
+ </div>
18
+
19
+ # Model Details
20
+
21
+ Ethereal Rainbow is an 8B parameter merge of various Llama3-based finetunes created using mergekit. The purpose of Ethereal Rainbow is to
22
+ create an uncensored Llama3 variant which is capable of writing creative prose, and engaging in SFW as well as NSFW roleplay and
23
+ storytelling, with a strong focus on long-form responses &amp; adherence to prompts.
24
+
25
+ I don't consider v0.2 to be a 'finished product' by any means, but I've been getting some good results out of it so I wanted to share.
26
+
27
+ # Formats
28
+
29
+ [Safetensors](https://huggingface.co/invisietch/EtherealRainbow-v0.2-8B) | [GGUF](https://huggingface.co/invisietch/EtherealRainbow-v0.2-8B-GGUF)
30
+
31
+ # Disclaimer
32
+
33
+ This model is built on an abliterated base and as such is largely uncensored. It can generate explicit, disturbing or offensive responses.
34
+ Use responsibly. I am not responsible for your use of this model.
35
+
36
+ # Known Issues
37
+
38
+ There are a couple of known issues with the model which I hope to fix in a future release:
39
+
40
+ * It prefers asterisks for actions and bare speech roleplay style. You can work around this by swiping or editing the first 2-3 responses
41
+ to match the style you want, and then it tends to follow your style from there.
42
+ * It runs 'hot' by default, so the temperature you use for other models is likely to be too high. I recommend starting with 0.7 &amp;
43
+ slowly increasing from there.
44
+
45
+ # Merge Details
46
+
47
+ ## Merge Methodology
48
+
49
+ EtherealRainbow is a two-stage merge. I create two base models -- **Rain** & **Sun** -- via `dare_ties`, and then using **Rain** as base, I use
50
+ `slerp` to bring in part of **Sun**.
51
+
52
+ ## Rain
53
+
54
+ The purpose of **Rain** is to create an uncensored model focused purely on roleplay. For this, I chose four component models:
55
+
56
+ * **mlabonne/NeuralDaredevil-8B-abliterated (base)** because it's a mostly-uncensored, coherent, high MMLU Llama-3 base.
57
+ * **Sao10K/L3-8B-Stheno-v3.2** because it has my favorite creative writing of all Llama-3 RP variants.
58
+ * **Nitral-AI/Hathor-L3-8B-v.02** to broaden the model's knowledge of creative writing styles.
59
+ * **grimjim/Llama-3-Luminurse-v0.2-OAS-8B** to improve ERP with more anatomical knowledge.
60
+
61
+ This is the mergekit config that I used to build **Rain**:
62
+
63
+ ```yaml
64
+ models:
65
+ - model: mlabonne/NeuralDaredevil-8B-abliterated
66
+ - model: Sao10K/L3-8B-Stheno-v3.2
67
+ parameters:
68
+ density: 0.53
69
+ weight: 0.4
70
+ - model: Nitral-AI/Hathor-L3-8B-v.02
71
+ parameters:
72
+ density: 0.37
73
+ weight: 0.4
74
+ - model: grimjim/Llama-3-Luminurse-v0.2-OAS-8B
75
+ parameters:
76
+ density: 0.26
77
+ weight: 0.2
78
+ merge_method: dare_ties
79
+ base_model: mlabonne/NeuralDaredevil-8B-abliterated
80
+ parameters:
81
+ int8_mask: true
82
+ dtype: bfloat16
83
+ ```
84
+
85
+ ## Sun
86
+
87
+ The purpose of **Sun** is to improve intelligence, particularly on two fronts: creative writing ability and anatomical knowledge.
88
+ For this, I chose four more component models:
89
+
90
+ * **Gryphe/Pantheon-RP-1.0-8b-Llama-3** because it's trained on large RP datasets that aren't in other models I used.
91
+ * **aaditya/Llama3-OpenBioLLM-8B** for its anatomical and medical knowledge.
92
+ * **Blackroot/Llama-3-LongStory** because it's uniquely trained on much longer story texts, to improve output length.
93
+ * **Locutusque/Llama-3-Hercules-5.0-8B** for similar reasons to Blackroot/Llama-3-LongStory.
94
+
95
+ This is the config that I used to create **Sun**:
96
+
97
+ ```yaml
98
+ models:
99
+ - model: Gryphe/Pantheon-RP-1.0-8b-Llama-3
100
+ - model: aaditya/Llama3-OpenBioLLM-8B
101
+ parameters:
102
+ density: 0.36
103
+ weight: 0.2
104
+ - model: Blackroot/Llama-3-LongStory
105
+ parameters:
106
+ density: 0.40
107
+ weight: 0.3
108
+ - model: Locutusque/Llama-3-Hercules-5.0-8B
109
+ parameters:
110
+ density: 0.49
111
+ weight: 0.5
112
+ merge_method: dare_ties
113
+ base_model: Gryphe/Pantheon-RP-1.0-8b-Llama-3
114
+ parameters:
115
+ int8_mask: true
116
+ dtype: bfloat16
117
+ ```
118
+
119
+ *Note: **Sun** is unusable by itself, it's very prone to endless generation issues, but after 400+ messages I have not seen*
120
+ *this issue with **EtherealRainbow**.*
121
+
122
+ ## EtherealRainbow
123
+
124
+ This is the config that I used to create **EtherealRainbow v0.2**:
125
+
126
+ ```yaml
127
+ models:
128
+ - model: ./fp16/rain-8B
129
+ - model: ./fp16/sun-8B
130
+ merge_method: slerp
131
+ base_model: ./fp16/rain-8B
132
+ parameters:
133
+ t:
134
+ - value: [0, 0, 0.1, 0.3, 0.5, 0.7, 0.5, 0.3, 0.1, 0, 0]
135
+ embed_slerp: true
136
+ dtype: bfloat16
137
+ tokenizer_source: model:./fp16/rain-8B
138
+ ```