tannedbum commited on
Commit
a0c0a1e
1 Parent(s): 849f315

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +153 -57
README.md CHANGED
@@ -1,57 +1,153 @@
1
- ---
2
- base_model:
3
- - tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B
4
- - tannedbum/L3-Stheno-3SOME-8B
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
-
10
- ---
11
- # L3-Nymeria-v2-8B
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
-
18
- This model was merged using the SLERP merge method.
19
-
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
- * [tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B](https://huggingface.co/tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B)
24
- * [tannedbum/L3-Stheno-3SOME-8B](https://huggingface.co/tannedbum/L3-Stheno-3SOME-8B)
25
-
26
- ### Configuration
27
-
28
- The following YAML configuration was used to produce this model:
29
-
30
- ```yaml
31
- slices:
32
- - sources:
33
- - model: tannedbum/L3-Stheno-3SOME-8B
34
- layer_range: [0, 32]
35
- - model: tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B
36
- layer_range: [0, 32]
37
- merge_method: slerp
38
- base_model: tannedbum/L3-Stheno-3SOME-8B
39
- parameters:
40
- t:
41
- - filter: self_attn
42
- value: [0.4, 0.3, 0.3, 0.6]
43
- - filter: mlp
44
- value: [0.6, 0.7, 0.7, 0.4]
45
- - value: 0.4
46
- dtype: bfloat16
47
-
48
-
49
-
50
-
51
-
52
-
53
-
54
-
55
-
56
-
57
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - Sao10K/L3-8B-Stheno-v3.2
4
+ - chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
5
+ - TheDrummer/Llama-3SOME-8B-v2
6
+ - NeverSleep/Llama-3-Lumimaid-8B-v0.1
7
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
8
+ library_name: transformers
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ - roleplay
13
+ - sillytavern
14
+ - llama3
15
+ - not-for-all-audiences
16
+ license: cc-by-nc-4.0
17
+ language:
18
+ - en
19
+ ---
20
+ ![Nymeria](https://huggingface.co/tannedbum/L3-Nymeria-v2-8B/resolve/main/Nymeria_v2.png?)
21
+
22
+
23
+ - Upgraded SimPO.
24
+ - A touch of 3SOME, Lumimaid and Jamet Blackroot resulting a slightly different prose and wider RP vocab.
25
+ - She can get moody. Don't make her angry.
26
+ - Leans slightly more on nsfw than the original.
27
+
28
+
29
+ ## SillyTavern
30
+
31
+ ## Text Completion presets
32
+ ```
33
+ temp 0.9
34
+ top_k 30
35
+ top_p 0.75
36
+ min_p 0.2
37
+ rep_pen 1.1
38
+ smooth_factor 0.25
39
+ smooth_curve 1
40
+ ```
41
+ ## Advanced Formatting
42
+
43
+ [Context & Instruct preset by Virt-io](https://huggingface.co/Virt-io/SillyTavern-Presets/tree/main/Prompts/LLAMA-3/v2.0)
44
+
45
+ Instruct Mode: Enabled
46
+
47
+
48
+
49
+ # merge
50
+
51
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
52
+
53
+ This model was merged using the slerp merge method.
54
+
55
+ ### Models Merged
56
+
57
+ The following models were included in the merge:
58
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
59
+ * [chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO](https://huggingface.co/chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO)
60
+ * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)
61
+ * [NeverSleep/Llama-3-Lumimaid-8B-v0.1](https://huggingface.co/NeverSleep/Llama-3-Lumimaid-8B-v0.1)
62
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
63
+
64
+ ### Configuration
65
+
66
+ The following YAML configuration was used to produce this model:
67
+
68
+ ```yaml
69
+
70
+ slices:
71
+ - sources:
72
+ - model: NeverSleep/Llama-3-Lumimaid-8B-v0.1
73
+ layer_range: [0, 32]
74
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
75
+ layer_range: [0, 32]
76
+ merge_method: slerp
77
+ base_model: NeverSleep/Llama-3-Lumimaid-8B-v0.1
78
+ parameters:
79
+ t:
80
+ - filter: self_attn
81
+ value: [0.7, 0.3, 0.3, 0.3]
82
+ - filter: mlp
83
+ value: [0.3, 0.7, 0.7, 0.7]
84
+ - value: 0.4
85
+ dtype: bfloat16
86
+
87
+ L3-Lumimaid-Jamet-Blackroot-8B
88
+
89
+
90
+ slices:
91
+ - sources:
92
+ - model: tannedbum/L3-Lumimaid-Jamet-Blackroot-8B
93
+ layer_range: [0, 32]
94
+ - model: chujiezheng/Llama-3-Instruct-8B-SimPO-ExPO
95
+ layer_range: [0, 32]
96
+ merge_method: slerp
97
+ base_model: tannedbum/L3-Lumimaid-Jamet-Blackroot-8B
98
+ parameters:
99
+ t:
100
+ - filter: self_attn
101
+ value: [0.3, 0.7, 0.7, 0.7]
102
+ - filter: mlp
103
+ value: [0.7, 0.3, 0.3, 0.3]
104
+ - value: 0.6
105
+ dtype: bfloat16
106
+
107
+ L3-SimPO-Lumimaid-Jamet-Blackroot-8B
108
+
109
+
110
+ slices:
111
+ - sources:
112
+ - model: Sao10K/L3-8B-Stheno-v3.2
113
+ layer_range: [0, 32]
114
+ - model: TheDrummer/Llama-3SOME-8B-v2
115
+ layer_range: [0, 32]
116
+ merge_method: slerp
117
+ base_model: Sao10K/L3-8B-Stheno-v3.2
118
+ parameters:
119
+ t:
120
+ - filter: self_attn
121
+ value: [0.3, 0.3, 0.7, 0.3]
122
+ - filter: mlp
123
+ value: [0.7, 0.7, 0.3, 0.7]
124
+ - value: 0.4
125
+ dtype: bfloat16
126
+
127
+ L3-Stheno-3SOME-8B
128
+
129
+
130
+ slices:
131
+ - sources:
132
+ - model: tannedbum/L3-Stheno-3SOME-8B
133
+ layer_range: [0, 32]
134
+ - model: tannedbum/L3-SimPO-Lumimaid-Jamet-Blackroot-8B
135
+ layer_range: [0, 32]
136
+ merge_method: slerp
137
+ base_model: tannedbum/L3-Stheno-3SOME-8B
138
+ parameters:
139
+ t:
140
+ - filter: self_attn
141
+ value: [0.4, 0.3, 0.3, 0.6]
142
+ - filter: mlp
143
+ value: [0.6, 0.7, 0.7, 0.4]
144
+ - value: 0.4
145
+ dtype: bfloat16
146
+
147
+ L3-Nymeria-v2-8B
148
+
149
+
150
+
151
+ ```
152
+
153
+ Want to support my work ? My Ko-fi page: https://ko-fi.com/tannedbum