DavidAU commited on
Commit
871b089
·
verified ·
1 Parent(s): 2817b57

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +187 -0
README.md ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - zh
6
+ tags:
7
+ - MOE
8
+ - Qwen 2.5 MOE
9
+ - Mixture of Experts
10
+ - Uncensored
11
+ - 2X1.5B
12
+ - deepseek
13
+ - reasoning
14
+ - thinking
15
+ - creative
16
+ - 128k context
17
+ - general usage
18
+ - problem solving
19
+ - brainstorming
20
+ - solve riddles
21
+ - story generation
22
+ - plot generation
23
+ - storytelling
24
+ - fiction story
25
+ - story
26
+ - writing
27
+ - fiction
28
+ - Qwen 2.5
29
+ - mergekit
30
+ pipeline_tag: text-generation
31
+ ---
32
+
33
+ (quants uploading, examples to be added)
34
+
35
+ <H2>Qwen2.5-MOE-2X7B-DeepSeek-Abliterated-Censored-15B-gguf</H2>
36
+
37
+ <img src="qwen-tiny.jpg" style="float:right; width:300px; height:300px; padding:5px;">
38
+
39
+ This is a Qwen2.5 MOE (Mixture of Experts) model comprised of TWO Qwen 2.5 Deepseek (Censored/Normal AND Uncensored) 7B models
40
+ creating a 15B model with the "Abliterated" (Uncensored) version of Deepseek Qwen 2.5 7B "in charge" so to speak.
41
+
42
+ The model is just over 15B because of the unqiue "shared expert" (roughly 2.5 models here) used in Qwen MOEs.
43
+
44
+ The oddball configuration yields interesting "thinking/reasoning" which is stronger than either 1.5B model on its own.
45
+
46
+ Example generations at the bottom of this page.
47
+
48
+ This model can be used for all use cases, and is also (mostly) uncensored.
49
+
50
+ Context: 128k.
51
+
52
+ You need to use the "Jinja Template" encoded in the GGUF to use this model. You might be able to use Llama 3, and/or Chatml templates
53
+ if your AI/LLM app can not access the "Jinja Template".
54
+
55
+ In Lmstudio the "Jinja Template" should load by default.
56
+
57
+ In other apps - use the Deepseek Tokenizer and/or "Jinja Template".
58
+
59
+ This model contains 2 times the power of DeepSeek Distill reasoning/thinking and shows exceptional performance.
60
+
61
+ Also, the DeepSeek Qwen 7B model is based on Qwen's 7B Math model so this model is slanted more towards math/logic problem solving
62
+ and I would also say more "sciency" too.
63
+
64
+ This does not mean it will not work for your use case.
65
+
66
+ Also, because of how this model works (uncensored and censored in the same model) you may want to try 1-4 generations depending
67
+ on your use case because even the "right" response will vary widely, and in many cases be more "interesting".
68
+
69
+ Examples below so you have some idea what this model can do.
70
+
71
+ Keep in mind this model is two 7B parameters models working together, and will come close but may not have the power of a 14B or 32B reasoning/thinking model.
72
+
73
+ However, sometimes it will generate truly "out of the park" responses.
74
+
75
+ Temp of .4 to .8 is suggested (for best reasoning/thinking), however it will still operate at much higher temps like 1.8, 2.6 etc.
76
+
77
+ Depending on your prompt change temp SLOWLY: IE: .41,.42,.43 ... etc etc.
78
+
79
+ The model MAY function better if you breakdown the reasoning/thinking task(s) into smaller pieces :
80
+
81
+ "IE: Instead of asking for 6 plots FOR theme XYZ, ASK IT for ONE plot for theme XYZ at a time".
82
+
83
+ Also set context limit at 4k minimum, 8K+ suggested.
84
+
85
+ I also suggest quant of IQ4/Q4 or higher, as larger quants will reasoning/thinking and perform much better.
86
+
87
+ If you can run Q6/Q8, please use these one(s).
88
+
89
+ IQ4XS will give very different responses VS other quants.
90
+
91
+ ---
92
+
93
+ <B> Additional Support / Documents for this model to assist with generation / performance: </b>
94
+
95
+ Document #1:
96
+
97
+ Details how to use reasoning/thinking models and get maximum performance from them, and includes links to all reasoning/thinking models - GGUF and source, as well as adapters to turn any "regular" model into a "reasoning/thinking" model.
98
+
99
+ [ https://huggingface.co/DavidAU/How-To-Use-Reasoning-Thinking-Models-and-Create-Them ]
100
+
101
+ Document #2:
102
+
103
+ Document detailing all parameters, settings, samplers and advanced samplers to use not only my models to their maximum potential - but all models (and quants) online (regardless of the repo) to their maximum potential. Included quick start and detailed notes, include AI / LLM apps and other critical information and references too. A must read if you are using any AI/LLM right now.
104
+
105
+ [ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
106
+
107
+ Software:
108
+
109
+ SOFTWARE patch (by me) for Silly Tavern (front end to connect to multiple AI apps / connect to AIs- like Koboldcpp, Lmstudio, Text Gen Web UI and other APIs) to control and improve output generation of ANY AI model. Also designed to control/wrangle some of my more "creative" models and make them perform perfectly with little to no parameter/samplers adjustments too.
110
+
111
+ [ https://huggingface.co/DavidAU/AI_Autocorrect__Auto-Creative-Enhancement__Auto-Low-Quant-Optimization__gguf-exl2-hqq-SOFTWARE ]
112
+
113
+ ---
114
+
115
+ <h2>Example Generation:</h2>
116
+
117
+ IQ4XS Quant, Temp 1.5, rep pen 1.06, topp: .95, minp: .05, topk: 40
118
+
119
+ ---
120
+
121
+ EXAMPLE #1:
122
+
123
+ ---
124
+
125
+ <B>
126
+
127
+ </B>
128
+
129
+ [[[Thinking Start]]]
130
+
131
+
132
+ [[[Thinking End]]]
133
+
134
+ OUTPUT:
135
+
136
+ ---
137
+
138
+ EXAMPLE #2:
139
+
140
+ ---
141
+
142
+ <B>
143
+
144
+ </B>
145
+
146
+ [[[Thinking Start]]]
147
+
148
+
149
+ [[[Thinking End]]]
150
+
151
+ OUTPUT:
152
+
153
+ ---
154
+
155
+ EXAMPLE #3:
156
+
157
+ ---
158
+
159
+ <B>
160
+
161
+ </B>
162
+
163
+ [[[Thinking Start]]]
164
+
165
+
166
+ [[[Thinking End]]]
167
+
168
+ OUTPUT:
169
+
170
+ ---
171
+
172
+ EXAMPLE #4:
173
+
174
+ ---
175
+
176
+ <B>
177
+
178
+ </B>
179
+
180
+ [[[Thinking Start]]]
181
+
182
+
183
+ [[[Thinking End]]]
184
+
185
+ OUTPUT:
186
+
187
+