GGUF
Not-For-All-Audiences
nsfw
Undi95 commited on
Commit
19b4e96
1 Parent(s): cd4c82c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -0
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - not-for-all-audiences
5
+ - nsfw
6
+ ---
7
+
8
+ ## MiquMaid v2 2x70 DPO
9
+
10
+ <center>[<a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B-GGUF">V2-70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO-GGUF">V2-70B-DPO</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-GGUF">V2-2x70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-DPO-GGUF">V2-2x70B-DPO</a>]
11
+ </br>
12
+ <div style="width: 100%;">
13
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/Wbzwoko-IZbOJfvPaImre.png" style="display: block; margin: auto;">
14
+ </div></center>
15
+
16
+ This model uses the Alpaca **prompting format**
17
+
18
+ Then, we have done a MoE, made of MiquMaid-v2-70B-DPO and Miqu-70B-DPO base, making the model using the finetune AND the base model for each token, working together.
19
+
20
+ The two model have been trained on DPO for uncensoring, more info on Miqu-70B-DPO [here](Undi95/Miqu-70B-Alpaca-DPO-GGUF)
21
+
22
+ We have seen a significant improvement, so we decided to share that, even if the model is very big.
23
+
24
+ ## Credits:
25
+ - Undi
26
+ - IkariDev
27
+
28
+ ## Description
29
+
30
+ This repo contains GGUF files of MiquMaid-v2-2x70B-DPO.
31
+
32
+ Switch: [FP16](https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-DPO) - [GGUF](https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-DPO-GGUF)
33
+
34
+ ## Training data used:
35
+ - [Aesir datasets](https://huggingface.co/MinervaAI)
36
+ - [NoRobots](https://huggingface.co/datasets/Doctor-Shotgun/no-robots-sharegpt)
37
+ - [limarp](https://huggingface.co/datasets/lemonilia/LimaRP)
38
+ - [toxic-dpo-v0.1-sharegpt](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-sharegpt)
39
+ - [ToxicQAFinal](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicQAFinal)
40
+
41
+ ## DPO training data used:
42
+ - [ToxicDPOqa](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicDPOqa)
43
+ - [toxic-dpo-v0.1-NoWarning](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-NoWarning)
44
+
45
+ ### Custom format:
46
+ ```
47
+ ### Instruction:
48
+ {system prompt}
49
+
50
+ ### Input:
51
+ {input}
52
+
53
+ ### Response:
54
+ {reply}
55
+ ```
56
+
57
+ ## Others
58
+
59
+ Undi: If you want to support us, you can [here](https://ko-fi.com/undiai).
60
+
61
+ IkariDev: Visit my [retro/neocities style website](https://ikaridevgit.github.io/) please kek