rAIfle commited on
Commit
0939353
1 Parent(s): 6a1f08f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +93 -0
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ ```
6
+ e88 88e d8
7
+ d888 888b 8888 8888 ,"Y88b 888 8e d88
8
+ C8888 8888D 8888 8888 "8" 888 888 88b d88888
9
+ Y888 888P Y888 888P ,ee 888 888 888 888
10
+ "88 88" "88 88" "88 888 888 888 888
11
+ b
12
+ 8b,
13
+
14
+ e88'Y88 d8 888
15
+ d888 'Y ,"Y88b 888,8, d88 ,e e, 888
16
+ C8888 "8" 888 888 " d88888 d88 88b 888
17
+ Y888 ,d ,ee 888 888 888 888 , 888
18
+ "88,d88 "88 888 888 888 "YeeP" 888
19
+
20
+ PROUDLY PRESENTS
21
+ ```
22
+ # Dendrite-L3-10B-exl2-rpcal
23
+
24
+ Quantized using 200 samples of 8192 tokens from an RP-oriented [PIPPA](https://huggingface.co/datasets/royallab/PIPPA-cleaned) dataset.
25
+
26
+ Branches:
27
+ - `main` -- `measurement.json`
28
+ - `8b8h` -- 8bpw, 8bit lm_head
29
+ - `6b6h` -- 6bpw, 6bit lm_head
30
+ - `4b6h` -- 4bpw, 6bit lm_head
31
+
32
+ Original model link: [Envoid/Dendrite-L3-10B](https://huggingface.co/Envoid/Dendrite-L3-10B)
33
+
34
+ Original model README below.
35
+
36
+ -----
37
+
38
+ # This model is experimental and thus results cannot be gauranteed.
39
+
40
+ ![](https://files.catbox.moe/rx5tfs.jpg)
41
+ # Dendrite-L3-10B
42
+
43
+ In a similar vein to [Libra-19B](https://huggingface.co/Envoid/Libra-19B) this model was created by taking all of the layers of one model and stacking along with them the first number of layers (8 in this case) from a donor model but in the reverse order.
44
+
45
+ In this case the base model used was [Poppy_Porpoise-DADA-8B](https://huggingface.co/Envoid/Poppy_Porpoise-DADA-8B) and the donor model used was [Llama-3-8B-Instruct-DADA](https://huggingface.co/Envoid/Llama-3-8B-Instruct-DADA)
46
+
47
+ It was then finetuned for 10 epochs on the Dendrite dataset at a low learning rate to repair the disorder and integrate the donor layers.
48
+
49
+ The following mergekit config was used:
50
+ ```
51
+ slices:
52
+ - sources:
53
+ - model: ./Poppy_Porpoise-DADA-8B
54
+ layer_range: [0, 32]
55
+ - sources:
56
+ - model: ./Llama-3-8B-Instruct-DADA
57
+ layer_range: [7, 8]
58
+ - sources:
59
+ - model: ./Llama-3-8B-Instruct-DADA
60
+ layer_range: [6, 7]
61
+ - sources:
62
+ - model: ./Llama-3-8B-Instruct-DADA
63
+ layer_range: [5, 6]
64
+ - sources:
65
+ - model: ./Llama-3-8B-Instruct-DADA
66
+ layer_range: [4, 5]
67
+ - sources:
68
+ - model: ./Llama-3-8B-Instruct-DADA
69
+ layer_range: [3, 4]
70
+ - sources:
71
+ - model: ./Llama-3-8B-Instruct-DADA
72
+ layer_range: [2, 3]
73
+ - sources:
74
+ - model: ./Llama-3-8B-Instruct-DADA
75
+ layer_range: [1, 2]
76
+ - sources:
77
+ - model: ./Llama-3-8B-Instruct-DADA
78
+ layer_range: [0, 1]
79
+ merge_method: passthrough
80
+ dtype: float16
81
+ ```
82
+
83
+ Unlike in the case of Libra-19B this models moral alignment seems very much intact.
84
+
85
+ In order to get the best results from this model you should uncheck "skip special tokens" on your front-end and add "<|eot_id|>" to your custom stopping strings.
86
+
87
+ It has been tested with a number of different Llama-3 prompt templates and seems to work well.
88
+
89
+ It regained its base assistant personality during the retraining process, however, using assistant style prompt templates and assistant cards in SillyTavern gives it fairly interesting replies.
90
+
91
+ It has been tested in RP, assistant and creative writing use cases and at a quick glance seems to work well.
92
+
93
+ Training was done using [qlora-pipe](https://github.com/tdrussell/qlora-pipe)