digitous commited on
Commit
9e95aa6
1 Parent(s): 83a29cb

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -0
README.md ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - llama
4
+ - alpaca
5
+ - vicuna
6
+ - uncensored
7
+ - cot
8
+ - chain of thought
9
+ - story
10
+ - adventure
11
+ - roleplay
12
+ - rp
13
+ - merge
14
+ - mix
15
+ - instruct
16
+ - wizardlm
17
+ - superhot
18
+ - supercot
19
+ - manticore
20
+ - hippogriff
21
+ ---
22
+
23
+ ## 30B-Epsilon
24
+
25
+ Epsilon is an instruct based general purpose model assembled from hand picked models and LoRAs.
26
+ There is no censorship and it follows instructions in the Alpaca format. This means you can create
27
+ your own rules in the context memory of your inference system of choice [mainly KoboldAI or Text
28
+ Generation Webui and chat UIs like SillyTavern and so on].
29
+
30
+ ## Composition:
31
+
32
+ This model is the result of an experimental use of LoRAs on language models and model merges.
33
+ [] = applied as LoRA to a composite model | () = combined as composite models
34
+ 30B-Epsilon = [SuperCOT[SuperHOT-prototype13b-8192[(wizardlmuncensored+((hippogriff+manticore)+(StoryV2))]
35
+
36
+ Alpaca's instruct format can be used to do many things, including control of the terms of behavior
37
+ between a user and a response from an agent in chat. Below is an example of a command injected into
38
+ memory.
39
+
40
+ ```
41
+ ### Instruction:
42
+ Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
43
+ Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
44
+ whatever quest or other information to keep consistent in the interaction).
45
+
46
+ ### Response:
47
+ {an empty new line here}
48
+ ```
49
+
50
+ All datasets from all models and LoRAs used were documented and reviewed as model candidates for merging.
51
+ Model candidates were based on five core principles: creativity, logic, inference, instruction following,
52
+ and longevity of trained responses. SuperHOTv2 did not have datasets public at the time it was selected,
53
+ however it was included as a key component of Epsilon considering the datasets its predecessor
54
+ included, which are no longer available to view as it has been removed [kaiokendev-SuperHOT-LoRA-prototype30b-8192].
55
+ SuperHot 2 is utilized regardless in light of that. The GPT4Alpaca LoRA from Chansung was removed from this
56
+ amalgam following a thorough review of where censorship and railroading the user came from in 33B-Lazarus.
57
+ This is not a reflection of ChanSung's excellent work - it merely did not fit the purpose of this model.
58
+
59
+ ## Language Models and LoRAs Used Credits:
60
+
61
+ manticore-30b-chat-pyg-alpha [Epoch0.4] by openaccess-ai-collective
62
+
63
+ https://huggingface.co/openaccess-ai-collective/manticore-30b-chat-pyg-alpha
64
+
65
+ hippogriff-30b-chat by openaccess-ai-collective
66
+
67
+ https://huggingface.co/openaccess-ai-collective/hippogriff-30b-chat
68
+
69
+ WizardLM-33B-V1.0-Uncensored by ehartford
70
+
71
+ https://huggingface.co/ehartford/WizardLM-33B-V1.0-Uncensored
72
+
73
+ Storytelling-LLaMa-LoRA [30B, Version 2] by GamerUnTouch
74
+
75
+ https://huggingface.co/GamerUntouch/Storytelling-LLaMa-LoRAs
76
+
77
+ SuperCOT-LoRA [30B] by kaiokendev
78
+
79
+ https://huggingface.co/kaiokendev/SuperCOT-LoRA
80
+
81
+ SuperHOT-LoRA-prototype30b-8192 [30b, not 8K version, but a removed prototype] by kaiokendev
82
+
83
+ https://huggingface.co/kaiokendev/superhot-30b-8k-no-rlhf-test [Similar LoRA to one since removed that was used in making this model.]
84
+
85
+ Also thanks to Meta for LLaMA and to each and every one of you
86
+ who developed these fine-tunes and LoRAs.