kubernetes-bad
commited on
Commit
•
eda3e8d
1
Parent(s):
807b26c
Update README.md
Browse files
README.md
CHANGED
@@ -1,16 +1,29 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
4 |
|
5 |
-
#
|
6 |
|
7 |
-
|
8 |
|
9 |
-
|
10 |
|
11 |
-
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
```
|
16 |
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
@@ -38,4 +51,30 @@ Fuckana is a friendly and talkative catgirl that has enormous breasts. Her voice
|
|
38 |
|
39 |
### Response:
|
40 |
|
41 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
license: llama2
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
tags:
|
6 |
+
- roleplay
|
7 |
+
- characters
|
8 |
---
|
9 |
|
10 |
+
# CharGen v1
|
11 |
|
12 |
+
> A model for creating characters for role play.
|
13 |
|
14 |
+
Trained on *lots* of character cards both from chub and janitor, with some post-processing.
|
15 |
|
16 |
+
For now, it only supports plaintext cards. Any other variation like plist/w++/etc is entirely untested.
|
17 |
|
18 |
+
Address {{user}} as `User`. Character descriptions work best if they begin with `CharacterName is a ...` - for example *"Martha is a middle-aged woman who is ..."*
|
19 |
+
|
20 |
+
NB: This model is **NOT** for roleplay directly. It creates characters that can then be used in roleplay with some other model like [MythoMax](https://huggingface.co/Gryphe/MythoMax-L2-13b).
|
21 |
+
|
22 |
+
It was trained on dynamic prompt template, so it should be able to accommodate your changes to the prompt.
|
23 |
+
|
24 |
+
Trained as a LoRA, the released model is a merge with [Airoboros 2.2](https://huggingface.co/jondurbin/airoboros-l2-13b-2.2) for extra-good instruction following.
|
25 |
+
|
26 |
+
Prompt template:
|
27 |
|
28 |
```
|
29 |
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
|
|
51 |
|
52 |
### Response:
|
53 |
|
54 |
+
```
|
55 |
+
|
56 |
+
### Dataset
|
57 |
+
|
58 |
+
~34,000 cards from CharacterHub and another ~80,000 cards from Janitor were used as initial dataset, as captured in period between August and September 2023.
|
59 |
+
|
60 |
+
Dataset will not be released, unless explicit permission to do so would be granted from both Chub and Janitor.
|
61 |
+
|
62 |
+
|
63 |
+
## Training procedure
|
64 |
+
|
65 |
+
|
66 |
+
The following `bitsandbytes` quantization config was used during training:
|
67 |
+
- quant_method: bitsandbytes
|
68 |
+
- load_in_8bit: True
|
69 |
+
- load_in_4bit: False
|
70 |
+
- llm_int8_threshold: 6.0
|
71 |
+
- llm_int8_skip_modules: None
|
72 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
73 |
+
- llm_int8_has_fp16_weight: False
|
74 |
+
- bnb_4bit_quant_type: fp4
|
75 |
+
- bnb_4bit_use_double_quant: False
|
76 |
+
- bnb_4bit_compute_dtype: float32
|
77 |
+
### Framework versions
|
78 |
+
|
79 |
+
|
80 |
+
- PEFT 0.6.0.dev0
|