Doctor-Shotgun commited on
Commit
e302c4f
·
1 Parent(s): 147562c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -9
README.md CHANGED
@@ -2,27 +2,70 @@
2
  tags:
3
  - generated_from_trainer
4
  model-index:
5
- - name: limarp-lora-out
6
  results: []
 
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
11
-
12
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
13
- # limarp-lora-out
 
 
14
 
15
- This model was trained from scratch on the None dataset.
16
  It achieves the following results on the evaluation set:
17
  - Loss: 1.8232
18
 
19
  ## Model description
20
 
21
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
23
  ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
  ## Training and evaluation data
28
 
@@ -77,4 +120,4 @@ The following hyperparameters were used during training:
77
  - Transformers 4.35.0.dev0
78
  - Pytorch 2.0.1+cu118
79
  - Datasets 2.14.5
80
- - Tokenizers 0.14.1
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
5
+ - name: limarpv3-llama2-70b-qlora
6
  results: []
7
+ license: apache-2.0
8
  ---
9
 
 
 
 
10
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
11
+ # limarpv3-llama2-70b-qlora
12
+
13
+ This model is an unofficial Llama 2 70B training on the LimaRP v3 dataset by [lemonilia](https://huggingface.co/lemonilia). It does not include the pretraining stage using stories.
14
 
 
15
  It achieves the following results on the evaluation set:
16
  - Loss: 1.8232
17
 
18
  ## Model description
19
 
20
+ For more details about LimaRP, see the model page for the [previously released v2 version for Llama-2](https://huggingface.co/lemonilia/limarp-llama2-v2). Most details written there apply for this version as well. Generally speaking, LimaRP is a longform-oriented, novel-style roleplaying chat model intended to replicate the experience of 1-on-1 roleplay on Internet forums. Short-form, IRC/Discord-style RP (aka "Markdown format") is not supported yet. The model does not include instruction tuning, only manually picked and slightly edited RP conversations with persona and scenario data.
21
+
22
+ Prompt format is the [extended Alpaca format](https://github.com/tatsu-lab/stanford_alpaca):
23
+
24
+ ```
25
+ ### Instruction:
26
+ Character's Persona: {bot character description}
27
+ User's Persona: {user character description}
28
+ Scenario: {what happens in the story}
29
+ Play the role of Character. You must engage in a roleplaying chat with User below this line. Do not write dialogues and narration for User.
30
+ ### Input:
31
+ User: {utterance}
32
+ ### Response:
33
+ Character: {utterance}
34
+ ### Input
35
+ User: {utterance}
36
+ ### Response:
37
+ Character: {utterance}
38
+ (etc.)
39
+ ```
40
+
41
+ Inspired by the previously named "Roleplay" preset in SillyTavern, with this version of LimaRP it is possible to append a length modifier to the response instruction sequence, like this:
42
+
43
+ ```
44
+ ### Input
45
+ User: {utterance}
46
+
47
+ ### Response: (length = medium)
48
+ Character: {utterance}
49
+ ```
50
+
51
+ This has an immediately noticeable effect on bot responses. The lengths using during training are:
52
+ `micro`, `tiny`, `short`, `medium`, `long`, `massive`, `huge`, `enormous`, `humongous`, `unlimited`.
53
+ **The recommended starting length is medium**. Keep in mind that the AI can ramble or impersonate
54
+ the user with very long messages.
55
+
56
+ The length control effect is reproducible, but the messages will not necessarily follow
57
+ lengths very precisely, rather follow certain ranges on average, as seen in this table
58
+ with data from tests made with one reply at the beginning of the conversation:
59
+
60
+ ![lengths](https://i.imgur.com/2WXGgaV.png)
61
+
62
+ Response length control appears to work well also deep into the conversation. **By omitting
63
+ the modifier, the model will choose the most appropriate response length** (although it might
64
+ not necessarily be what the user desires).
65
 
66
  ## Intended uses & limitations
67
 
68
+ The model will show biases similar to those observed in niche roleplaying forums on the Internet, besides those exhibited by the base model.
69
 
70
  ## Training and evaluation data
71
 
 
120
  - Transformers 4.35.0.dev0
121
  - Pytorch 2.0.1+cu118
122
  - Datasets 2.14.5
123
+ - Tokenizers 0.14.1