Update README.md
Browse files
README.md
CHANGED
@@ -4,24 +4,27 @@ datasets:
|
|
4 |
- ludis/geepeetee4
|
5 |
---
|
6 |
|
7 |
-
|
8 |
|
9 |
-
|
10 |
|
11 |
-
|
12 |
|
13 |
-
|
14 |
|
15 |
-
|
16 |
|
17 |
-
|
18 |
|
19 |
-
|
20 |
|
21 |
-
|
22 |
|
23 |
-
|
24 |
|
25 |
-
|
|
|
|
|
|
|
|
|
26 |
|
27 |
-
The `<|system|>` prompt can be used to inject out-of-channel information behind the scenes, while the `<|user|>` prompt should be used to indicate user input. The `<|model|>` token should then be used to indicate that the model should generate a response. These tokens can happen multiple times and be chained up to form a conversation history.
|
|
|
4 |
- ludis/geepeetee4
|
5 |
---
|
6 |
|
7 |
+
## GGUF
|
8 |
|
9 |
+
gguf quants for ludis/tsukasa-limarp-7b
|
10 |
|
11 |
+
## Prompting
|
12 |
|
13 |
+
https://rentry.org/v43eo - reccomended prompts and gen settings
|
14 |
|
15 |
+
The current model version has been trained on prompts using three different roles, which are denoted by the following tokens: `<|system|>`, `<|user|>` and `<|model|>`.
|
16 |
|
17 |
+
The `<|system|>` prompt can be used to inject out-of-channel information behind the scenes, while the `<|user|>` prompt should be used to indicate user input. The `<|model|>` token should then be used to indicate that the model should generate a response. These tokens can happen multiple times and be chained up to form a conversation history.
|
18 |
|
19 |
+
## Training
|
20 |
|
21 |
+
base model (llama-2-13b-hf)
|
22 |
|
23 |
+
tuned on koishi dataset (commit c83d922) for 1 epoch
|
24 |
|
25 |
+
then tuned on pippa dataset (commit 6412b0c) for 1 epoch
|
26 |
+
|
27 |
+
then tuned on geepeetee4 dataset (commit c83d922) for 1 epoch
|
28 |
+
|
29 |
+
then tuned on limarp (without ponyville, lolicit, and all the fallen subsets. Version 2023-09-14) for 2 epochs
|
30 |
|
|