Update README.md
Browse files
README.md
CHANGED
@@ -11,9 +11,13 @@ Below is one use case example of what is possible when using Alpac(ino) in a con
|
|
11 |
An infinite text-based adventure game can be setup by switching to chat mode (in this case username is Player and AI is Narrator) and tailoring the following
|
12 |
template in the context/memory section of the interface:
|
13 |
|
|
|
|
|
14 |
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
|
15 |
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
|
16 |
whatever quest or other information to keep consistent in the interaction).
|
|
|
|
|
17 |
|
18 |
If using either KAI or TGUI for model inference, testing subjectively suggests ideal presets are "Storywriter" (temp raised to 1.1) or "Godlike" with context tokens
|
19 |
maxed at 2048 and max generation tokens ~680 or greater. This model is intelligent enough to determine when to stop writing, and will rarely use half those tokens;
|
|
|
11 |
An infinite text-based adventure game can be setup by switching to chat mode (in this case username is Player and AI is Narrator) and tailoring the following
|
12 |
template in the context/memory section of the interface:
|
13 |
|
14 |
+
|
15 |
+
### Instruction:
|
16 |
Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
|
17 |
Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
|
18 |
whatever quest or other information to keep consistent in the interaction).
|
19 |
+
### Response:
|
20 |
+
(carriage return after response)
|
21 |
|
22 |
If using either KAI or TGUI for model inference, testing subjectively suggests ideal presets are "Storywriter" (temp raised to 1.1) or "Godlike" with context tokens
|
23 |
maxed at 2048 and max generation tokens ~680 or greater. This model is intelligent enough to determine when to stop writing, and will rarely use half those tokens;
|