digitous commited on
Commit
f23e899
1 Parent(s): df126f7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -11,9 +11,13 @@ Below is one use case example of what is possible when using Alpac(ino) in a con
11
  An infinite text-based adventure game can be setup by switching to chat mode (in this case username is Player and AI is Narrator) and tailoring the following
12
  template in the context/memory section of the interface:
13
 
 
 
14
  Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
15
  Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
16
  whatever quest or other information to keep consistent in the interaction).
 
 
17
 
18
  If using either KAI or TGUI for model inference, testing subjectively suggests ideal presets are "Storywriter" (temp raised to 1.1) or "Godlike" with context tokens
19
  maxed at 2048 and max generation tokens ~680 or greater. This model is intelligent enough to determine when to stop writing, and will rarely use half those tokens;
 
11
  An infinite text-based adventure game can be setup by switching to chat mode (in this case username is Player and AI is Narrator) and tailoring the following
12
  template in the context/memory section of the interface:
13
 
14
+
15
+ ### Instruction:
16
  Make Narrator function as a text based adventure game that responds with verbose, detailed, and creative descriptions of what happens next after Player's response.
17
  Make Player function as the player input for Narrator's text based adventure game, controlling a character named (insert character name here, their short bio, and
18
  whatever quest or other information to keep consistent in the interaction).
19
+ ### Response:
20
+ (carriage return after response)
21
 
22
  If using either KAI or TGUI for model inference, testing subjectively suggests ideal presets are "Storywriter" (temp raised to 1.1) or "Godlike" with context tokens
23
  maxed at 2048 and max generation tokens ~680 or greater. This model is intelligent enough to determine when to stop writing, and will rarely use half those tokens;