Gryphe commited on
Commit
70a6df2
1 Parent(s): 73facb7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,14 +22,14 @@ I originally never intended to publish this model but over time I've become curi
22
  Quantized versions are available from Bartowski: [GGUF](https://huggingface.co/bartowski/Pantheon-RP-1.0-8b-Llama-3-GGUF) - [EXL2](https://huggingface.co/bartowski/Pantheon-RP-1.0-8b-Llama-3-exl2)
23
 
24
  ## Model details
25
- This model features a highly diverse collection of datasets, totaling ~22 million tokens;
26
 
27
  - For general instructions I created GPT 4 and Claude Opus variations of the No-Robots dataset. I actually ended up not including NoRo itself as it made the model worse.
28
  - For roleplay I used an extensive collection of GPT 4 and Claude Opus data, augmented by the always popular LimaRP for the "human factor".
29
  - The Pantheon Roleplay personas were made using Claude 1.3 data, further diversifying the outputs of this model.
30
  - Aiva's persona includes additional datasets featuring questions related to DM world building, Python coding and RSS summarization. (She summarizes my daily news every day!)
31
 
32
- Roughly 25% of the training data was instructional, with the rest being focused on roleplay. Each of these datasets was then carefully balanced to ensure diversity, removing examples where deemed necessary.
33
 
34
  **TLDR;** Download. ChatML prompt format. Have fun! Leave feedback!
35
 
 
22
  Quantized versions are available from Bartowski: [GGUF](https://huggingface.co/bartowski/Pantheon-RP-1.0-8b-Llama-3-GGUF) - [EXL2](https://huggingface.co/bartowski/Pantheon-RP-1.0-8b-Llama-3-exl2)
23
 
24
  ## Model details
25
+ This model features a highly diverse collection of datasets, totaling ~24 million tokens;
26
 
27
  - For general instructions I created GPT 4 and Claude Opus variations of the No-Robots dataset. I actually ended up not including NoRo itself as it made the model worse.
28
  - For roleplay I used an extensive collection of GPT 4 and Claude Opus data, augmented by the always popular LimaRP for the "human factor".
29
  - The Pantheon Roleplay personas were made using Claude 1.3 data, further diversifying the outputs of this model.
30
  - Aiva's persona includes additional datasets featuring questions related to DM world building, Python coding and RSS summarization. (She summarizes my daily news every day!)
31
 
32
+ Roughly 30% of the training data was instructional, with another 25% being used by the Pantheon Persona data. The remaining 45% was filled with roleplay scenarios covering a huge spectrum of situations. Each of these datasets was then carefully balanced to ensure diversity, removing examples where deemed necessary.
33
 
34
  **TLDR;** Download. ChatML prompt format. Have fun! Leave feedback!
35