totally-not-an-llm commited on
Commit
825f454
·
1 Parent(s): 44748e3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -12,6 +12,8 @@ Introducing EverythingLM, a llama-2 based, general-purpose 13b model with 16k co
12
 
13
  The model is completely uncensored.
14
 
 
 
15
  ### Notable features:
16
  - Automatically triggered CoT reasoning
17
  - Verbose and detailed replies
@@ -29,8 +31,6 @@ ASSISTANT:
29
 
30
  Training took about 1 hour using QLoRa on 1xA100, so this model can be recreated for about $3. QLoRa model can be found here: https://huggingface.co/totally-not-an-llm/EverythingLM-13b-peft.
31
 
32
- This is an early test, so here are some things to note on the model:
33
-
34
  ### Model quirks:
35
  - Due to the nature of the dataset, it does better with more detail. I've found it gives much better stories when I provide more requirements
36
  - It really likes to use numbered lists. I don't necessarilly have a problem with this but it's something to note when training on the dataset
 
12
 
13
  The model is completely uncensored.
14
 
15
+ This model is an early test of the EverythingLM dataset and some new experimental principles, so don't consider it SOTA.
16
+
17
  ### Notable features:
18
  - Automatically triggered CoT reasoning
19
  - Verbose and detailed replies
 
31
 
32
  Training took about 1 hour using QLoRa on 1xA100, so this model can be recreated for about $3. QLoRa model can be found here: https://huggingface.co/totally-not-an-llm/EverythingLM-13b-peft.
33
 
 
 
34
  ### Model quirks:
35
  - Due to the nature of the dataset, it does better with more detail. I've found it gives much better stories when I provide more requirements
36
  - It really likes to use numbered lists. I don't necessarilly have a problem with this but it's something to note when training on the dataset