DavidAU commited on
Commit
3424995
1 Parent(s): 097d21b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -58,12 +58,14 @@ Example outputs below.
58
  - This is not a "happy ever after" model. It has a negative bias.
59
  - Output length will vary however this model prefers shortly outputs unless you state the size.
60
  - For creative uses, different quants will produce slightly different output.
 
61
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
62
  - Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
63
 
64
  Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
65
 
66
- The Imatrix versions of this model have even lower perplexity then both this model and Llama3 Instruct and enhanced output.
 
67
 
68
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
69
  However this can be extended using "rope" settings up to 32k.
 
58
  - This is not a "happy ever after" model. It has a negative bias.
59
  - Output length will vary however this model prefers shortly outputs unless you state the size.
60
  - For creative uses, different quants will produce slightly different output.
61
+ - Due to the high stability and compressed nature of this model, all quants will operate at above average levels.
62
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
63
  - Source code for this model (Bfloat16), Float 32 master GGUFs (and source), and Imatrix GGUFs versions will be uploaded shortly at separate repos.
64
 
65
  Note the "float32" version of this model behaves VERY differently which is why it was not uploaded first.
66
 
67
+ The Imatrix versions of this model have even lower perplexity (1/2 level of magnitude lower than this model, 1 full level of magnitude
68
+ lower than LLama3 Instruct) then both this model and Llama3 Instruct and enhanced output.
69
 
70
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 8k / 8192.
71
  However this can be extended using "rope" settings up to 32k.