Pinkstack commited on
Commit
8ba7766
·
verified ·
1 Parent(s): 290c871

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -24,13 +24,12 @@ We are proud to announce, our new high quality flagship model series - ***PARM2*
24
  🧀 Which quant is right for you?
25
 
26
  - ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
27
- - ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its slightly slower than Q4.
 
28
 
29
  This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
30
  if you benchmarked this model let me know
31
 
32
- ⚠️ it may think it's name is Claude if you use our prompt format, due to the training data. we are sorry for this issue but is shouldn't effect the quality of the responses. If you use a chat ml prompt the quality of the responses would be lower but it won't have this issue.
33
-
34
  This is a pretty lite model which can be run on high end phones pretty quickly using the q4 quant.
35
 
36
  # Passes "strawberry" test! (Q8 w/ msty & rtx 3080 10gb) ✅
 
24
  🧀 Which quant is right for you?
25
 
26
  - ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
27
+ - ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its slightly slower than Q4.
28
+ *other formats were not included as Q4,Q8 have the best performance, quality.*
29
 
30
  This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
31
  if you benchmarked this model let me know
32
 
 
 
33
  This is a pretty lite model which can be run on high end phones pretty quickly using the q4 quant.
34
 
35
  # Passes "strawberry" test! (Q8 w/ msty & rtx 3080 10gb) ✅