Pinkstack commited on
Commit
2cfcc60
1 Parent(s): 71460fc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -13,19 +13,18 @@ license: apache-2.0
13
  language:
14
  - en
15
  - zh
16
- - ru
17
  pipeline_tag: text-generation
18
  ---
19
 
20
  ![PARM-2.png](https://cdn-uploads.huggingface.co/production/uploads/6710ba6af1279fe0dfe33afe/9wMB_c4WmaJR91f-ybFJl.png)
21
 
22
- We are proud to announce, our new high quality flagship model series - ***PARM2***, Very high quality reasoning, math and code for a small size, that **anyone** can run on their device for free.
23
 
24
 
25
  🧀 Which quant is right for you?
26
 
27
  - ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
28
- - ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its noticeably slower than q4
29
 
30
  This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
31
  if you benchmarked this model let me know
 
13
  language:
14
  - en
15
  - zh
 
16
  pipeline_tag: text-generation
17
  ---
18
 
19
  ![PARM-2.png](https://cdn-uploads.huggingface.co/production/uploads/6710ba6af1279fe0dfe33afe/9wMB_c4WmaJR91f-ybFJl.png)
20
 
21
+ We are proud to announce, our new high quality flagship model series - ***PARM2***, Very high quality reasoning, math and coding abilities for a small size, that **anyone** can run on their device for free.
22
 
23
 
24
  🧀 Which quant is right for you?
25
 
26
  - ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
27
+ - ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its slightly slower than Q4.
28
 
29
  This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
30
  if you benchmarked this model let me know