Update README.md
Browse files
README.md
CHANGED
@@ -13,19 +13,18 @@ license: apache-2.0
|
|
13 |
language:
|
14 |
- en
|
15 |
- zh
|
16 |
-
- ru
|
17 |
pipeline_tag: text-generation
|
18 |
---
|
19 |
|
20 |
![PARM-2.png](https://cdn-uploads.huggingface.co/production/uploads/6710ba6af1279fe0dfe33afe/9wMB_c4WmaJR91f-ybFJl.png)
|
21 |
|
22 |
-
We are proud to announce, our new high quality flagship model series - ***PARM2***, Very high quality reasoning, math and
|
23 |
|
24 |
|
25 |
🧀 Which quant is right for you?
|
26 |
|
27 |
- ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
|
28 |
-
- ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its
|
29 |
|
30 |
This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
|
31 |
if you benchmarked this model let me know
|
|
|
13 |
language:
|
14 |
- en
|
15 |
- zh
|
|
|
16 |
pipeline_tag: text-generation
|
17 |
---
|
18 |
|
19 |
![PARM-2.png](https://cdn-uploads.huggingface.co/production/uploads/6710ba6af1279fe0dfe33afe/9wMB_c4WmaJR91f-ybFJl.png)
|
20 |
|
21 |
+
We are proud to announce, our new high quality flagship model series - ***PARM2***, Very high quality reasoning, math and coding abilities for a small size, that **anyone** can run on their device for free.
|
22 |
|
23 |
|
24 |
🧀 Which quant is right for you?
|
25 |
|
26 |
- ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable.
|
27 |
+
- ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its slightly slower than Q4.
|
28 |
|
29 |
This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c)
|
30 |
if you benchmarked this model let me know
|