--- base_model: Pinkstack/PARM-V1.5-QwQ-Qwen-2.5-o1-3B-VLLM tags: - text-generation-inference - transformers - unsloth - qwen2 - gguf - Reasoning - o1 - qwq license: apache-2.0 language: - en pipeline_tag: text-generation --- ![Pinkstack.png](https://cdn-uploads.huggingface.co/production/uploads/6710ba6af1279fe0dfe33afe/2xMulpuSlZ3C1vpGgsAYi.png) ⚠️ it may think it's name is Claude due to the training data, we are sorry for this issue but is shouldn't effect the quality of the responses. 🧀 Which quant is right for you? - ***Q4:*** This model should be used on edge devices like high end phones or laptops due to its very compact size, quality is okay but fully usable. - ***Q8:*** This model should be used on most high end modern devices like rtx 3080, Responses are very high quality, but its noticeably slower than q4 This Parm v2 is based on Qwen 2.5 3B which has gotten many extra reasoning training parameters so it would have similar outputs to qwen QwQ / O.1 mini (only much, smaller.). We've trained it using the datasets [here](https://huggingface.co/collections/Pinkstackorg/pram-v2-67612d3c542b9121bf15891c) This is a pretty heavy to run model if you want on device ai's for phones I'd recommend using the 0.5B version of this model (coming soon) To use this model, you must use a service which supports the GGUF file format. Additionaly, this is the Prompt Template: it uses the qwen2 template. ``` {{ if .System }}<|system|> {{ .System }}<|end|> {{ end }}{{ if .Prompt }}<|user|> {{ .Prompt }}<|end|> {{ end }}<|assistant|> {{ .Response }}<|end|> ``` Or if you are using an anti prompt: <|end|><|assistant|> Highly recommended to use with a system prompt. # Uploaded model - **Developed by:** Pinkstack - **License:** apache-2.0 - **Finetuned from model :** Pinkstack/PARM-V1.5-QwQ-Qwen-2.5-o1-3B-VLLM This ai model was trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.