Update README.md
Browse files
README.md
CHANGED
@@ -18,6 +18,25 @@ language:
|
|
18 |
- **License:** apache-2.0
|
19 |
- **Finetuned from model :** unsloth/qwen2.5-14b-bnb-4bit
|
20 |
|
21 |
-
This
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
|
23 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
|
|
18 |
- **License:** apache-2.0
|
19 |
- **Finetuned from model :** unsloth/qwen2.5-14b-bnb-4bit
|
20 |
|
21 |
+
This model is a fine-tuned version of **Qwen 2.5-14B**, trained on QwQ 32B Preview's responses to questions from the **NuminaMathCoT** dataset.
|
22 |
+
|
23 |
+
---
|
24 |
+
|
25 |
+
#### Training Details
|
26 |
+
|
27 |
+
- **Base Model**: Qwen 2.5-14B
|
28 |
+
- **Fine-Tuning Dataset**: Verified subset of **NuminaMathCoT** using Qwen 2.5 3B Instruct as a judge.
|
29 |
+
- **QLoRA Configuration**:
|
30 |
+
- **Rank**: 32
|
31 |
+
- **Rank Stabilization**: Enabled
|
32 |
+
- **Optimization Settings**:
|
33 |
+
- Batch Size: 8
|
34 |
+
- Gradient Accumulation Steps: 2 (Effective Batch Size: 16)
|
35 |
+
- Warm-Up Steps: 5
|
36 |
+
- Weight Decay: 0.01
|
37 |
+
- **Training Steps**: 500 steps
|
38 |
+
|
39 |
+
---
|
40 |
+
|
41 |
|
42 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|