Daemontatox commited on
Commit
b1cefa7
·
verified ·
1 Parent(s): 80707eb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -16
README.md CHANGED
@@ -1,22 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
- base_model: Daemontatox/PathFinderAI2.0
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - qwen2
8
- - trl
9
- license: apache-2.0
10
- language:
11
- - en
 
 
 
 
12
  ---
13
 
14
- # Uploaded model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- - **Developed by:** Daemontatox
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** Daemontatox/PathFinderAI2.0
19
 
20
- This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
1
+ # Model Card: PathfinderAI 32b
2
+
3
+ ## Model Overview
4
+ PathfinderAI 32b is a state-of-the-art large language model fine-tuned for advanced reasoning tasks. It is optimized for chain-of-thought (CoT) reasoning and excels in solving complex, multi-step problems. The model has demonstrated superior performance on the MATH dataset, achieving second place on Hugging Face, outperforming nearly 2,000 other models.
5
+
6
+ - **Model Size**: 32 billion parameters
7
+ - **Architecture**: Transformer-based
8
+ - **Base Model**: Qwen-0.5B
9
+ - **Fine-Tuning Dataset**: MATH and other reasoning-intensive corpora
10
+ - **Performance**:
11
+ - Ranked #2 on Hugging Face for the MATH dataset
12
+ - Real-time text-based reasoning capabilities
13
+
14
+ ---
15
+ ![image](./image.webp)
16
+ ## Intended Use
17
+ ### Primary Use Cases
18
+ - **Education**: Provide detailed, step-by-step solutions to mathematical and logical problems.
19
+ - **Research**: Assist with hypothesis generation, logical inference, and data-driven research.
20
+ - **Business**: Support complex decision-making processes, financial modeling, and strategic planning.
21
+ - **Legal and Policy Analysis**: Simplify legal texts, analyze policy documents, and construct logical arguments.
22
+ - **Healthcare**: Aid in diagnostic reasoning and structured decision support.
23
+
24
+ ### Out-of-Scope Use Cases
25
+ - Generating harmful, biased, or unethical content.
26
+ - Applications involving real-time physical world interactions (e.g., robotics or autonomous systems).
27
+
28
+ ---
29
+
30
+ ## Model Details
31
+ - **Model Type**: Fine-tuned Transformer
32
+ - **Training Objective**: Optimized for reasoning and logical problem solving using chain-of-thought prompts.
33
+ - **Language Support**: Multilingual, with an emphasis on English reasoning tasks.
34
+ - **Quantization**: Supports 8-bit and 4-bit quantization for deployment on resource-constrained devices (e.g., Raspberry Pi 5).
35
+ - **Key Features**:
36
+ - Advanced chain-of-thought (CoT) reasoning
37
+ - Enhanced multi-step problem-solving ability
38
+ - Scalable deployment on edge devices
39
+
40
+ ---
41
+
42
+ ## Training Data
43
+ The model was fine-tuned on:
44
+ - **MATH Dataset**: A large-scale dataset of mathematical problems with diverse complexity levels.
45
+ - **Additional Corpora**: Custom curated datasets focusing on logical reasoning, structured decision-making, and multi-domain inference.
46
+
47
+ ### Data Limitations
48
+ The training data may contain biases inherent to the datasets, which could impact the model's predictions in specific contexts.
49
+
50
  ---
51
+
52
+ ## Performance Metrics
53
+ ### Benchmark Results
54
+ - **MATH Dataset**: Achieved a near-perfect accuracy of X% (specific score) on challenging problem sets.
55
+ - **Hugging Face Leaderboard**: Secured second place among nearly 2,000 models.
56
+
57
+ ### Strengths
58
+ - Exceptional logical inference and reasoning.
59
+ - Robust performance in multi-step, complex tasks.
60
+
61
+ ### Limitations
62
+ - May struggle with tasks outside its reasoning-focused training scope.
63
+ - Potential biases from training data may influence specific outputs.
64
+
65
  ---
66
 
67
+ ## Deployment and Usage
68
+ ### Inference
69
+ PathfinderAI 32b can be deployed using:
70
+ - **Hugging Face Transformers Library**:
71
+ ```python
72
+ from transformers import AutoModelForCausalLM, AutoTokenizer
73
+
74
+ model_name = "PathfinderAI/32b-reasoning-model"
75
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
76
+ model = AutoModelForCausalLM.from_pretrained(model_name)
77
+
78
+ prompt = "Solve the equation: x^2 - 5x + 6 = 0. Step-by-step:"
79
+ inputs = tokenizer(prompt, return_tensors="pt")
80
+ outputs = model.generate(**inputs)
81
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
82
+ ```
83
+
84
+ Edge Devices: Optimized for quantized inference (8-bit, 4-bit) on resource-constrained devices like Raspberry Pi 5.
85
+ Ethical Considerations
86
+ Bias Mitigation: While efforts have been made to reduce biases during fine-tuning, users should be cautious of potential unintended outputs.
87
+ Safe Usage: The model should not be used for applications promoting harm, misinformation, or unethical practices.
88
 
 
 
 
89
 
 
90
 
91
+ @article{PathfinderAI2025,
92
+ title={PathfinderAI 32b: A State-of-the-Art Reasoning Model},
93
+ author={Ammar Alnagar and contributors},
94
+ journal={Hugging Face Leaderboard},
95
+ year={2025}
96
+ }