File size: 9,140 Bytes
93324c7
 
 
 
 
 
 
 
 
 
 
 
 
3efe24f
1621aff
f946e88
93324c7
f946e88
93324c7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5242d13
93324c7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
b0921d1
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
---
license: apache-2.0
base_model:
- nidum/Nidum-Llama-3.2-3B-Uncensored
- meta-llama/Llama-3.2-3B
library_name: adapter-transformers
tags:
- chemistry
- biology
- legal
- code
- medical
- finance
- roleplay
- uncensored
pipeline_tag: text-generation
---

### Nidum-Llama-3.2-3B-Uncensored  

### Welcome to Nidum!  
At Nidum, we believe in pushing the boundaries of innovation by providing advanced and unrestricted AI models for every application. Dive into our world of possibilities and experience the freedom of **Nidum-Llama-3.2-3B-Uncensored**, tailored to meet diverse needs with exceptional performance.

---

[![GitHub Icon](https://upload.wikimedia.org/wikipedia/commons/thumb/9/95/Font_Awesome_5_brands_github.svg/232px-Font_Awesome_5_brands_github.svg.png)](https://github.com/NidumAI-Inc)  
**Explore Nidum's Open-Source Projects on GitHub**: [https://github.com/NidumAI-Inc](https://github.com/NidumAI-Inc)

---
### Key Features

1. **Uncensored Responses**: Capable of addressing any query without content restrictions, offering detailed and uninhibited answers.
2. **Versatility**: Excels in diverse use cases, from complex technical queries to engaging casual conversations.
3. **Advanced Contextual Understanding**: Draws from an expansive knowledge base for accurate and context-aware outputs.
4. **Extended Context Handling**: Optimized for handling long-context interactions for improved continuity and depth.
5. **Customizability**: Adaptable to specific tasks and user preferences through fine-tuning.

---

### Use Cases

- **Open-Ended Q&A**  
- **Creative Writing and Ideation**  
- **Research Assistance**  
- **Educational Queries**  
- **Casual Conversations**  
- **Mathematical Problem Solving**  
- **Long-Context Dialogues**  

---

### How to Use

To start using **Nidum-Llama-3.2-3B-Uncensored**, follow the sample code below:

```python
import torch
from transformers import pipeline

pipe = pipeline(
    "text-generation",
    model="nidum/Nidum-Llama-3.2-3B-Uncensored",
    model_kwargs={"torch_dtype": torch.bfloat16},
    device="cuda",  # replace with "mps" to run on a Mac device
)

messages = [
    {"role": "user", "content": "Tell me something fascinating."},
]

outputs = pipe(messages, max_new_tokens=256)
assistant_response = outputs[0]["generated_text"][-1]["content"].strip()
print(assistant_response)
```

---
#### Quantized Models Available for Download

| **Quantized Model Version**                                                                                       | **Description**                                                                 |
|-------------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------|
| [**Nidum-Llama-3.2-3B-Uncensored-F16.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/Nidum-Llama-3.2-3B-Uncensored-F16.gguf) | Full 16-bit floating point precision for maximum accuracy on high-end GPUs.     |
| [**model-Q2_K.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q2_K.gguf)               | Optimized for minimal memory usage with lower precision, suitable for edge cases.|
| [**model-Q3_K_L.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_L.gguf)           | Balanced precision with enhanced memory efficiency for medium-range devices.    |
| [**model-Q3_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_M.gguf)           | Mid-range quantization for moderate precision and memory usage balance.         |
| [**model-Q3_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_S.gguf)           | Smaller quantization steps, offering moderate precision with reduced memory use.|
| [**model-Q4_0_4_4.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_4_4.gguf)       | Performance-optimized for low memory, ideal for lightweight deployment.         |
| [**model-Q4_0_4_8.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_4_8.gguf)       | Extended quantization balancing memory use and inference speed.                 |
| [**model-Q4_0_8_8.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_8_8.gguf)       | Advanced memory precision targeting larger contexts.                            |
| [**model-Q4_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_K_M.gguf)           | High-efficiency quantization for moderate GPU resources.                        |
| [**model-Q4_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_K_S.gguf)           | Optimized for smaller-scale operations with compact memory footprint.           |
| [**model-Q5_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q5_K_M.gguf)           | Balances performance and precision, ideal for robust inferencing environments.  |
| [**model-Q5_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q5_K_S.gguf)           | Moderate quantization targeting performance with minimal resource usage.        |
| [**model-Q6_K.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q6_K.gguf)               | High-precision quantization for accurate and stable inferencing tasks.          |
| [**model-TQ1_0.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-TQ1_0.gguf)             | Experimental quantization for targeted applications in test environments.       |
| [**model-TQ2_0.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-TQ2_0.gguf)             | High-performance tuning for experimental use cases and flexible precision.      |

---
### Datasets and Fine-Tuning

The following fine-tuning datasets are leveraged to enhance specific model capabilities:

- **Uncensored Data**: Enables unrestricted and uninhibited responses.
- **RAG-Based Fine-Tuning**: Optimizes retrieval-augmented generation for knowledge-intensive tasks.
- **Long Context Fine-Tuning**: Enhances the model's ability to process and maintain coherence in extended conversations.
- **Math-Instruct Data**: Specially curated for precise and contextually accurate mathematical reasoning.

---

### Benchmarks  

After fine-tuning with **uncensored data**, **Nidum-Llama-3.2-3B** demonstrates **superior performance compared to the original LLaMA model**, particularly in accuracy and handling diverse, unrestricted scenarios.

#### Benchmark Summary Table

| **Benchmark**    | **Metric**                       | **LLaMA 3.2 3B** | **Nidum 3.2 3B** | **Observation**                                                                                     |
|-------------------|-----------------------------------|--------------|--------------|-----------------------------------------------------------------------------------------------------|
| **GPQA**         | Exact Match (Flexible)           | 0.3          | 0.5          | Nidum 3B demonstrates significant improvement, particularly in **generative tasks**.                |
|                  | Accuracy                         | 0.4          | 0.5          | Consistent improvement, especially in **zero-shot** scenarios.                                      |
| **HellaSwag**    | Accuracy                         | 0.3          | 0.4          | Better performance in **common sense reasoning** tasks.                                             |
|                  | Normalized Accuracy              | 0.3          | 0.4          | Enhanced ability to understand and predict context in sentence completion.                          |
|                  | Normalized Accuracy (Stderr)     | 0.15275      | 0.1633       | Slightly improved consistency in normalized accuracy.                                               |
|                  | Accuracy (Stderr)                | 0.15275      | 0.1633       | Shows robustness in reasoning accuracy compared to LLaMA 3B.                                        |

---

### Insights:
1. **GPQA Results**: Fine-tuning on uncensored data has boosted **Nidum 3B's Exact Match and Accuracy**, particularly excelling in **generative** and **zero-shot** tasks involving domain-specific knowledge.
2. **HellaSwag Results**: **Nidum 3B** consistently outperforms **LLaMA 3B** in **common sense reasoning benchmarks**, indicating enhanced contextual and semantic understanding.

---

### Contributing

We welcome contributions to improve and extend the model’s capabilities. Stay tuned for updates on how to contribute.

---

### Contact

For inquiries, collaborations, or further information, please reach out to us at **info@nidum.ai**.

---

### Explore the Possibilities

Dive into unrestricted creativity and innovation with **Nidum Llama 3.2 3B Uncensored**!