File size: 4,404 Bytes
7347351 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 |
---
license: creativeml-openrail-m
datasets:
- mlabonne/lmsys-arena-human-preference-55k-sharegpt
language:
- en
base_model:
- meta-llama/Llama-3.2-3B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- Llama
- Llama-Cpp
- Llama3.2
- Instruct
- 3B
- bin
- Sentient
---
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
# QuantFactory/Llama-Sentient-3.2-3B-Instruct-GGUF
This is quantized version of [prithivMLmods/Llama-Sentient-3.2-3B-Instruct](https://huggingface.co/prithivMLmods/Llama-Sentient-3.2-3B-Instruct) created using llama.cpp
# Original Model Card
## Llama-Sentient-3.2-3B-Instruct Modelfile
| File Name | Size | Description | Upload Status |
|-----------------------------------------|--------------|-----------------------------------------|----------------|
| `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
| `README.md` | 42 Bytes | Initial commit README | Uploaded |
| `config.json` | 1.04 kB | Configuration file | Uploaded |
| `generation_config.json` | 248 Bytes | Generation configuration file | Uploaded |
| `pytorch_model-00001-of-00002.bin` | 4.97 GB | PyTorch model file (part 1) | Uploaded (LFS) |
| `pytorch_model-00002-of-00002.bin` | 1.46 GB | PyTorch model file (part 2) | Uploaded (LFS) |
| `pytorch_model.bin.index.json` | 21.2 kB | Model index file | Uploaded |
| `special_tokens_map.json` | 477 Bytes | Special tokens mapping | Uploaded |
| `tokenizer.json` | 17.2 MB | Tokenizer JSON file | Uploaded (LFS) |
| `tokenizer_config.json` | 57.4 kB | Tokenizer configuration file | Uploaded |
| Model Type | Size | Context Length | Link |
|------------|------|----------------|------|
| GGUF | 3B | - | [🤗 Llama-Sentient-3.2-3B-Instruct-GGUF](https://huggingface.co/prithivMLmods/Llama-Sentient-3.2-3B-Instruct-GGUF) |
The **Llama-Sentient-3.2-3B-Instruct** model is a fine-tuned version of the **Llama-3.2-3B-Instruct** model, optimized for **text generation** tasks, particularly where instruction-following abilities are critical. This model is trained on the **mlabonne/lmsys-arena-human-preference-55k-sharegpt** dataset, which enhances its performance in conversational and advisory contexts, making it suitable for a wide range of applications.
### Key Use Cases:
1. **Conversational AI**: Engage in intelligent dialogue, offering coherent responses and following instructions, useful for customer support and virtual assistants.
2. **Text Generation**: Generate high-quality, contextually appropriate content such as articles, summaries, explanations, and other forms of written communication based on user prompts.
3. **Instruction Following**: Follow specific instructions with accuracy, making it ideal for tasks that require structured guidance, such as technical troubleshooting or educational assistance.
The model uses a **PyTorch-based architecture** and includes a range of necessary files such as configuration files, tokenizer files, and model weight files for deployment.
### Intended Applications:
- **Chatbots** for virtual assistance, customer support, or as personal digital assistants.
- **Content Creation Tools**, aiding in the generation of written materials, blog posts, or automated responses based on user inputs.
- **Educational and Training Systems**, providing explanations and guided learning experiences in various domains.
- **Human-AI Interaction** platforms, where the model can follow user instructions to provide personalized assistance or perform specific tasks.
With its strong foundation in instruction-following and conversational contexts, the **Llama-Sentient-3.2-3B-Instruct** model offers versatile applications for both general and specialized domains.
|