File size: 1,516 Bytes
4ebfc57 ba827fd 4ebfc57 ba827fd 4ebfc57 9071207 7647425 1abd9c7 7647425 2cf8706 dc9e26b cf1272e 4ebfc57 4582c43 4ebfc57 6d2e67b e229982 cf1272e |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
---
license: cc-by-nc-nd-4.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
tags:
- Mistral
- Pygmalion
- llama-2
- llama-2-7b
---
# MistralPy-7b
This is a merger focusing on preserving the roleplay abilities of Pygmalion while gaining the improved results from Mistral. This model works best for roleplay but is still fairly capable assistant. The smaller (7b) size does mean it isn't perfect at more complex reasoning tasks, but this should be addressed in the larger version that I'll upload soon (when I can get Mistral to play along).
[GGUF version done by TheBloke](https://huggingface.co/TheBloke/Mistral-Pygmalion-7B-GGUF)
# LLM Leaderboard Evaluation
| Metric | Value |
|-----------------------|---------------------------|
| Avg. | 44.58 |
| ARC (25-shot) | 54.44 |
| HellaSwag (10-shot) | 78.48 |
| MMLU (5-shot) | 49.23 |
| TruthfulQA (0-shot) | 41.82 |
| Winogrande (5-shot) | 75.3 |
| GSM8K (5-shot) | 6.82 |
| DROP (3-shot) | 5.94 |
### Prompt Template
```
### Instruction:
{Prompt & Backstory}
### Assistant:
{Output}
```
Example:
```
### Instruction:
You are Sally, a fun 19 year old woman. Her favorite animal is "cat". Her favoritate color is "blue". She enjoys grape juice and cake.
### Assistant:
Sally: Hi, how are you?
User: Okay, you?
```
# Send a message
[Steam](https://steamcommunity.com/id/delcos/)
#### Discord: delcos69 |