File size: 1,516 Bytes
4ebfc57
 
 
 
 
 
 
 
ba827fd
4ebfc57
 
 
 
 
ba827fd
4ebfc57
9071207
7647425
1abd9c7
7647425
2cf8706
 
dc9e26b
cf1272e
 
 
 
 
 
 
 
 
 
 
 
 
4ebfc57
 
 
 
 
 
 
4582c43
 
 
 
 
 
 
 
 
4ebfc57
6d2e67b
e229982
cf1272e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
---
license: cc-by-nc-nd-4.0
language:
- en
library_name: transformers
pipeline_tag: text-generation
tags:
- Mistral
- Pygmalion
- llama-2
- llama-2-7b
---


# MistralPy-7b

This is a merger focusing on preserving the roleplay abilities of Pygmalion while gaining the improved results from Mistral. This model works best for roleplay but is still fairly capable assistant. The smaller (7b) size does mean it isn't perfect at more complex reasoning tasks, but this should be addressed in the larger version that I'll upload soon (when I can get Mistral to play along).

[GGUF version done by TheBloke](https://huggingface.co/TheBloke/Mistral-Pygmalion-7B-GGUF)



# LLM Leaderboard Evaluation

| Metric                | Value                     |
|-----------------------|---------------------------|
| Avg.                  | 44.58   |
| ARC (25-shot)         | 54.44          |
| HellaSwag (10-shot)   | 78.48    |
| MMLU (5-shot)         | 49.23         |
| TruthfulQA (0-shot)   | 41.82   |
| Winogrande (5-shot)   | 75.3   |
| GSM8K (5-shot)        | 6.82        |
| DROP (3-shot)         | 5.94         |


### Prompt Template
```
### Instruction:
{Prompt & Backstory}
### Assistant:
{Output}
```
Example:

```
### Instruction:
You are Sally, a fun 19 year old woman. Her favorite animal is "cat". Her favoritate color is "blue". She enjoys grape juice and cake.
### Assistant:
Sally: Hi, how are you?
User: Okay, you?
```

# Send a message
[Steam](https://steamcommunity.com/id/delcos/)
#### Discord: delcos69