File size: 3,999 Bytes
7999b89
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0b3dd0f
 
 
 
7999b89
8aacc8b
7999b89
 
 
 
 
 
 
8d0e843
7999b89
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
---
base_model: meta-llama/Meta-Llama-3-8B-Instruct
inference: false
pipeline_tag: text-generation
language:
- en
license: other
license_name: llama3
license_link: https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/LICENSE
model_creator: meta-llama
model_name: Meta-Llama-3-8B-Instruct
model_type: llama
tags:
- facebook
- meta
- pytorch
- llama
- llama-3
quantized_by: brittlewis12

---

# Meta-Llama-3-8B-Instruct GGUF

**Original model**: [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct)

**Model creator**: [Meta](https://huggingface.co/meta-llama)

> Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Further, in developing these models, we took great care to optimize helpfulness and safety.

> Llama 3 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align with human preferences for helpfulness and safety.

This repo contains GGUF format model files for Meta’s Llama-3-8B-Instruct,
**updated as of 2024-04-29** to incorporate [tokenization improvements](https://github.com/ggerganov/llama.cpp/pull/6920), as well as previous interventions to handle the `<|eot_id|>` special token as EOS token.

Learn more on Meta’s [Llama 3 page](https://llama.meta.com/llama3).

### What is GGUF?

GGUF is a file format for representing AI models. It is the third version of the format, 
introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp. 
Converted with llama.cpp build 2763 (revision [ffe666](https://github.com/ggerganov/llama.cpp/commits/ffe666572f98a686b17a2cd1dbf4c0a982e5ac0a)),
using [autogguf](https://github.com/brittlewis12/autogguf).

### Prompt template

```
<|start_header_id|>system<|end_header_id|>

{{system_prompt}}<|eot_id|><|start_header_id|>user<|end_header_id|>

{{prompt}}<|eot_id|><|start_header_id|>assistant<|end_header_id|>


```

---

## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!

![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)

[cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
- create & save **Characters** with custom system prompts & temperature settings
- download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
- make it your own with custom **Theme colors**
- powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
- **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
- follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date

---

## Original Model Evaluation

<table>
  <tr>
   <td><strong>Benchmark</strong>
   </td>
   <td><strong>Llama 3 8B</strong>
   </td>
   <td><strong>Llama 2 7B</strong>
   </td>
   <td><strong>Llama 2 13B</strong>
   </td>
  </tr>
  <tr>
   <td>MMLU (5-shot)
   </td>
   <td><b>68.4</b>
   </td>
   <td>34.1
   </td>
   <td>47.8
   </td>
  </tr>
  <tr>
   <td>GPQA (0-shot)
   </td>
   <td><b>34.2</b>
   </td>
   <td>21.7
   </td>
   <td>22.3
   </td>
  </tr>
  <tr>
   <td>HumanEval (0-shot)
   </td>
   <td><b>62.2</b>
   </td>
   <td>7.9
   </td>
   <td>14.0
   </td>
  </tr>
  <tr>
   <td>GSM-8K (8-shot, CoT)
   </td>
   <td><b>79.6</b>
   </td>
   <td>25.7
   </td>
   <td>77.4
   </td>
  </tr>
  <tr>
   <td>MATH (4-shot, CoT)
   </td>
   <td><b>30.0</b>
   </td>
   <td>3.8
   </td>
   <td>6.7
   </td>
  </tr>
</table>