File size: 2,848 Bytes
0ebf784
 
35aa8b2
 
 
 
 
02789ee
 
be13597
08ba578
31fbd44
7f1f13d
 
 
08ba578
63488b7
 
 
 
 
 
08ba578
33b1b37
6a21ffd
921dd1a
6a21ffd
 
08ba578
8611d21
 
5642849
428542b
bbe25c0
169a850
 
 
 
 
 
428542b
7f1f13d
 
 
 
 
 
 
 
 
 
 
31fbd44
63488b7
08ba578
 
 
 
 
 
 
 
 
 
 
 
 
 
 
be13597
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: cc-by-nc-4.0
datasets:
- meta-math/MetaMathQA
language:
- en
pipeline_tag: text-generation
tags:
- Math
- exl2
---
# Merged-AGI-7B

- Model creator: [Talha Rüzgar Akkuş](https://huggingface.co/Q-bert)
- Original model: [Merged AGI 7B](https://huggingface.co/Q-bert/Merged-AGI-7B)

## EXL2 Quants

- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw)
- [4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main)
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw)
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw)

Zipped Quantization (if you want to download a single file)
- [3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw-zip)
- [4.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/4.0bpw-zip)
- [6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw-zip)
- [8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw-zip)

## Calibration Dataset
[wikitext-103-v1](https://huggingface.co/datasets/wikitext/blob/9a9e482b5987f9d25b3a9b2883fc6cc9fd8071b3/wikitext-103-v1/wikitext-test.parquet)

## Memory Usage
Measured using ExLlamav2_HF and 4096 max_seq_len with [Oobabooga's Text Generation WebUI](https://github.com/oobabooga/text-generation-webui/tree/main).
| Branch | BPW | VRAM Usage | Description |
| ------ | --- | ---------- | ----------- |
[3.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/3.0bpw)|3.0|3.7 GB|For >=6GB VRAM cards
[4.0bpw (main)](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/main)|4.0|4.4 GB|For >=6GB VRAM cards
[6.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/6.0bpw)|6.0|6.1 GB|For >=8GB VRAM cards
[8.0bpw](https://huggingface.co/hgloow/Merged-AGI-7B-EXL2/tree/8.0bpw)|8.0|7.7 GB|For >=10GB VRAM cards

## Prompt template: ChatML

```
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant

```

# Original Info

Merge [Q-bert/MetaMath-Cybertron-Starling](https://huggingface.co/Q-bert/MetaMath-Cybertron-Starling) and [fblgit/juanako-7b-UNA](https://huggingface.co/fblgit/juanako-7b-UNA) using slerp merge.

You can use ChatML format.

# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [Coming soon]()

| Metric                | Value                     |
|-----------------------|---------------------------|
| Avg.                  | Coming soon               |
| ARC (25-shot)         | Coming soon               |
| HellaSwag (10-shot)   | Coming soon               |
| MMLU (5-shot)         | Coming soon               |
| TruthfulQA (0-shot)   | Coming soon               |
| Winogrande (5-shot)   | Coming soon               |
| GSM8K (5-shot)        | Coming soon               |