aashish1904 commited on
Commit
9a11936
·
verified ·
1 Parent(s): becd952

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +189 -0
README.md ADDED
@@ -0,0 +1,189 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ base_model:
9
+ - ifable/gemma-2-Ifable-9B
10
+ - jsgreenawalt/gemma-2-9B-it-advanced-v2.1
11
+ model-index:
12
+ - name: Gemma-2-Ataraxy-v2-9B
13
+ results:
14
+ - task:
15
+ type: text-generation
16
+ name: Text Generation
17
+ dataset:
18
+ name: IFEval (0-Shot)
19
+ type: HuggingFaceH4/ifeval
20
+ args:
21
+ num_few_shot: 0
22
+ metrics:
23
+ - type: inst_level_strict_acc and prompt_level_strict_acc
24
+ value: 21.36
25
+ name: strict accuracy
26
+ source:
27
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
28
+ name: Open LLM Leaderboard
29
+ - task:
30
+ type: text-generation
31
+ name: Text Generation
32
+ dataset:
33
+ name: BBH (3-Shot)
34
+ type: BBH
35
+ args:
36
+ num_few_shot: 3
37
+ metrics:
38
+ - type: acc_norm
39
+ value: 39.8
40
+ name: normalized accuracy
41
+ source:
42
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
43
+ name: Open LLM Leaderboard
44
+ - task:
45
+ type: text-generation
46
+ name: Text Generation
47
+ dataset:
48
+ name: MATH Lvl 5 (4-Shot)
49
+ type: hendrycks/competition_math
50
+ args:
51
+ num_few_shot: 4
52
+ metrics:
53
+ - type: exact_match
54
+ value: 0.83
55
+ name: exact match
56
+ source:
57
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
58
+ name: Open LLM Leaderboard
59
+ - task:
60
+ type: text-generation
61
+ name: Text Generation
62
+ dataset:
63
+ name: GPQA (0-shot)
64
+ type: Idavidrein/gpqa
65
+ args:
66
+ num_few_shot: 0
67
+ metrics:
68
+ - type: acc_norm
69
+ value: 12.3
70
+ name: acc_norm
71
+ source:
72
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
73
+ name: Open LLM Leaderboard
74
+ - task:
75
+ type: text-generation
76
+ name: Text Generation
77
+ dataset:
78
+ name: MuSR (0-shot)
79
+ type: TAUR-Lab/MuSR
80
+ args:
81
+ num_few_shot: 0
82
+ metrics:
83
+ - type: acc_norm
84
+ value: 4.88
85
+ name: acc_norm
86
+ source:
87
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
88
+ name: Open LLM Leaderboard
89
+ - task:
90
+ type: text-generation
91
+ name: Text Generation
92
+ dataset:
93
+ name: MMLU-PRO (5-shot)
94
+ type: TIGER-Lab/MMLU-Pro
95
+ config: main
96
+ split: test
97
+ args:
98
+ num_few_shot: 5
99
+ metrics:
100
+ - type: acc
101
+ value: 35.79
102
+ name: accuracy
103
+ source:
104
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=lemon07r/Gemma-2-Ataraxy-v2-9B
105
+ name: Open LLM Leaderboard
106
+
107
+ ---
108
+
109
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
110
+
111
+
112
+ # QuantFactory/Gemma-2-Ataraxy-v2-9B-GGUF
113
+ This is quantized version of [lemon07r/Gemma-2-Ataraxy-v2-9B](https://huggingface.co/lemon07r/Gemma-2-Ataraxy-v2-9B) created using llama.cpp
114
+
115
+ # Original Model Card
116
+
117
+ # Gemma 2 Ataraxy v2 9B
118
+
119
+ Finally, after much testing, a sucessor to the first Gemma 2 Ataraxy 9B. Same kind of recipe, using the same principles, same concept as the last Ataraxy but using better models this time.
120
+
121
+ ![Ataraxy](https://i.imgur.com/P2F9XN9.png)
122
+
123
+ ## GGUF / EXL2 Quants
124
+
125
+ Bartowski quants (imatrix): https://huggingface.co/bartowski/Gemma-2-Ataraxy-v2-9B-GGUF
126
+
127
+ Mradermacher quants (static): https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v2-9B-GGUF
128
+
129
+ Mradermacher quants (imatrix): https://huggingface.co/mradermacher/Gemma-2-Ataraxy-v2-9B-i1-GGUF
130
+
131
+ Bartowski and mradermacher use different calibration data for their imatrix quants I believe, and the static quant of course uses none. Pick your poison.
132
+
133
+ More coming soon.
134
+
135
+ ## Format
136
+
137
+ Use Gemma 2 format.
138
+
139
+ ## Merge Details
140
+ ### Merge Method
141
+
142
+ This model was merged using the SLERP merge method.
143
+
144
+ ### Models Merged
145
+
146
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
147
+
148
+ The following models were included in the merge:
149
+ * [ifable/gemma-2-Ifable-9B](https://huggingface.co/ifable/gemma-2-Ifable-9B)
150
+ * [jsgreenawalt/gemma-2-9B-it-advanced-v2.1](https://huggingface.co/jsgreenawalt/gemma-2-9B-it-advanced-v2.1)
151
+
152
+ ### Configuration
153
+
154
+ The following YAML configuration was used to produce this model:
155
+
156
+ ```yaml
157
+ base_model: ifable/gemma-2-Ifable-9B
158
+ dtype: bfloat16
159
+ merge_method: slerp
160
+ parameters:
161
+ t:
162
+ - filter: self_attn
163
+ value: [0.0, 0.5, 0.3, 0.7, 1.0]
164
+ - filter: mlp
165
+ value: [1.0, 0.5, 0.7, 0.3, 0.0]
166
+ - value: 0.5
167
+ slices:
168
+ - sources:
169
+ - layer_range: [0, 42]
170
+ model: jsgreenawalt/gemma-2-9B-it-advanced-v2.1
171
+ - layer_range: [0, 42]
172
+ model: ifable/gemma-2-Ifable-9B
173
+ ```
174
+
175
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
176
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_lemon07r__Gemma-2-Ataraxy-v2-9B)
177
+
178
+ | Metric |Value|
179
+ |-------------------|----:|
180
+ |Avg. |19.16|
181
+ |IFEval (0-Shot) |21.36|
182
+ |BBH (3-Shot) |39.80|
183
+ |MATH Lvl 5 (4-Shot)| 0.83|
184
+ |GPQA (0-shot) |12.30|
185
+ |MuSR (0-shot) | 4.88|
186
+ |MMLU-PRO (5-shot) |35.79|
187
+
188
+ Second highest ranked open weight model in EQ-Bench.
189
+