weezywitasneezy commited on
Commit
febc3d9
1 Parent(s): 981a0ca

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -13
README.md CHANGED
@@ -120,6 +120,24 @@ Oxide-F1-7B-slerp is a merge of the following models using [LazyMergekit](https:
120
  * [weezywitasneezy/BenchmarkEngineering-7B-slerp](https://huggingface.co/weezywitasneezy/BenchmarkEngineering-7B-slerp)
121
  * [weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp](https://huggingface.co/weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp)
122
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
  ## 🧩 Configuration
124
 
125
  ```yaml
@@ -165,16 +183,3 @@ pipeline = transformers.pipeline(
165
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
166
  print(outputs[0]["generated_text"])
167
  ```
168
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
169
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__Oxide-F1-7B-slerp)
170
-
171
- | Metric |Value|
172
- |---------------------------------|----:|
173
- |Avg. |73.64|
174
- |AI2 Reasoning Challenge (25-Shot)|69.71|
175
- |HellaSwag (10-Shot) |87.82|
176
- |MMLU (5-Shot) |64.77|
177
- |TruthfulQA (0-shot) |70.00|
178
- |Winogrande (5-shot) |81.85|
179
- |GSM8k (5-shot) |67.70|
180
-
 
120
  * [weezywitasneezy/BenchmarkEngineering-7B-slerp](https://huggingface.co/weezywitasneezy/BenchmarkEngineering-7B-slerp)
121
  * [weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp](https://huggingface.co/weezywitasneezy/OxytocinErosEngineeringFX-7B-slerp)
122
 
123
+
124
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
125
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__Oxide-F1-7B-slerp)
126
+
127
+
128
+ | Metric |Value|
129
+ |---------------------------------|----:|
130
+ |Avg. |73.64|
131
+ |AI2 Reasoning Challenge (25-Shot)|69.71|
132
+ |HellaSwag (10-Shot) |87.82|
133
+ |MMLU (5-Shot) |64.77|
134
+ |TruthfulQA (0-shot) |70.00|
135
+ |Winogrande (5-shot) |81.85|
136
+ |GSM8k (5-shot) |67.70|
137
+
138
+
139
+
140
+
141
  ## 🧩 Configuration
142
 
143
  ```yaml
 
183
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
184
  print(outputs[0]["generated_text"])
185
  ```