barthfab commited on
Commit
217b779
1 Parent(s): 10bf4d8

Delete jpacifico/French-Alpaca-Llama3-8B-Instruct-v1.0

Browse files
jpacifico/French-Alpaca-Llama3-8B-Instruct-v1.0/results_2024_06_03T12-24-56.json DELETED
@@ -1,570 +0,0 @@
1
- {
2
- "config_general": {
3
- "lighteval_sha": "",
4
- "num_few_shot_default": 0,
5
- "num_fewshot_seeds": 1,
6
- "override_batch_size": "auto:6",
7
- "max_samples": "null",
8
- "job_id": "",
9
- "model_name": "jpacifico/French-Alpaca-Llama3-8B-Instruct-v1.0",
10
- "model_sha": "",
11
- "model_dtype": "torch.bfloat16",
12
- "model_size": ""
13
- },
14
- "results": {
15
- "harness|hellaswag_fr|10": {
16
- "acc,none": 0.5063182694367102,
17
- "acc_stderr,none": 0.005174062227988366,
18
- "acc_norm,none": 0.67551938316556,
19
- "acc_norm_stderr,none": 0.004845176560636152,
20
- "alias": "hellaswag_fr"
21
- },
22
- "harness|arc_challenge_m_it|25": {
23
- "acc,none": 0.5089820359281437,
24
- "acc_stderr,none": 0.014627782577773944,
25
- "acc_norm,none": 0.5397775876817793,
26
- "acc_norm_stderr,none": 0.01458377263416605,
27
- "alias": "arc_challenge_m_it"
28
- },
29
- "harness|gsm8k|5": {
30
- "exact_match,get-answer": 0.66565579984837,
31
- "exact_match_stderr,get-answer": 0.012994634003332763,
32
- "alias": "gsm8k"
33
- },
34
- "harness|truthfulqa_mc2|0": {
35
- "acc,none": 0.49419688087237895,
36
- "acc_stderr,none": 0.014691972386008712,
37
- "alias": "truthfulqa_mc2"
38
- },
39
- "harness|arc_challenge_m_fr|25": {
40
- "acc,none": 0.49871685201026517,
41
- "acc_stderr,none": 0.01463009522328876,
42
- "acc_norm,none": 0.5431993156544055,
43
- "acc_norm_stderr,none": 0.014575436127048211,
44
- "alias": "arc_challenge_m_fr"
45
- },
46
- "harness|mmlu_m_fr|5": {
47
- "acc,none": 0.5703154839202506,
48
- "acc_stderr,none": 0.004326758046899862,
49
- "alias": "mmlu_m_fr"
50
- },
51
- "harness|truthfulqa_mc2_m_es|0": {
52
- "acc,none": 0.2674271229404309,
53
- "acc_stderr,none": 0.01576757970059163,
54
- "alias": "truthfulqa_mc2_m_es"
55
- },
56
- "harness|belebele_eng_Latn|5": {
57
- "acc,none": 0.9055555555555556,
58
- "acc_stderr,none": 0.009753621805733555,
59
- "acc_norm,none": 0.9055555555555556,
60
- "acc_norm_stderr,none": 0.009753621805733555,
61
- "alias": "belebele_eng_Latn"
62
- },
63
- "harness|mmlu_m_es|5": {
64
- "acc,none": 0.5768711564421779,
65
- "acc_stderr,none": 0.004278699519688541,
66
- "alias": "mmlu_m_es"
67
- },
68
- "harness|hellaswag_it|10": {
69
- "acc,none": 0.4840639617099967,
70
- "acc_stderr,none": 0.005212478778642557,
71
- "acc_norm,none": 0.6560426411399978,
72
- "acc_norm_stderr,none": 0.004954653970651795,
73
- "alias": "hellaswag_it"
74
- },
75
- "harness|belebele_fra_Latn|5": {
76
- "acc,none": 0.8477777777777777,
77
- "acc_stderr,none": 0.011981196673569642,
78
- "acc_norm,none": 0.8477777777777777,
79
- "acc_norm_stderr,none": 0.011981196673569642,
80
- "alias": "belebele_fra_Latn"
81
- },
82
- "harness|hellaswag_de|10": {
83
- "acc,none": 0.4738471391972673,
84
- "acc_stderr,none": 0.005159110613278454,
85
- "acc_norm,none": 0.6286293766011956,
86
- "acc_norm_stderr,none": 0.004992302000579806,
87
- "alias": "hellaswag_de"
88
- },
89
- "harness|arc_challenge_m_de|25": {
90
- "acc,none": 0.4739093242087254,
91
- "acc_stderr,none": 0.014610211661428537,
92
- "acc_norm,none": 0.5183917878528657,
93
- "acc_norm_stderr,none": 0.014620242527326811,
94
- "alias": "arc_challenge_m_de"
95
- },
96
- "harness|mmlu_m_it|5": {
97
- "acc,none": 0.5661403641308453,
98
- "acc_stderr,none": 0.004307827500632795,
99
- "alias": "mmlu_m_it"
100
- },
101
- "harness|hellaswag|10": {
102
- "acc,none": 0.6217884883489345,
103
- "acc_stderr,none": 0.004839497020536618,
104
- "acc_norm,none": 0.8193586934873531,
105
- "acc_norm_stderr,none": 0.003839344497192083,
106
- "alias": "hellaswag"
107
- },
108
- "harness|arc_challenge|25": {
109
- "acc,none": 0.5947098976109215,
110
- "acc_stderr,none": 0.014346869060229318,
111
- "acc_norm,none": 0.6407849829351536,
112
- "acc_norm_stderr,none": 0.014020224155839157,
113
- "alias": "arc_challenge"
114
- },
115
- "harness|arc_challenge_m_es|25": {
116
- "acc,none": 0.5153846153846153,
117
- "acc_stderr,none": 0.01461696032622132,
118
- "acc_norm,none": 0.5547008547008547,
119
- "acc_norm_stderr,none": 0.014536106383401226,
120
- "alias": "arc_challenge_m_es"
121
- },
122
- "harness|hendrycksTest|5": {
123
- "acc,none": 0.6615866685657313,
124
- "acc_stderr,none": 0.12826541916322984,
125
- "alias": "mmlu"
126
- },
127
- "harness|hendrycksTest-humanities|5": {
128
- "acc,none": 0.6615866685657313,
129
- "acc_stderr,none": 0.12826541916322984,
130
- "alias": "mmlu"
131
- },
132
- "harness|hendrycksTest-formal_logic|5": {
133
- "acc,none": 0.6615866685657313,
134
- "acc_stderr,none": 0.12826541916322984,
135
- "alias": "mmlu"
136
- },
137
- "harness|hendrycksTest-high_school_european_history|5": {
138
- "acc,none": 0.6615866685657313,
139
- "acc_stderr,none": 0.12826541916322984,
140
- "alias": "mmlu"
141
- },
142
- "harness|hendrycksTest-high_school_us_history|5": {
143
- "acc,none": 0.6615866685657313,
144
- "acc_stderr,none": 0.12826541916322984,
145
- "alias": "mmlu"
146
- },
147
- "harness|hendrycksTest-high_school_world_history|5": {
148
- "acc,none": 0.6615866685657313,
149
- "acc_stderr,none": 0.12826541916322984,
150
- "alias": "mmlu"
151
- },
152
- "harness|hendrycksTest-international_law|5": {
153
- "acc,none": 0.6615866685657313,
154
- "acc_stderr,none": 0.12826541916322984,
155
- "alias": "mmlu"
156
- },
157
- "harness|hendrycksTest-jurisprudence|5": {
158
- "acc,none": 0.6615866685657313,
159
- "acc_stderr,none": 0.12826541916322984,
160
- "alias": "mmlu"
161
- },
162
- "harness|hendrycksTest-logical_fallacies|5": {
163
- "acc,none": 0.6615866685657313,
164
- "acc_stderr,none": 0.12826541916322984,
165
- "alias": "mmlu"
166
- },
167
- "harness|hendrycksTest-moral_disputes|5": {
168
- "acc,none": 0.6615866685657313,
169
- "acc_stderr,none": 0.12826541916322984,
170
- "alias": "mmlu"
171
- },
172
- "harness|hendrycksTest-moral_scenarios|5": {
173
- "acc,none": 0.6615866685657313,
174
- "acc_stderr,none": 0.12826541916322984,
175
- "alias": "mmlu"
176
- },
177
- "harness|hendrycksTest-philosophy|5": {
178
- "acc,none": 0.6615866685657313,
179
- "acc_stderr,none": 0.12826541916322984,
180
- "alias": "mmlu"
181
- },
182
- "harness|hendrycksTest-prehistory|5": {
183
- "acc,none": 0.6615866685657313,
184
- "acc_stderr,none": 0.12826541916322984,
185
- "alias": "mmlu"
186
- },
187
- "harness|hendrycksTest-professional_law|5": {
188
- "acc,none": 0.6615866685657313,
189
- "acc_stderr,none": 0.12826541916322984,
190
- "alias": "mmlu"
191
- },
192
- "harness|hendrycksTest-world_religions|5": {
193
- "acc,none": 0.6615866685657313,
194
- "acc_stderr,none": 0.12826541916322984,
195
- "alias": "mmlu"
196
- },
197
- "harness|hendrycksTest-other|5": {
198
- "acc,none": 0.6615866685657313,
199
- "acc_stderr,none": 0.12826541916322984,
200
- "alias": "mmlu"
201
- },
202
- "harness|hendrycksTest-business_ethics|5": {
203
- "acc,none": 0.6615866685657313,
204
- "acc_stderr,none": 0.12826541916322984,
205
- "alias": "mmlu"
206
- },
207
- "harness|hendrycksTest-clinical_knowledge|5": {
208
- "acc,none": 0.6615866685657313,
209
- "acc_stderr,none": 0.12826541916322984,
210
- "alias": "mmlu"
211
- },
212
- "harness|hendrycksTest-college_medicine|5": {
213
- "acc,none": 0.6615866685657313,
214
- "acc_stderr,none": 0.12826541916322984,
215
- "alias": "mmlu"
216
- },
217
- "harness|hendrycksTest-global_facts|5": {
218
- "acc,none": 0.6615866685657313,
219
- "acc_stderr,none": 0.12826541916322984,
220
- "alias": "mmlu"
221
- },
222
- "harness|hendrycksTest-human_aging|5": {
223
- "acc,none": 0.6615866685657313,
224
- "acc_stderr,none": 0.12826541916322984,
225
- "alias": "mmlu"
226
- },
227
- "harness|hendrycksTest-management|5": {
228
- "acc,none": 0.6615866685657313,
229
- "acc_stderr,none": 0.12826541916322984,
230
- "alias": "mmlu"
231
- },
232
- "harness|hendrycksTest-marketing|5": {
233
- "acc,none": 0.6615866685657313,
234
- "acc_stderr,none": 0.12826541916322984,
235
- "alias": "mmlu"
236
- },
237
- "harness|hendrycksTest-medical_genetics|5": {
238
- "acc,none": 0.6615866685657313,
239
- "acc_stderr,none": 0.12826541916322984,
240
- "alias": "mmlu"
241
- },
242
- "harness|hendrycksTest-miscellaneous|5": {
243
- "acc,none": 0.6615866685657313,
244
- "acc_stderr,none": 0.12826541916322984,
245
- "alias": "mmlu"
246
- },
247
- "harness|hendrycksTest-nutrition|5": {
248
- "acc,none": 0.6615866685657313,
249
- "acc_stderr,none": 0.12826541916322984,
250
- "alias": "mmlu"
251
- },
252
- "harness|hendrycksTest-professional_accounting|5": {
253
- "acc,none": 0.6615866685657313,
254
- "acc_stderr,none": 0.12826541916322984,
255
- "alias": "mmlu"
256
- },
257
- "harness|hendrycksTest-professional_medicine|5": {
258
- "acc,none": 0.6615866685657313,
259
- "acc_stderr,none": 0.12826541916322984,
260
- "alias": "mmlu"
261
- },
262
- "harness|hendrycksTest-virology|5": {
263
- "acc,none": 0.6615866685657313,
264
- "acc_stderr,none": 0.12826541916322984,
265
- "alias": "mmlu"
266
- },
267
- "harness|hendrycksTest-social_sciences|5": {
268
- "acc,none": 0.6615866685657313,
269
- "acc_stderr,none": 0.12826541916322984,
270
- "alias": "mmlu"
271
- },
272
- "harness|hendrycksTest-econometrics|5": {
273
- "acc,none": 0.6615866685657313,
274
- "acc_stderr,none": 0.12826541916322984,
275
- "alias": "mmlu"
276
- },
277
- "harness|hendrycksTest-high_school_geography|5": {
278
- "acc,none": 0.6615866685657313,
279
- "acc_stderr,none": 0.12826541916322984,
280
- "alias": "mmlu"
281
- },
282
- "harness|hendrycksTest-high_school_government_and_politics|5": {
283
- "acc,none": 0.6615866685657313,
284
- "acc_stderr,none": 0.12826541916322984,
285
- "alias": "mmlu"
286
- },
287
- "harness|hendrycksTest-high_school_macroeconomics|5": {
288
- "acc,none": 0.6615866685657313,
289
- "acc_stderr,none": 0.12826541916322984,
290
- "alias": "mmlu"
291
- },
292
- "harness|hendrycksTest-high_school_microeconomics|5": {
293
- "acc,none": 0.6615866685657313,
294
- "acc_stderr,none": 0.12826541916322984,
295
- "alias": "mmlu"
296
- },
297
- "harness|hendrycksTest-high_school_psychology|5": {
298
- "acc,none": 0.6615866685657313,
299
- "acc_stderr,none": 0.12826541916322984,
300
- "alias": "mmlu"
301
- },
302
- "harness|hendrycksTest-human_sexuality|5": {
303
- "acc,none": 0.6615866685657313,
304
- "acc_stderr,none": 0.12826541916322984,
305
- "alias": "mmlu"
306
- },
307
- "harness|hendrycksTest-professional_psychology|5": {
308
- "acc,none": 0.6615866685657313,
309
- "acc_stderr,none": 0.12826541916322984,
310
- "alias": "mmlu"
311
- },
312
- "harness|hendrycksTest-public_relations|5": {
313
- "acc,none": 0.6615866685657313,
314
- "acc_stderr,none": 0.12826541916322984,
315
- "alias": "mmlu"
316
- },
317
- "harness|hendrycksTest-security_studies|5": {
318
- "acc,none": 0.6615866685657313,
319
- "acc_stderr,none": 0.12826541916322984,
320
- "alias": "mmlu"
321
- },
322
- "harness|hendrycksTest-sociology|5": {
323
- "acc,none": 0.6615866685657313,
324
- "acc_stderr,none": 0.12826541916322984,
325
- "alias": "mmlu"
326
- },
327
- "harness|hendrycksTest-us_foreign_policy|5": {
328
- "acc,none": 0.6615866685657313,
329
- "acc_stderr,none": 0.12826541916322984,
330
- "alias": "mmlu"
331
- },
332
- "harness|hendrycksTest-stem|5": {
333
- "acc,none": 0.6615866685657313,
334
- "acc_stderr,none": 0.12826541916322984,
335
- "alias": "mmlu"
336
- },
337
- "harness|hendrycksTest-abstract_algebra|5": {
338
- "acc,none": 0.6615866685657313,
339
- "acc_stderr,none": 0.12826541916322984,
340
- "alias": "mmlu"
341
- },
342
- "harness|hendrycksTest-anatomy|5": {
343
- "acc,none": 0.6615866685657313,
344
- "acc_stderr,none": 0.12826541916322984,
345
- "alias": "mmlu"
346
- },
347
- "harness|hendrycksTest-astronomy|5": {
348
- "acc,none": 0.6615866685657313,
349
- "acc_stderr,none": 0.12826541916322984,
350
- "alias": "mmlu"
351
- },
352
- "harness|hendrycksTest-college_biology|5": {
353
- "acc,none": 0.6615866685657313,
354
- "acc_stderr,none": 0.12826541916322984,
355
- "alias": "mmlu"
356
- },
357
- "harness|hendrycksTest-college_chemistry|5": {
358
- "acc,none": 0.6615866685657313,
359
- "acc_stderr,none": 0.12826541916322984,
360
- "alias": "mmlu"
361
- },
362
- "harness|hendrycksTest-college_computer_science|5": {
363
- "acc,none": 0.6615866685657313,
364
- "acc_stderr,none": 0.12826541916322984,
365
- "alias": "mmlu"
366
- },
367
- "harness|hendrycksTest-college_mathematics|5": {
368
- "acc,none": 0.6615866685657313,
369
- "acc_stderr,none": 0.12826541916322984,
370
- "alias": "mmlu"
371
- },
372
- "harness|hendrycksTest-college_physics|5": {
373
- "acc,none": 0.6615866685657313,
374
- "acc_stderr,none": 0.12826541916322984,
375
- "alias": "mmlu"
376
- },
377
- "harness|hendrycksTest-computer_security|5": {
378
- "acc,none": 0.6615866685657313,
379
- "acc_stderr,none": 0.12826541916322984,
380
- "alias": "mmlu"
381
- },
382
- "harness|hendrycksTest-conceptual_physics|5": {
383
- "acc,none": 0.6615866685657313,
384
- "acc_stderr,none": 0.12826541916322984,
385
- "alias": "mmlu"
386
- },
387
- "harness|hendrycksTest-electrical_engineering|5": {
388
- "acc,none": 0.6615866685657313,
389
- "acc_stderr,none": 0.12826541916322984,
390
- "alias": "mmlu"
391
- },
392
- "harness|hendrycksTest-elementary_mathematics|5": {
393
- "acc,none": 0.6615866685657313,
394
- "acc_stderr,none": 0.12826541916322984,
395
- "alias": "mmlu"
396
- },
397
- "harness|hendrycksTest-high_school_biology|5": {
398
- "acc,none": 0.6615866685657313,
399
- "acc_stderr,none": 0.12826541916322984,
400
- "alias": "mmlu"
401
- },
402
- "harness|hendrycksTest-high_school_chemistry|5": {
403
- "acc,none": 0.6615866685657313,
404
- "acc_stderr,none": 0.12826541916322984,
405
- "alias": "mmlu"
406
- },
407
- "harness|hendrycksTest-high_school_computer_science|5": {
408
- "acc,none": 0.6615866685657313,
409
- "acc_stderr,none": 0.12826541916322984,
410
- "alias": "mmlu"
411
- },
412
- "harness|hendrycksTest-high_school_mathematics|5": {
413
- "acc,none": 0.6615866685657313,
414
- "acc_stderr,none": 0.12826541916322984,
415
- "alias": "mmlu"
416
- },
417
- "harness|hendrycksTest-high_school_physics|5": {
418
- "acc,none": 0.6615866685657313,
419
- "acc_stderr,none": 0.12826541916322984,
420
- "alias": "mmlu"
421
- },
422
- "harness|hendrycksTest-high_school_statistics|5": {
423
- "acc,none": 0.6615866685657313,
424
- "acc_stderr,none": 0.12826541916322984,
425
- "alias": "mmlu"
426
- },
427
- "harness|hendrycksTest-machine_learning|5": {
428
- "acc,none": 0.6615866685657313,
429
- "acc_stderr,none": 0.12826541916322984,
430
- "alias": "mmlu"
431
- },
432
- "harness|belebele_deu_Latn|5": {
433
- "acc,none": 0.8488888888888889,
434
- "acc_stderr,none": 0.011945209697456624,
435
- "acc_norm,none": 0.8488888888888889,
436
- "acc_norm_stderr,none": 0.011945209697456624,
437
- "alias": "belebele_deu_Latn"
438
- },
439
- "harness|truthfulqa_mc2_m_fr|0": {
440
- "acc,none": 0.2782719186785261,
441
- "acc_stderr,none": 0.015984910901571035,
442
- "alias": "truthfulqa_mc2_m_fr"
443
- },
444
- "harness|mmlu_m_de|5": {
445
- "acc,none": 0.5568713229748077,
446
- "acc_stderr,none": 0.0043143932405325025,
447
- "alias": "mmlu_m_de"
448
- },
449
- "harness|belebele_ita_Latn|5": {
450
- "acc,none": 0.8211111111111111,
451
- "acc_stderr,none": 0.012782411716899067,
452
- "acc_norm,none": 0.8211111111111111,
453
- "acc_norm_stderr,none": 0.012782411716899067,
454
- "alias": "belebele_ita_Latn"
455
- },
456
- "harness|truthfulqa_mc2_m_de|0": {
457
- "acc,none": 0.24619289340101522,
458
- "acc_stderr,none": 0.015356084872692898,
459
- "alias": "truthfulqa_mc2_m_de"
460
- },
461
- "harness|belebele_spa_Latn|5": {
462
- "acc,none": 0.8366666666666667,
463
- "acc_stderr,none": 0.012329168844652513,
464
- "acc_norm,none": 0.8366666666666667,
465
- "acc_norm_stderr,none": 0.012329168844652513,
466
- "alias": "belebele_spa_Latn"
467
- },
468
- "harness|hellaswag_es|10": {
469
- "acc,none": 0.5134414337529336,
470
- "acc_stderr,none": 0.00516266219718063,
471
- "acc_norm,none": 0.6941540430979305,
472
- "acc_norm_stderr,none": 0.004759266148217961,
473
- "alias": "hellaswag_es"
474
- },
475
- "harness|truthfulqa_mc2_m_it|0": {
476
- "acc,none": 0.2950191570881226,
477
- "acc_stderr,none": 0.016308363772932724,
478
- "alias": "truthfulqa_mc2_m_it"
479
- }
480
- },
481
- "versions": {
482
- "harness|hellaswag_fr|10": 1.0,
483
- "harness|arc_challenge_m_it|25": 1.0,
484
- "harness|gsm8k|5": 2.0,
485
- "harness|truthfulqa_mc2|0": 2.0,
486
- "harness|arc_challenge_m_fr|25": 1.0,
487
- "harness|mmlu_m_fr|5": "Yaml",
488
- "harness|truthfulqa_mc2_m_es|0": "Yaml",
489
- "harness|belebele_eng_Latn|5": 0.0,
490
- "harness|mmlu_m_es|5": "Yaml",
491
- "harness|hellaswag_it|10": 1.0,
492
- "harness|belebele_fra_Latn|5": 0.0,
493
- "harness|hellaswag_de|10": 1.0,
494
- "harness|arc_challenge_m_de|25": 1.0,
495
- "harness|mmlu_m_it|5": "Yaml",
496
- "harness|hellaswag|10": 1.0,
497
- "harness|arc_challenge|25": 1.0,
498
- "harness|arc_challenge_m_es|25": 1.0,
499
- "harness|hendrycksTest|5": "N/A",
500
- "harness|hendrycksTest-humanities|5": "N/A",
501
- "harness|hendrycksTest-formal_logic|5": "N/A",
502
- "harness|hendrycksTest-high_school_european_history|5": "N/A",
503
- "harness|hendrycksTest-high_school_us_history|5": "N/A",
504
- "harness|hendrycksTest-high_school_world_history|5": "N/A",
505
- "harness|hendrycksTest-international_law|5": "N/A",
506
- "harness|hendrycksTest-jurisprudence|5": "N/A",
507
- "harness|hendrycksTest-logical_fallacies|5": "N/A",
508
- "harness|hendrycksTest-moral_disputes|5": "N/A",
509
- "harness|hendrycksTest-moral_scenarios|5": "N/A",
510
- "harness|hendrycksTest-philosophy|5": "N/A",
511
- "harness|hendrycksTest-prehistory|5": "N/A",
512
- "harness|hendrycksTest-professional_law|5": "N/A",
513
- "harness|hendrycksTest-world_religions|5": "N/A",
514
- "harness|hendrycksTest-other|5": "N/A",
515
- "harness|hendrycksTest-business_ethics|5": "N/A",
516
- "harness|hendrycksTest-clinical_knowledge|5": "N/A",
517
- "harness|hendrycksTest-college_medicine|5": "N/A",
518
- "harness|hendrycksTest-global_facts|5": "N/A",
519
- "harness|hendrycksTest-human_aging|5": "N/A",
520
- "harness|hendrycksTest-management|5": "N/A",
521
- "harness|hendrycksTest-marketing|5": "N/A",
522
- "harness|hendrycksTest-medical_genetics|5": "N/A",
523
- "harness|hendrycksTest-miscellaneous|5": "N/A",
524
- "harness|hendrycksTest-nutrition|5": "N/A",
525
- "harness|hendrycksTest-professional_accounting|5": "N/A",
526
- "harness|hendrycksTest-professional_medicine|5": "N/A",
527
- "harness|hendrycksTest-virology|5": "N/A",
528
- "harness|hendrycksTest-social_sciences|5": "N/A",
529
- "harness|hendrycksTest-econometrics|5": "N/A",
530
- "harness|hendrycksTest-high_school_geography|5": "N/A",
531
- "harness|hendrycksTest-high_school_government_and_politics|5": "N/A",
532
- "harness|hendrycksTest-high_school_macroeconomics|5": "N/A",
533
- "harness|hendrycksTest-high_school_microeconomics|5": "N/A",
534
- "harness|hendrycksTest-high_school_psychology|5": "N/A",
535
- "harness|hendrycksTest-human_sexuality|5": "N/A",
536
- "harness|hendrycksTest-professional_psychology|5": "N/A",
537
- "harness|hendrycksTest-public_relations|5": "N/A",
538
- "harness|hendrycksTest-security_studies|5": "N/A",
539
- "harness|hendrycksTest-sociology|5": "N/A",
540
- "harness|hendrycksTest-us_foreign_policy|5": "N/A",
541
- "harness|hendrycksTest-stem|5": "N/A",
542
- "harness|hendrycksTest-abstract_algebra|5": "N/A",
543
- "harness|hendrycksTest-anatomy|5": "N/A",
544
- "harness|hendrycksTest-astronomy|5": "N/A",
545
- "harness|hendrycksTest-college_biology|5": "N/A",
546
- "harness|hendrycksTest-college_chemistry|5": "N/A",
547
- "harness|hendrycksTest-college_computer_science|5": "N/A",
548
- "harness|hendrycksTest-college_mathematics|5": "N/A",
549
- "harness|hendrycksTest-college_physics|5": "N/A",
550
- "harness|hendrycksTest-computer_security|5": "N/A",
551
- "harness|hendrycksTest-conceptual_physics|5": "N/A",
552
- "harness|hendrycksTest-electrical_engineering|5": "N/A",
553
- "harness|hendrycksTest-elementary_mathematics|5": "N/A",
554
- "harness|hendrycksTest-high_school_biology|5": "N/A",
555
- "harness|hendrycksTest-high_school_chemistry|5": "N/A",
556
- "harness|hendrycksTest-high_school_computer_science|5": "N/A",
557
- "harness|hendrycksTest-high_school_mathematics|5": "N/A",
558
- "harness|hendrycksTest-high_school_physics|5": "N/A",
559
- "harness|hendrycksTest-high_school_statistics|5": "N/A",
560
- "harness|hendrycksTest-machine_learning|5": "N/A",
561
- "harness|belebele_deu_Latn|5": 0.0,
562
- "harness|truthfulqa_mc2_m_fr|0": "Yaml",
563
- "harness|mmlu_m_de|5": "Yaml",
564
- "harness|belebele_ita_Latn|5": 0.0,
565
- "harness|truthfulqa_mc2_m_de|0": "Yaml",
566
- "harness|belebele_spa_Latn|5": 0.0,
567
- "harness|hellaswag_es|10": 1.0,
568
- "harness|truthfulqa_mc2_m_it|0": "Yaml"
569
- }
570
- }