File size: 3,875 Bytes
3389d64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
{
    "name": "heavy",
    "uuid": "70f3f719-28b5-46f5-b58e-bd89765d1e40",
    "model": "/net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1220000-tokens5117B",
    "creation_date": "2024_08_06-17_05_21",
    "eval_metrics": {
        "icl": {
            "mmlu_zeroshot": 0.4273025756865217,
            "hellaswag_zeroshot": 0.7597092390060425,
            "jeopardy": 0.4871806979179382,
            "triviaqa_sm_sub": 0.5303333401679993,
            "gsm8k_cot": 0.07354056090116501,
            "agi_eval_sat_math_cot": 0.040909089148044586,
            "aqua_cot": 0.02857142873108387,
            "svamp_cot": 0.2800000011920929,
            "bigbench_qa_wikidata": 0.6884503960609436,
            "arc_easy": 0.7845118045806885,
            "arc_challenge": 0.5341296792030334,
            "mmlu_fewshot": 0.5194499257363772,
            "bigbench_misconceptions": 0.5525113940238953,
            "copa": 0.800000011920929,
            "siqa": 0.707267165184021,
            "commonsense_qa": 0.6699426770210266,
            "piqa": 0.8117519021034241,
            "openbook_qa": 0.4399999976158142,
            "bigbench_novel_concepts": 0.625,
            "bigbench_strange_stories": 0.6839080452919006,
            "bigbench_strategy_qa": 0.5810397267341614,
            "lambada_openai": 0.7219095826148987,
            "hellaswag": 0.7750447988510132,
            "winograd": 0.831501841545105,
            "winogrande": 0.6850828528404236,
            "bigbench_conlang_translation": 0.060975611209869385,
            "bigbench_language_identification": 0.34049999713897705,
            "bigbench_conceptual_combinations": 0.5048543810844421,
            "bigbench_elementary_math_qa": 0.27025681734085083,
            "bigbench_dyck_languages": 0.1589999943971634,
            "agi_eval_lsat_ar": 0.2652173936367035,
            "bigbench_cs_algorithms": 0.5015151500701904,
            "bigbench_logical_deduction": 0.25333333015441895,
            "bigbench_operators": 0.34285715222358704,
            "bigbench_repeat_copy_logic": 0.15625,
            "simple_arithmetic_nospaces": 0.1809999942779541,
            "simple_arithmetic_withspaces": 0.20600000023841858,
            "math_qa": 0.2708682417869568,
            "logi_qa": 0.3425499200820923,
            "pubmed_qa_labeled": 0.4659999907016754,
            "squad": 0.5244086980819702,
            "agi_eval_lsat_rc": 0.43656715750694275,
            "agi_eval_lsat_lr": 0.386274516582489,
            "coqa": 0.4436928331851959,
            "bigbench_understanding_fables": 0.4444444477558136,
            "boolq": 0.7281345725059509,
            "agi_eval_sat_en": 0.5485436916351318,
            "winogender_mc_female": 0.46666666865348816,
            "winogender_mc_male": 0.5833333134651184,
            "enterprise_pii_classification": 0.5372606515884399,
            "bbq": 0.48323566534302453,
            "gpqa_main": 0.2232142835855484,
            "gpqa_diamond": 0.21212121844291687
        }
    },
    "missing tasks": "[]",
    "aggregated_task_categories_centered": {
        "commonsense reasoning": 0.46081640452671535,
        "language understanding": 0.4762512398893946,
        "reading comprehension": 0.36885401178478144,
        "safety": 0.03524814952503552,
        "symbolic problem solving": 0.15957477013304083,
        "world knowledge": 0.34119598718414534
    },
    "aggregated_centered_results": 0.30931975984045923,
    "aggregated_results": 0.4599646118255447,
    "rw_small": 0.7152613600095113,
    "rw_small_centered": 0.5043107818441781,
    "95%_CI_above": 0.5589368432469957,
    "95%_CI_above_centered": 0.41426754302993024,
    "99%_CI_above": 0.5701544312031374,
    "99%_CI_above_centered": 0.45654352726393493,
    "low_variance_datasets": 0.5659450578418646,
    "low_variance_datasets_centered": 0.46520260353587645
}