File size: 3,875 Bytes
3389d64 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 |
{
"name": "heavy",
"uuid": "70f3f719-28b5-46f5-b58e-bd89765d1e40",
"model": "/net/nfs.cirrascale/allennlp/davidw/checkpoints/moe-release/OLMoE-7B-A1B/step1220000-tokens5117B",
"creation_date": "2024_08_06-17_05_21",
"eval_metrics": {
"icl": {
"mmlu_zeroshot": 0.4273025756865217,
"hellaswag_zeroshot": 0.7597092390060425,
"jeopardy": 0.4871806979179382,
"triviaqa_sm_sub": 0.5303333401679993,
"gsm8k_cot": 0.07354056090116501,
"agi_eval_sat_math_cot": 0.040909089148044586,
"aqua_cot": 0.02857142873108387,
"svamp_cot": 0.2800000011920929,
"bigbench_qa_wikidata": 0.6884503960609436,
"arc_easy": 0.7845118045806885,
"arc_challenge": 0.5341296792030334,
"mmlu_fewshot": 0.5194499257363772,
"bigbench_misconceptions": 0.5525113940238953,
"copa": 0.800000011920929,
"siqa": 0.707267165184021,
"commonsense_qa": 0.6699426770210266,
"piqa": 0.8117519021034241,
"openbook_qa": 0.4399999976158142,
"bigbench_novel_concepts": 0.625,
"bigbench_strange_stories": 0.6839080452919006,
"bigbench_strategy_qa": 0.5810397267341614,
"lambada_openai": 0.7219095826148987,
"hellaswag": 0.7750447988510132,
"winograd": 0.831501841545105,
"winogrande": 0.6850828528404236,
"bigbench_conlang_translation": 0.060975611209869385,
"bigbench_language_identification": 0.34049999713897705,
"bigbench_conceptual_combinations": 0.5048543810844421,
"bigbench_elementary_math_qa": 0.27025681734085083,
"bigbench_dyck_languages": 0.1589999943971634,
"agi_eval_lsat_ar": 0.2652173936367035,
"bigbench_cs_algorithms": 0.5015151500701904,
"bigbench_logical_deduction": 0.25333333015441895,
"bigbench_operators": 0.34285715222358704,
"bigbench_repeat_copy_logic": 0.15625,
"simple_arithmetic_nospaces": 0.1809999942779541,
"simple_arithmetic_withspaces": 0.20600000023841858,
"math_qa": 0.2708682417869568,
"logi_qa": 0.3425499200820923,
"pubmed_qa_labeled": 0.4659999907016754,
"squad": 0.5244086980819702,
"agi_eval_lsat_rc": 0.43656715750694275,
"agi_eval_lsat_lr": 0.386274516582489,
"coqa": 0.4436928331851959,
"bigbench_understanding_fables": 0.4444444477558136,
"boolq": 0.7281345725059509,
"agi_eval_sat_en": 0.5485436916351318,
"winogender_mc_female": 0.46666666865348816,
"winogender_mc_male": 0.5833333134651184,
"enterprise_pii_classification": 0.5372606515884399,
"bbq": 0.48323566534302453,
"gpqa_main": 0.2232142835855484,
"gpqa_diamond": 0.21212121844291687
}
},
"missing tasks": "[]",
"aggregated_task_categories_centered": {
"commonsense reasoning": 0.46081640452671535,
"language understanding": 0.4762512398893946,
"reading comprehension": 0.36885401178478144,
"safety": 0.03524814952503552,
"symbolic problem solving": 0.15957477013304083,
"world knowledge": 0.34119598718414534
},
"aggregated_centered_results": 0.30931975984045923,
"aggregated_results": 0.4599646118255447,
"rw_small": 0.7152613600095113,
"rw_small_centered": 0.5043107818441781,
"95%_CI_above": 0.5589368432469957,
"95%_CI_above_centered": 0.41426754302993024,
"99%_CI_above": 0.5701544312031374,
"99%_CI_above_centered": 0.45654352726393493,
"low_variance_datasets": 0.5659450578418646,
"low_variance_datasets_centered": 0.46520260353587645
} |