Sean Cho
test data
ec0015a
{
"results": {
"harness|ko_arc_challenge|25": {
"acc": 0.24829351535836178,
"acc_stderr": 0.012624912868089764,
"acc_norm": 0.2858361774744027,
"acc_norm_stderr": 0.013203196088537369
},
"harness|ko_hellaswag|10": {
"acc": 0.35371439952200756,
"acc_stderr": 0.004771447244095125,
"acc_norm": 0.4420434176458873,
"acc_norm_stderr": 0.004956147046108963
},
"harness|ko_mmlu_world_religions|5": {
"acc": 0.28654970760233917,
"acc_stderr": 0.034678266857038245,
"acc_norm": 0.28654970760233917,
"acc_norm_stderr": 0.034678266857038245
},
"harness|ko_mmlu_management|5": {
"acc": 0.21359223300970873,
"acc_stderr": 0.040580420156460344,
"acc_norm": 0.21359223300970873,
"acc_norm_stderr": 0.040580420156460344
},
"harness|ko_mmlu_miscellaneous|5": {
"acc": 0.2681992337164751,
"acc_stderr": 0.015842430835269445,
"acc_norm": 0.2681992337164751,
"acc_norm_stderr": 0.015842430835269445
},
"harness|ko_mmlu_anatomy|5": {
"acc": 0.2074074074074074,
"acc_stderr": 0.03502553170678317,
"acc_norm": 0.2074074074074074,
"acc_norm_stderr": 0.03502553170678317
},
"harness|ko_mmlu_abstract_algebra|5": {
"acc": 0.27,
"acc_stderr": 0.0446196043338474,
"acc_norm": 0.27,
"acc_norm_stderr": 0.0446196043338474
},
"harness|ko_mmlu_conceptual_physics|5": {
"acc": 0.23404255319148937,
"acc_stderr": 0.027678452578212373,
"acc_norm": 0.23404255319148937,
"acc_norm_stderr": 0.027678452578212373
},
"harness|ko_mmlu_virology|5": {
"acc": 0.2469879518072289,
"acc_stderr": 0.03357351982064536,
"acc_norm": 0.2469879518072289,
"acc_norm_stderr": 0.03357351982064536
},
"harness|ko_mmlu_philosophy|5": {
"acc": 0.26366559485530544,
"acc_stderr": 0.02502553850053234,
"acc_norm": 0.26366559485530544,
"acc_norm_stderr": 0.02502553850053234
},
"harness|ko_mmlu_human_aging|5": {
"acc": 0.18385650224215247,
"acc_stderr": 0.025998379092356513,
"acc_norm": 0.18385650224215247,
"acc_norm_stderr": 0.025998379092356513
},
"harness|ko_mmlu_human_sexuality|5": {
"acc": 0.21374045801526717,
"acc_stderr": 0.0359546161177469,
"acc_norm": 0.21374045801526717,
"acc_norm_stderr": 0.0359546161177469
},
"harness|ko_mmlu_medical_genetics|5": {
"acc": 0.24,
"acc_stderr": 0.042923469599092816,
"acc_norm": 0.24,
"acc_norm_stderr": 0.042923469599092816
},
"harness|ko_mmlu_high_school_geography|5": {
"acc": 0.25757575757575757,
"acc_stderr": 0.03115626951964684,
"acc_norm": 0.25757575757575757,
"acc_norm_stderr": 0.03115626951964684
},
"harness|ko_mmlu_electrical_engineering|5": {
"acc": 0.27586206896551724,
"acc_stderr": 0.03724563619774632,
"acc_norm": 0.27586206896551724,
"acc_norm_stderr": 0.03724563619774632
},
"harness|ko_mmlu_college_physics|5": {
"acc": 0.10784313725490197,
"acc_stderr": 0.030864282122060136,
"acc_norm": 0.10784313725490197,
"acc_norm_stderr": 0.030864282122060136
},
"harness|ko_mmlu_high_school_microeconomics|5": {
"acc": 0.24369747899159663,
"acc_stderr": 0.02788682807838056,
"acc_norm": 0.24369747899159663,
"acc_norm_stderr": 0.02788682807838056
},
"harness|ko_mmlu_high_school_macroeconomics|5": {
"acc": 0.24102564102564103,
"acc_stderr": 0.0216855466653332,
"acc_norm": 0.24102564102564103,
"acc_norm_stderr": 0.0216855466653332
},
"harness|ko_mmlu_computer_security|5": {
"acc": 0.26,
"acc_stderr": 0.0440844002276808,
"acc_norm": 0.26,
"acc_norm_stderr": 0.0440844002276808
},
"harness|ko_mmlu_global_facts|5": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"harness|ko_mmlu_jurisprudence|5": {
"acc": 0.2037037037037037,
"acc_stderr": 0.03893542518824847,
"acc_norm": 0.2037037037037037,
"acc_norm_stderr": 0.03893542518824847
},
"harness|ko_mmlu_high_school_chemistry|5": {
"acc": 0.21182266009852216,
"acc_stderr": 0.028748983689941072,
"acc_norm": 0.21182266009852216,
"acc_norm_stderr": 0.028748983689941072
},
"harness|ko_mmlu_high_school_biology|5": {
"acc": 0.2645161290322581,
"acc_stderr": 0.025091892378859275,
"acc_norm": 0.2645161290322581,
"acc_norm_stderr": 0.025091892378859275
},
"harness|ko_mmlu_marketing|5": {
"acc": 0.2606837606837607,
"acc_stderr": 0.028760348956523414,
"acc_norm": 0.2606837606837607,
"acc_norm_stderr": 0.028760348956523414
},
"harness|ko_mmlu_clinical_knowledge|5": {
"acc": 0.22641509433962265,
"acc_stderr": 0.025757559893106727,
"acc_norm": 0.22641509433962265,
"acc_norm_stderr": 0.025757559893106727
},
"harness|ko_mmlu_public_relations|5": {
"acc": 0.22727272727272727,
"acc_stderr": 0.04013964554072775,
"acc_norm": 0.22727272727272727,
"acc_norm_stderr": 0.04013964554072775
},
"harness|ko_mmlu_high_school_mathematics|5": {
"acc": 0.26296296296296295,
"acc_stderr": 0.026842057873833713,
"acc_norm": 0.26296296296296295,
"acc_norm_stderr": 0.026842057873833713
},
"harness|ko_mmlu_high_school_physics|5": {
"acc": 0.2582781456953642,
"acc_stderr": 0.035737053147634576,
"acc_norm": 0.2582781456953642,
"acc_norm_stderr": 0.035737053147634576
},
"harness|ko_mmlu_sociology|5": {
"acc": 0.31343283582089554,
"acc_stderr": 0.03280188205348642,
"acc_norm": 0.31343283582089554,
"acc_norm_stderr": 0.03280188205348642
},
"harness|ko_mmlu_college_medicine|5": {
"acc": 0.23699421965317918,
"acc_stderr": 0.03242414757483098,
"acc_norm": 0.23699421965317918,
"acc_norm_stderr": 0.03242414757483098
},
"harness|ko_mmlu_elementary_mathematics|5": {
"acc": 0.2671957671957672,
"acc_stderr": 0.022789673145776575,
"acc_norm": 0.2671957671957672,
"acc_norm_stderr": 0.022789673145776575
},
"harness|ko_mmlu_college_biology|5": {
"acc": 0.2638888888888889,
"acc_stderr": 0.03685651095897532,
"acc_norm": 0.2638888888888889,
"acc_norm_stderr": 0.03685651095897532
},
"harness|ko_mmlu_college_chemistry|5": {
"acc": 0.19,
"acc_stderr": 0.03942772444036622,
"acc_norm": 0.19,
"acc_norm_stderr": 0.03942772444036622
},
"harness|ko_mmlu_us_foreign_policy|5": {
"acc": 0.29,
"acc_stderr": 0.04560480215720684,
"acc_norm": 0.29,
"acc_norm_stderr": 0.04560480215720684
},
"harness|ko_mmlu_moral_disputes|5": {
"acc": 0.26011560693641617,
"acc_stderr": 0.023618678310069363,
"acc_norm": 0.26011560693641617,
"acc_norm_stderr": 0.023618678310069363
},
"harness|ko_mmlu_logical_fallacies|5": {
"acc": 0.3128834355828221,
"acc_stderr": 0.03642914578292404,
"acc_norm": 0.3128834355828221,
"acc_norm_stderr": 0.03642914578292404
},
"harness|ko_mmlu_prehistory|5": {
"acc": 0.3148148148148148,
"acc_stderr": 0.025842248700902168,
"acc_norm": 0.3148148148148148,
"acc_norm_stderr": 0.025842248700902168
},
"harness|ko_mmlu_college_mathematics|5": {
"acc": 0.25,
"acc_stderr": 0.04351941398892446,
"acc_norm": 0.25,
"acc_norm_stderr": 0.04351941398892446
},
"harness|ko_mmlu_high_school_government_and_politics|5": {
"acc": 0.24352331606217617,
"acc_stderr": 0.030975436386845436,
"acc_norm": 0.24352331606217617,
"acc_norm_stderr": 0.030975436386845436
},
"harness|ko_mmlu_econometrics|5": {
"acc": 0.19298245614035087,
"acc_stderr": 0.037124548537213684,
"acc_norm": 0.19298245614035087,
"acc_norm_stderr": 0.037124548537213684
},
"harness|ko_mmlu_high_school_psychology|5": {
"acc": 0.21100917431192662,
"acc_stderr": 0.01749392240411265,
"acc_norm": 0.21100917431192662,
"acc_norm_stderr": 0.01749392240411265
},
"harness|ko_mmlu_formal_logic|5": {
"acc": 0.16666666666666666,
"acc_stderr": 0.03333333333333338,
"acc_norm": 0.16666666666666666,
"acc_norm_stderr": 0.03333333333333338
},
"harness|ko_mmlu_nutrition|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.024630048979824765,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.024630048979824765
},
"harness|ko_mmlu_business_ethics|5": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"harness|ko_mmlu_international_law|5": {
"acc": 0.371900826446281,
"acc_stderr": 0.04412015806624503,
"acc_norm": 0.371900826446281,
"acc_norm_stderr": 0.04412015806624503
},
"harness|ko_mmlu_astronomy|5": {
"acc": 0.20394736842105263,
"acc_stderr": 0.032790004063100515,
"acc_norm": 0.20394736842105263,
"acc_norm_stderr": 0.032790004063100515
},
"harness|ko_mmlu_professional_psychology|5": {
"acc": 0.25326797385620914,
"acc_stderr": 0.01759348689536683,
"acc_norm": 0.25326797385620914,
"acc_norm_stderr": 0.01759348689536683
},
"harness|ko_mmlu_professional_accounting|5": {
"acc": 0.2375886524822695,
"acc_stderr": 0.025389512552729903,
"acc_norm": 0.2375886524822695,
"acc_norm_stderr": 0.025389512552729903
},
"harness|ko_mmlu_machine_learning|5": {
"acc": 0.24107142857142858,
"acc_stderr": 0.04059867246952689,
"acc_norm": 0.24107142857142858,
"acc_norm_stderr": 0.04059867246952689
},
"harness|ko_mmlu_high_school_statistics|5": {
"acc": 0.2361111111111111,
"acc_stderr": 0.02896370257079103,
"acc_norm": 0.2361111111111111,
"acc_norm_stderr": 0.02896370257079103
},
"harness|ko_mmlu_moral_scenarios|5": {
"acc": 0.28156424581005585,
"acc_stderr": 0.015042290171866132,
"acc_norm": 0.28156424581005585,
"acc_norm_stderr": 0.015042290171866132
},
"harness|ko_mmlu_college_computer_science|5": {
"acc": 0.24,
"acc_stderr": 0.042923469599092816,
"acc_norm": 0.24,
"acc_norm_stderr": 0.042923469599092816
},
"harness|ko_mmlu_high_school_computer_science|5": {
"acc": 0.27,
"acc_stderr": 0.044619604333847394,
"acc_norm": 0.27,
"acc_norm_stderr": 0.044619604333847394
},
"harness|ko_mmlu_professional_medicine|5": {
"acc": 0.34191176470588236,
"acc_stderr": 0.02881472242225418,
"acc_norm": 0.34191176470588236,
"acc_norm_stderr": 0.02881472242225418
},
"harness|ko_mmlu_security_studies|5": {
"acc": 0.18775510204081633,
"acc_stderr": 0.02500025603954621,
"acc_norm": 0.18775510204081633,
"acc_norm_stderr": 0.02500025603954621
},
"harness|ko_mmlu_high_school_world_history|5": {
"acc": 0.270042194092827,
"acc_stderr": 0.028900721906293426,
"acc_norm": 0.270042194092827,
"acc_norm_stderr": 0.028900721906293426
},
"harness|ko_mmlu_professional_law|5": {
"acc": 0.24185136897001303,
"acc_stderr": 0.010936550813827054,
"acc_norm": 0.24185136897001303,
"acc_norm_stderr": 0.010936550813827054
},
"harness|ko_mmlu_high_school_us_history|5": {
"acc": 0.24509803921568626,
"acc_stderr": 0.03019028245350195,
"acc_norm": 0.24509803921568626,
"acc_norm_stderr": 0.03019028245350195
},
"harness|ko_mmlu_high_school_european_history|5": {
"acc": 0.2606060606060606,
"acc_stderr": 0.03427743175816525,
"acc_norm": 0.2606060606060606,
"acc_norm_stderr": 0.03427743175816525
},
"harness|ko_truthfulqa_mc|0": {
"mc1": 0.22399020807833536,
"mc1_stderr": 0.014594964329474203,
"mc2": 0.4106638009419967,
"mc2_stderr": 0.015724386722290755
},
"harness|ko_commongen_v2|2": {
"acc": 0.3187721369539551,
"acc_stderr": 0.01602142705530959,
"acc_norm": 0.38488783943329397,
"acc_norm_stderr": 0.01672857970149866
}
},
"versions": {
"all": 0,
"harness|ko_arc_challenge|25": 0,
"harness|ko_hellaswag|10": 0,
"harness|ko_mmlu_world_religions|5": 1,
"harness|ko_mmlu_management|5": 1,
"harness|ko_mmlu_miscellaneous|5": 1,
"harness|ko_mmlu_anatomy|5": 1,
"harness|ko_mmlu_abstract_algebra|5": 1,
"harness|ko_mmlu_conceptual_physics|5": 1,
"harness|ko_mmlu_virology|5": 1,
"harness|ko_mmlu_philosophy|5": 1,
"harness|ko_mmlu_human_aging|5": 1,
"harness|ko_mmlu_human_sexuality|5": 1,
"harness|ko_mmlu_medical_genetics|5": 1,
"harness|ko_mmlu_high_school_geography|5": 1,
"harness|ko_mmlu_electrical_engineering|5": 1,
"harness|ko_mmlu_college_physics|5": 1,
"harness|ko_mmlu_high_school_microeconomics|5": 1,
"harness|ko_mmlu_high_school_macroeconomics|5": 1,
"harness|ko_mmlu_computer_security|5": 1,
"harness|ko_mmlu_global_facts|5": 1,
"harness|ko_mmlu_jurisprudence|5": 1,
"harness|ko_mmlu_high_school_chemistry|5": 1,
"harness|ko_mmlu_high_school_biology|5": 1,
"harness|ko_mmlu_marketing|5": 1,
"harness|ko_mmlu_clinical_knowledge|5": 1,
"harness|ko_mmlu_public_relations|5": 1,
"harness|ko_mmlu_high_school_mathematics|5": 1,
"harness|ko_mmlu_high_school_physics|5": 1,
"harness|ko_mmlu_sociology|5": 1,
"harness|ko_mmlu_college_medicine|5": 1,
"harness|ko_mmlu_elementary_mathematics|5": 1,
"harness|ko_mmlu_college_biology|5": 1,
"harness|ko_mmlu_college_chemistry|5": 1,
"harness|ko_mmlu_us_foreign_policy|5": 1,
"harness|ko_mmlu_moral_disputes|5": 1,
"harness|ko_mmlu_logical_fallacies|5": 1,
"harness|ko_mmlu_prehistory|5": 1,
"harness|ko_mmlu_college_mathematics|5": 1,
"harness|ko_mmlu_high_school_government_and_politics|5": 1,
"harness|ko_mmlu_econometrics|5": 1,
"harness|ko_mmlu_high_school_psychology|5": 1,
"harness|ko_mmlu_formal_logic|5": 1,
"harness|ko_mmlu_nutrition|5": 1,
"harness|ko_mmlu_business_ethics|5": 1,
"harness|ko_mmlu_international_law|5": 1,
"harness|ko_mmlu_astronomy|5": 1,
"harness|ko_mmlu_professional_psychology|5": 1,
"harness|ko_mmlu_professional_accounting|5": 1,
"harness|ko_mmlu_machine_learning|5": 1,
"harness|ko_mmlu_high_school_statistics|5": 1,
"harness|ko_mmlu_moral_scenarios|5": 1,
"harness|ko_mmlu_college_computer_science|5": 1,
"harness|ko_mmlu_high_school_computer_science|5": 1,
"harness|ko_mmlu_professional_medicine|5": 1,
"harness|ko_mmlu_security_studies|5": 1,
"harness|ko_mmlu_high_school_world_history|5": 1,
"harness|ko_mmlu_professional_law|5": 1,
"harness|ko_mmlu_high_school_us_history|5": 1,
"harness|ko_mmlu_high_school_european_history|5": 1,
"harness|ko_truthfulqa_mc|0": 0,
"harness|ko_commongen_v2|2": 1
},
"config_general": {
"model_name": "GUSSSSSSSSSSS/polyglot-ko-12.8b-instruction",
"model_sha": "5dd983e0688b676b814f4b9a02810de2d31dafb3",
"model_dtype": "torch.float16",
"lighteval_sha": "",
"num_few_shot_default": 0,
"num_fewshot_seeds": 1,
"override_batch_size": 1,
"max_samples": null
}
}