{ "results": { "harness|ko_arc_challenge|25": { "acc": 0.19795221843003413, "acc_stderr": 0.011643990971573401, "acc_norm": 0.24744027303754265, "acc_norm_stderr": 0.01261035266329267 }, "harness|ko_hellaswag|10": { "acc": 0.255327623979287, "acc_stderr": 0.004351540603988567, "acc_norm": 0.2583150766779526, "acc_norm_stderr": 0.004368135676213556 }, "harness|ko_mmlu_world_religions|5": { "acc": 0.39766081871345027, "acc_stderr": 0.03753638955761691, "acc_norm": 0.39766081871345027, "acc_norm_stderr": 0.03753638955761691 }, "harness|ko_mmlu_management|5": { "acc": 0.32038834951456313, "acc_stderr": 0.046202840822800406, "acc_norm": 0.32038834951456313, "acc_norm_stderr": 0.046202840822800406 }, "harness|ko_mmlu_miscellaneous|5": { "acc": 0.3103448275862069, "acc_stderr": 0.016543785026048315, "acc_norm": 0.3103448275862069, "acc_norm_stderr": 0.016543785026048315 }, "harness|ko_mmlu_anatomy|5": { "acc": 0.2518518518518518, "acc_stderr": 0.037498507091740234, "acc_norm": 0.2518518518518518, "acc_norm_stderr": 0.037498507091740234 }, "harness|ko_mmlu_abstract_algebra|5": { "acc": 0.26, "acc_stderr": 0.04408440022768078, "acc_norm": 0.26, "acc_norm_stderr": 0.04408440022768078 }, "harness|ko_mmlu_conceptual_physics|5": { "acc": 0.2765957446808511, "acc_stderr": 0.02924188386962882, "acc_norm": 0.2765957446808511, "acc_norm_stderr": 0.02924188386962882 }, "harness|ko_mmlu_virology|5": { "acc": 0.29518072289156627, "acc_stderr": 0.03550920185689629, "acc_norm": 0.29518072289156627, "acc_norm_stderr": 0.03550920185689629 }, "harness|ko_mmlu_philosophy|5": { "acc": 0.2861736334405145, "acc_stderr": 0.025670259242188943, "acc_norm": 0.2861736334405145, "acc_norm_stderr": 0.025670259242188943 }, "harness|ko_mmlu_human_aging|5": { "acc": 0.17937219730941703, "acc_stderr": 0.025749819569192794, "acc_norm": 0.17937219730941703, "acc_norm_stderr": 0.025749819569192794 }, "harness|ko_mmlu_human_sexuality|5": { "acc": 0.33587786259541985, "acc_stderr": 0.041423137719966634, "acc_norm": 0.33587786259541985, "acc_norm_stderr": 0.041423137719966634 }, "harness|ko_mmlu_medical_genetics|5": { "acc": 0.3, "acc_stderr": 0.046056618647183814, "acc_norm": 0.3, "acc_norm_stderr": 0.046056618647183814 }, "harness|ko_mmlu_high_school_geography|5": { "acc": 0.3181818181818182, "acc_stderr": 0.03318477333845331, "acc_norm": 0.3181818181818182, "acc_norm_stderr": 0.03318477333845331 }, "harness|ko_mmlu_electrical_engineering|5": { "acc": 0.41379310344827586, "acc_stderr": 0.04104269211806232, "acc_norm": 0.41379310344827586, "acc_norm_stderr": 0.04104269211806232 }, "harness|ko_mmlu_college_physics|5": { "acc": 0.19607843137254902, "acc_stderr": 0.03950581861179961, "acc_norm": 0.19607843137254902, "acc_norm_stderr": 0.03950581861179961 }, "harness|ko_mmlu_high_school_microeconomics|5": { "acc": 0.35714285714285715, "acc_stderr": 0.031124619309328177, "acc_norm": 0.35714285714285715, "acc_norm_stderr": 0.031124619309328177 }, "harness|ko_mmlu_high_school_macroeconomics|5": { "acc": 0.33589743589743587, "acc_stderr": 0.023946724741563976, "acc_norm": 0.33589743589743587, "acc_norm_stderr": 0.023946724741563976 }, "harness|ko_mmlu_computer_security|5": { "acc": 0.33, "acc_stderr": 0.047258156262526045, "acc_norm": 0.33, "acc_norm_stderr": 0.047258156262526045 }, "harness|ko_mmlu_global_facts|5": { "acc": 0.29, "acc_stderr": 0.04560480215720683, "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720683 }, "harness|ko_mmlu_jurisprudence|5": { "acc": 0.24074074074074073, "acc_stderr": 0.04133119440243838, "acc_norm": 0.24074074074074073, "acc_norm_stderr": 0.04133119440243838 }, "harness|ko_mmlu_high_school_chemistry|5": { "acc": 0.33004926108374383, "acc_stderr": 0.03308530426228257, "acc_norm": 0.33004926108374383, "acc_norm_stderr": 0.03308530426228257 }, "harness|ko_mmlu_high_school_biology|5": { "acc": 0.31290322580645163, "acc_stderr": 0.02637756702864586, "acc_norm": 0.31290322580645163, "acc_norm_stderr": 0.02637756702864586 }, "harness|ko_mmlu_marketing|5": { "acc": 0.2863247863247863, "acc_stderr": 0.029614323690456648, "acc_norm": 0.2863247863247863, "acc_norm_stderr": 0.029614323690456648 }, "harness|ko_mmlu_clinical_knowledge|5": { "acc": 0.27547169811320754, "acc_stderr": 0.027495663683724043, "acc_norm": 0.27547169811320754, "acc_norm_stderr": 0.027495663683724043 }, "harness|ko_mmlu_public_relations|5": { "acc": 0.22727272727272727, "acc_stderr": 0.04013964554072773, "acc_norm": 0.22727272727272727, "acc_norm_stderr": 0.04013964554072773 }, "harness|ko_mmlu_high_school_mathematics|5": { "acc": 0.28888888888888886, "acc_stderr": 0.027634907264178544, "acc_norm": 0.28888888888888886, "acc_norm_stderr": 0.027634907264178544 }, "harness|ko_mmlu_high_school_physics|5": { "acc": 0.2980132450331126, "acc_stderr": 0.03734535676787198, "acc_norm": 0.2980132450331126, "acc_norm_stderr": 0.03734535676787198 }, "harness|ko_mmlu_sociology|5": { "acc": 0.3482587064676617, "acc_stderr": 0.033687874661154596, "acc_norm": 0.3482587064676617, "acc_norm_stderr": 0.033687874661154596 }, "harness|ko_mmlu_college_medicine|5": { "acc": 0.27167630057803466, "acc_stderr": 0.03391750322321659, "acc_norm": 0.27167630057803466, "acc_norm_stderr": 0.03391750322321659 }, "harness|ko_mmlu_elementary_mathematics|5": { "acc": 0.2671957671957672, "acc_stderr": 0.02278967314577656, "acc_norm": 0.2671957671957672, "acc_norm_stderr": 0.02278967314577656 }, "harness|ko_mmlu_college_biology|5": { "acc": 0.2361111111111111, "acc_stderr": 0.03551446610810826, "acc_norm": 0.2361111111111111, "acc_norm_stderr": 0.03551446610810826 }, "harness|ko_mmlu_college_chemistry|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_us_foreign_policy|5": { "acc": 0.35, "acc_stderr": 0.047937248544110196, "acc_norm": 0.35, "acc_norm_stderr": 0.047937248544110196 }, "harness|ko_mmlu_moral_disputes|5": { "acc": 0.2630057803468208, "acc_stderr": 0.023703099525258165, "acc_norm": 0.2630057803468208, "acc_norm_stderr": 0.023703099525258165 }, "harness|ko_mmlu_logical_fallacies|5": { "acc": 0.3128834355828221, "acc_stderr": 0.03642914578292406, "acc_norm": 0.3128834355828221, "acc_norm_stderr": 0.03642914578292406 }, "harness|ko_mmlu_prehistory|5": { "acc": 0.3148148148148148, "acc_stderr": 0.025842248700902168, "acc_norm": 0.3148148148148148, "acc_norm_stderr": 0.025842248700902168 }, "harness|ko_mmlu_college_mathematics|5": { "acc": 0.26, "acc_stderr": 0.0440844002276808, "acc_norm": 0.26, "acc_norm_stderr": 0.0440844002276808 }, "harness|ko_mmlu_high_school_government_and_politics|5": { "acc": 0.2694300518134715, "acc_stderr": 0.032018671228777947, "acc_norm": 0.2694300518134715, "acc_norm_stderr": 0.032018671228777947 }, "harness|ko_mmlu_econometrics|5": { "acc": 0.24561403508771928, "acc_stderr": 0.04049339297748141, "acc_norm": 0.24561403508771928, "acc_norm_stderr": 0.04049339297748141 }, "harness|ko_mmlu_high_school_psychology|5": { "acc": 0.3137614678899083, "acc_stderr": 0.01989472334146913, "acc_norm": 0.3137614678899083, "acc_norm_stderr": 0.01989472334146913 }, "harness|ko_mmlu_formal_logic|5": { "acc": 0.25396825396825395, "acc_stderr": 0.03893259610604675, "acc_norm": 0.25396825396825395, "acc_norm_stderr": 0.03893259610604675 }, "harness|ko_mmlu_nutrition|5": { "acc": 0.3137254901960784, "acc_stderr": 0.026568921015457162, "acc_norm": 0.3137254901960784, "acc_norm_stderr": 0.026568921015457162 }, "harness|ko_mmlu_business_ethics|5": { "acc": 0.24, "acc_stderr": 0.04292346959909283, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909283 }, "harness|ko_mmlu_international_law|5": { "acc": 0.39669421487603307, "acc_stderr": 0.044658697805310094, "acc_norm": 0.39669421487603307, "acc_norm_stderr": 0.044658697805310094 }, "harness|ko_mmlu_astronomy|5": { "acc": 0.35526315789473684, "acc_stderr": 0.03894734487013316, "acc_norm": 0.35526315789473684, "acc_norm_stderr": 0.03894734487013316 }, "harness|ko_mmlu_professional_psychology|5": { "acc": 0.2369281045751634, "acc_stderr": 0.01720166216978978, "acc_norm": 0.2369281045751634, "acc_norm_stderr": 0.01720166216978978 }, "harness|ko_mmlu_professional_accounting|5": { "acc": 0.2695035460992908, "acc_stderr": 0.026469036818590624, "acc_norm": 0.2695035460992908, "acc_norm_stderr": 0.026469036818590624 }, "harness|ko_mmlu_machine_learning|5": { "acc": 0.23214285714285715, "acc_stderr": 0.04007341809755807, "acc_norm": 0.23214285714285715, "acc_norm_stderr": 0.04007341809755807 }, "harness|ko_mmlu_high_school_statistics|5": { "acc": 0.4305555555555556, "acc_stderr": 0.03376922151252335, "acc_norm": 0.4305555555555556, "acc_norm_stderr": 0.03376922151252335 }, "harness|ko_mmlu_moral_scenarios|5": { "acc": 0.26927374301675977, "acc_stderr": 0.014835616582882601, "acc_norm": 0.26927374301675977, "acc_norm_stderr": 0.014835616582882601 }, "harness|ko_mmlu_college_computer_science|5": { "acc": 0.36, "acc_stderr": 0.048241815132442176, "acc_norm": 0.36, "acc_norm_stderr": 0.048241815132442176 }, "harness|ko_mmlu_high_school_computer_science|5": { "acc": 0.31, "acc_stderr": 0.04648231987117316, "acc_norm": 0.31, "acc_norm_stderr": 0.04648231987117316 }, "harness|ko_mmlu_professional_medicine|5": { "acc": 0.43014705882352944, "acc_stderr": 0.030074971917302875, "acc_norm": 0.43014705882352944, "acc_norm_stderr": 0.030074971917302875 }, "harness|ko_mmlu_security_studies|5": { "acc": 0.35918367346938773, "acc_stderr": 0.030713560455108493, "acc_norm": 0.35918367346938773, "acc_norm_stderr": 0.030713560455108493 }, "harness|ko_mmlu_high_school_world_history|5": { "acc": 0.22784810126582278, "acc_stderr": 0.02730348459906944, "acc_norm": 0.22784810126582278, "acc_norm_stderr": 0.02730348459906944 }, "harness|ko_mmlu_professional_law|5": { "acc": 0.24837027379400262, "acc_stderr": 0.011035212598034503, "acc_norm": 0.24837027379400262, "acc_norm_stderr": 0.011035212598034503 }, "harness|ko_mmlu_high_school_us_history|5": { "acc": 0.23039215686274508, "acc_stderr": 0.029554292605695063, "acc_norm": 0.23039215686274508, "acc_norm_stderr": 0.029554292605695063 }, "harness|ko_mmlu_high_school_european_history|5": { "acc": 0.2, "acc_stderr": 0.03123475237772118, "acc_norm": 0.2, "acc_norm_stderr": 0.03123475237772118 }, "harness|ko_truthfulqa_mc|0": { "mc1": 0.2729498164014688, "mc1_stderr": 0.015594753632006526, "mc2": 0.5147520134989033, "mc2_stderr": 0.016916247986138434 }, "harness|ko_commongen_v2|2": { "acc": 0.10743801652892562, "acc_stderr": 0.01064665060891188, "acc_norm": 0.31641086186540734, "acc_norm_stderr": 0.015989617951065477 } }, "versions": { "all": 0, "harness|ko_arc_challenge|25": 0, "harness|ko_hellaswag|10": 0, "harness|ko_mmlu_world_religions|5": 1, "harness|ko_mmlu_management|5": 1, "harness|ko_mmlu_miscellaneous|5": 1, "harness|ko_mmlu_anatomy|5": 1, "harness|ko_mmlu_abstract_algebra|5": 1, "harness|ko_mmlu_conceptual_physics|5": 1, "harness|ko_mmlu_virology|5": 1, "harness|ko_mmlu_philosophy|5": 1, "harness|ko_mmlu_human_aging|5": 1, "harness|ko_mmlu_human_sexuality|5": 1, "harness|ko_mmlu_medical_genetics|5": 1, "harness|ko_mmlu_high_school_geography|5": 1, "harness|ko_mmlu_electrical_engineering|5": 1, "harness|ko_mmlu_college_physics|5": 1, "harness|ko_mmlu_high_school_microeconomics|5": 1, "harness|ko_mmlu_high_school_macroeconomics|5": 1, "harness|ko_mmlu_computer_security|5": 1, "harness|ko_mmlu_global_facts|5": 1, "harness|ko_mmlu_jurisprudence|5": 1, "harness|ko_mmlu_high_school_chemistry|5": 1, "harness|ko_mmlu_high_school_biology|5": 1, "harness|ko_mmlu_marketing|5": 1, "harness|ko_mmlu_clinical_knowledge|5": 1, "harness|ko_mmlu_public_relations|5": 1, "harness|ko_mmlu_high_school_mathematics|5": 1, "harness|ko_mmlu_high_school_physics|5": 1, "harness|ko_mmlu_sociology|5": 1, "harness|ko_mmlu_college_medicine|5": 1, "harness|ko_mmlu_elementary_mathematics|5": 1, "harness|ko_mmlu_college_biology|5": 1, "harness|ko_mmlu_college_chemistry|5": 1, "harness|ko_mmlu_us_foreign_policy|5": 1, "harness|ko_mmlu_moral_disputes|5": 1, "harness|ko_mmlu_logical_fallacies|5": 1, "harness|ko_mmlu_prehistory|5": 1, "harness|ko_mmlu_college_mathematics|5": 1, "harness|ko_mmlu_high_school_government_and_politics|5": 1, "harness|ko_mmlu_econometrics|5": 1, "harness|ko_mmlu_high_school_psychology|5": 1, "harness|ko_mmlu_formal_logic|5": 1, "harness|ko_mmlu_nutrition|5": 1, "harness|ko_mmlu_business_ethics|5": 1, "harness|ko_mmlu_international_law|5": 1, "harness|ko_mmlu_astronomy|5": 1, "harness|ko_mmlu_professional_psychology|5": 1, "harness|ko_mmlu_professional_accounting|5": 1, "harness|ko_mmlu_machine_learning|5": 1, "harness|ko_mmlu_high_school_statistics|5": 1, "harness|ko_mmlu_moral_scenarios|5": 1, "harness|ko_mmlu_college_computer_science|5": 1, "harness|ko_mmlu_high_school_computer_science|5": 1, "harness|ko_mmlu_professional_medicine|5": 1, "harness|ko_mmlu_security_studies|5": 1, "harness|ko_mmlu_high_school_world_history|5": 1, "harness|ko_mmlu_professional_law|5": 1, "harness|ko_mmlu_high_school_us_history|5": 1, "harness|ko_mmlu_high_school_european_history|5": 1, "harness|ko_truthfulqa_mc|0": 0, "harness|ko_commongen_v2|2": 1 }, "config_general": { "model_name": "4yo1/llama3-pre1-pre2-inst3-ds-lora3", "model_sha": "d5fbcf408c849c5efd33e1dd1c36f9f583c9433b", "model_dtype": "torch.float16", "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null } }