|
{ |
|
"results": { |
|
"harness|ko_arc_challenge|25": { |
|
"acc": 0.3873720136518771, |
|
"acc_stderr": 0.01423587248790987, |
|
"acc_norm": 0.4274744027303754, |
|
"acc_norm_stderr": 0.014456862944650647 |
|
}, |
|
"harness|ko_hellaswag|10": { |
|
"acc": 0.392850029874527, |
|
"acc_stderr": 0.0048738583238407945, |
|
"acc_norm": 0.5120493925512846, |
|
"acc_norm_stderr": 0.004988332289642083 |
|
}, |
|
"harness|ko_mmlu_world_religions|5": { |
|
"acc": 0.52046783625731, |
|
"acc_stderr": 0.038316105328219295, |
|
"acc_norm": 0.52046783625731, |
|
"acc_norm_stderr": 0.038316105328219295 |
|
}, |
|
"harness|ko_mmlu_management|5": { |
|
"acc": 0.6116504854368932, |
|
"acc_stderr": 0.04825729337356389, |
|
"acc_norm": 0.6116504854368932, |
|
"acc_norm_stderr": 0.04825729337356389 |
|
}, |
|
"harness|ko_mmlu_miscellaneous|5": { |
|
"acc": 0.5057471264367817, |
|
"acc_stderr": 0.017878782326129224, |
|
"acc_norm": 0.5057471264367817, |
|
"acc_norm_stderr": 0.017878782326129224 |
|
}, |
|
"harness|ko_mmlu_anatomy|5": { |
|
"acc": 0.3851851851851852, |
|
"acc_stderr": 0.042039210401562783, |
|
"acc_norm": 0.3851851851851852, |
|
"acc_norm_stderr": 0.042039210401562783 |
|
}, |
|
"harness|ko_mmlu_abstract_algebra|5": { |
|
"acc": 0.25, |
|
"acc_stderr": 0.04351941398892446, |
|
"acc_norm": 0.25, |
|
"acc_norm_stderr": 0.04351941398892446 |
|
}, |
|
"harness|ko_mmlu_conceptual_physics|5": { |
|
"acc": 0.43829787234042555, |
|
"acc_stderr": 0.03243618636108102, |
|
"acc_norm": 0.43829787234042555, |
|
"acc_norm_stderr": 0.03243618636108102 |
|
}, |
|
"harness|ko_mmlu_virology|5": { |
|
"acc": 0.41566265060240964, |
|
"acc_stderr": 0.03836722176598052, |
|
"acc_norm": 0.41566265060240964, |
|
"acc_norm_stderr": 0.03836722176598052 |
|
}, |
|
"harness|ko_mmlu_philosophy|5": { |
|
"acc": 0.47266881028938906, |
|
"acc_stderr": 0.02835563356832818, |
|
"acc_norm": 0.47266881028938906, |
|
"acc_norm_stderr": 0.02835563356832818 |
|
}, |
|
"harness|ko_mmlu_human_aging|5": { |
|
"acc": 0.484304932735426, |
|
"acc_stderr": 0.0335412657542081, |
|
"acc_norm": 0.484304932735426, |
|
"acc_norm_stderr": 0.0335412657542081 |
|
}, |
|
"harness|ko_mmlu_human_sexuality|5": { |
|
"acc": 0.4580152671755725, |
|
"acc_stderr": 0.04369802690578756, |
|
"acc_norm": 0.4580152671755725, |
|
"acc_norm_stderr": 0.04369802690578756 |
|
}, |
|
"harness|ko_mmlu_medical_genetics|5": { |
|
"acc": 0.38, |
|
"acc_stderr": 0.048783173121456316, |
|
"acc_norm": 0.38, |
|
"acc_norm_stderr": 0.048783173121456316 |
|
}, |
|
"harness|ko_mmlu_high_school_geography|5": { |
|
"acc": 0.601010101010101, |
|
"acc_stderr": 0.03488901616852731, |
|
"acc_norm": 0.601010101010101, |
|
"acc_norm_stderr": 0.03488901616852731 |
|
}, |
|
"harness|ko_mmlu_electrical_engineering|5": { |
|
"acc": 0.503448275862069, |
|
"acc_stderr": 0.04166567577101579, |
|
"acc_norm": 0.503448275862069, |
|
"acc_norm_stderr": 0.04166567577101579 |
|
}, |
|
"harness|ko_mmlu_college_physics|5": { |
|
"acc": 0.28431372549019607, |
|
"acc_stderr": 0.04488482852329017, |
|
"acc_norm": 0.28431372549019607, |
|
"acc_norm_stderr": 0.04488482852329017 |
|
}, |
|
"harness|ko_mmlu_high_school_microeconomics|5": { |
|
"acc": 0.47478991596638653, |
|
"acc_stderr": 0.0324371805513741, |
|
"acc_norm": 0.47478991596638653, |
|
"acc_norm_stderr": 0.0324371805513741 |
|
}, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": { |
|
"acc": 0.46923076923076923, |
|
"acc_stderr": 0.025302958890850154, |
|
"acc_norm": 0.46923076923076923, |
|
"acc_norm_stderr": 0.025302958890850154 |
|
}, |
|
"harness|ko_mmlu_computer_security|5": { |
|
"acc": 0.52, |
|
"acc_stderr": 0.05021167315686779, |
|
"acc_norm": 0.52, |
|
"acc_norm_stderr": 0.05021167315686779 |
|
}, |
|
"harness|ko_mmlu_global_facts|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|ko_mmlu_jurisprudence|5": { |
|
"acc": 0.5, |
|
"acc_stderr": 0.04833682445228318, |
|
"acc_norm": 0.5, |
|
"acc_norm_stderr": 0.04833682445228318 |
|
}, |
|
"harness|ko_mmlu_high_school_chemistry|5": { |
|
"acc": 0.42857142857142855, |
|
"acc_stderr": 0.03481904844438803, |
|
"acc_norm": 0.42857142857142855, |
|
"acc_norm_stderr": 0.03481904844438803 |
|
}, |
|
"harness|ko_mmlu_high_school_biology|5": { |
|
"acc": 0.4806451612903226, |
|
"acc_stderr": 0.028422687404312107, |
|
"acc_norm": 0.4806451612903226, |
|
"acc_norm_stderr": 0.028422687404312107 |
|
}, |
|
"harness|ko_mmlu_marketing|5": { |
|
"acc": 0.7478632478632479, |
|
"acc_stderr": 0.028447965476231022, |
|
"acc_norm": 0.7478632478632479, |
|
"acc_norm_stderr": 0.028447965476231022 |
|
}, |
|
"harness|ko_mmlu_clinical_knowledge|5": { |
|
"acc": 0.4830188679245283, |
|
"acc_stderr": 0.030755120364119905, |
|
"acc_norm": 0.4830188679245283, |
|
"acc_norm_stderr": 0.030755120364119905 |
|
}, |
|
"harness|ko_mmlu_public_relations|5": { |
|
"acc": 0.5272727272727272, |
|
"acc_stderr": 0.04782001791380061, |
|
"acc_norm": 0.5272727272727272, |
|
"acc_norm_stderr": 0.04782001791380061 |
|
}, |
|
"harness|ko_mmlu_high_school_mathematics|5": { |
|
"acc": 0.2851851851851852, |
|
"acc_stderr": 0.027528599210340492, |
|
"acc_norm": 0.2851851851851852, |
|
"acc_norm_stderr": 0.027528599210340492 |
|
}, |
|
"harness|ko_mmlu_high_school_physics|5": { |
|
"acc": 0.3509933774834437, |
|
"acc_stderr": 0.03896981964257375, |
|
"acc_norm": 0.3509933774834437, |
|
"acc_norm_stderr": 0.03896981964257375 |
|
}, |
|
"harness|ko_mmlu_sociology|5": { |
|
"acc": 0.6218905472636815, |
|
"acc_stderr": 0.034288678487786564, |
|
"acc_norm": 0.6218905472636815, |
|
"acc_norm_stderr": 0.034288678487786564 |
|
}, |
|
"harness|ko_mmlu_college_medicine|5": { |
|
"acc": 0.4393063583815029, |
|
"acc_stderr": 0.03784271932887467, |
|
"acc_norm": 0.4393063583815029, |
|
"acc_norm_stderr": 0.03784271932887467 |
|
}, |
|
"harness|ko_mmlu_elementary_mathematics|5": { |
|
"acc": 0.38095238095238093, |
|
"acc_stderr": 0.025010749116137595, |
|
"acc_norm": 0.38095238095238093, |
|
"acc_norm_stderr": 0.025010749116137595 |
|
}, |
|
"harness|ko_mmlu_college_biology|5": { |
|
"acc": 0.3888888888888889, |
|
"acc_stderr": 0.04076663253918567, |
|
"acc_norm": 0.3888888888888889, |
|
"acc_norm_stderr": 0.04076663253918567 |
|
}, |
|
"harness|ko_mmlu_college_chemistry|5": { |
|
"acc": 0.39, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.39, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|ko_mmlu_us_foreign_policy|5": { |
|
"acc": 0.63, |
|
"acc_stderr": 0.048523658709391, |
|
"acc_norm": 0.63, |
|
"acc_norm_stderr": 0.048523658709391 |
|
}, |
|
"harness|ko_mmlu_moral_disputes|5": { |
|
"acc": 0.5202312138728323, |
|
"acc_stderr": 0.026897049996382868, |
|
"acc_norm": 0.5202312138728323, |
|
"acc_norm_stderr": 0.026897049996382868 |
|
}, |
|
"harness|ko_mmlu_logical_fallacies|5": { |
|
"acc": 0.44171779141104295, |
|
"acc_stderr": 0.03901591825836184, |
|
"acc_norm": 0.44171779141104295, |
|
"acc_norm_stderr": 0.03901591825836184 |
|
}, |
|
"harness|ko_mmlu_prehistory|5": { |
|
"acc": 0.4382716049382716, |
|
"acc_stderr": 0.027607914087400473, |
|
"acc_norm": 0.4382716049382716, |
|
"acc_norm_stderr": 0.027607914087400473 |
|
}, |
|
"harness|ko_mmlu_college_mathematics|5": { |
|
"acc": 0.37, |
|
"acc_stderr": 0.04852365870939099, |
|
"acc_norm": 0.37, |
|
"acc_norm_stderr": 0.04852365870939099 |
|
}, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": { |
|
"acc": 0.49740932642487046, |
|
"acc_stderr": 0.03608390745384487, |
|
"acc_norm": 0.49740932642487046, |
|
"acc_norm_stderr": 0.03608390745384487 |
|
}, |
|
"harness|ko_mmlu_econometrics|5": { |
|
"acc": 0.2631578947368421, |
|
"acc_stderr": 0.04142439719489358, |
|
"acc_norm": 0.2631578947368421, |
|
"acc_norm_stderr": 0.04142439719489358 |
|
}, |
|
"harness|ko_mmlu_high_school_psychology|5": { |
|
"acc": 0.48440366972477067, |
|
"acc_stderr": 0.02142689153920805, |
|
"acc_norm": 0.48440366972477067, |
|
"acc_norm_stderr": 0.02142689153920805 |
|
}, |
|
"harness|ko_mmlu_formal_logic|5": { |
|
"acc": 0.3412698412698413, |
|
"acc_stderr": 0.04240799327574924, |
|
"acc_norm": 0.3412698412698413, |
|
"acc_norm_stderr": 0.04240799327574924 |
|
}, |
|
"harness|ko_mmlu_nutrition|5": { |
|
"acc": 0.5294117647058824, |
|
"acc_stderr": 0.028580341065138296, |
|
"acc_norm": 0.5294117647058824, |
|
"acc_norm_stderr": 0.028580341065138296 |
|
}, |
|
"harness|ko_mmlu_business_ethics|5": { |
|
"acc": 0.51, |
|
"acc_stderr": 0.05024183937956912, |
|
"acc_norm": 0.51, |
|
"acc_norm_stderr": 0.05024183937956912 |
|
}, |
|
"harness|ko_mmlu_international_law|5": { |
|
"acc": 0.6776859504132231, |
|
"acc_stderr": 0.042664163633521664, |
|
"acc_norm": 0.6776859504132231, |
|
"acc_norm_stderr": 0.042664163633521664 |
|
}, |
|
"harness|ko_mmlu_astronomy|5": { |
|
"acc": 0.506578947368421, |
|
"acc_stderr": 0.040685900502249704, |
|
"acc_norm": 0.506578947368421, |
|
"acc_norm_stderr": 0.040685900502249704 |
|
}, |
|
"harness|ko_mmlu_professional_psychology|5": { |
|
"acc": 0.4068627450980392, |
|
"acc_stderr": 0.019873802005061177, |
|
"acc_norm": 0.4068627450980392, |
|
"acc_norm_stderr": 0.019873802005061177 |
|
}, |
|
"harness|ko_mmlu_professional_accounting|5": { |
|
"acc": 0.375886524822695, |
|
"acc_stderr": 0.028893955412115882, |
|
"acc_norm": 0.375886524822695, |
|
"acc_norm_stderr": 0.028893955412115882 |
|
}, |
|
"harness|ko_mmlu_machine_learning|5": { |
|
"acc": 0.4107142857142857, |
|
"acc_stderr": 0.04669510663875192, |
|
"acc_norm": 0.4107142857142857, |
|
"acc_norm_stderr": 0.04669510663875192 |
|
}, |
|
"harness|ko_mmlu_high_school_statistics|5": { |
|
"acc": 0.4166666666666667, |
|
"acc_stderr": 0.033622774366080424, |
|
"acc_norm": 0.4166666666666667, |
|
"acc_norm_stderr": 0.033622774366080424 |
|
}, |
|
"harness|ko_mmlu_moral_scenarios|5": { |
|
"acc": 0.27932960893854747, |
|
"acc_stderr": 0.015005762446786173, |
|
"acc_norm": 0.27932960893854747, |
|
"acc_norm_stderr": 0.015005762446786173 |
|
}, |
|
"harness|ko_mmlu_college_computer_science|5": { |
|
"acc": 0.45, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.45, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|ko_mmlu_high_school_computer_science|5": { |
|
"acc": 0.57, |
|
"acc_stderr": 0.04975698519562428, |
|
"acc_norm": 0.57, |
|
"acc_norm_stderr": 0.04975698519562428 |
|
}, |
|
"harness|ko_mmlu_professional_medicine|5": { |
|
"acc": 0.4338235294117647, |
|
"acc_stderr": 0.030105636570016636, |
|
"acc_norm": 0.4338235294117647, |
|
"acc_norm_stderr": 0.030105636570016636 |
|
}, |
|
"harness|ko_mmlu_security_studies|5": { |
|
"acc": 0.44081632653061226, |
|
"acc_stderr": 0.03178419114175363, |
|
"acc_norm": 0.44081632653061226, |
|
"acc_norm_stderr": 0.03178419114175363 |
|
}, |
|
"harness|ko_mmlu_high_school_world_history|5": { |
|
"acc": 0.6160337552742616, |
|
"acc_stderr": 0.031658678064106674, |
|
"acc_norm": 0.6160337552742616, |
|
"acc_norm_stderr": 0.031658678064106674 |
|
}, |
|
"harness|ko_mmlu_professional_law|5": { |
|
"acc": 0.3428943937418514, |
|
"acc_stderr": 0.012123463271585895, |
|
"acc_norm": 0.3428943937418514, |
|
"acc_norm_stderr": 0.012123463271585895 |
|
}, |
|
"harness|ko_mmlu_high_school_us_history|5": { |
|
"acc": 0.45588235294117646, |
|
"acc_stderr": 0.03495624522015474, |
|
"acc_norm": 0.45588235294117646, |
|
"acc_norm_stderr": 0.03495624522015474 |
|
}, |
|
"harness|ko_mmlu_high_school_european_history|5": { |
|
"acc": 0.4121212121212121, |
|
"acc_stderr": 0.03843566993588718, |
|
"acc_norm": 0.4121212121212121, |
|
"acc_norm_stderr": 0.03843566993588718 |
|
}, |
|
"harness|ko_truthfulqa_mc|0": { |
|
"mc1": 0.2937576499388005, |
|
"mc1_stderr": 0.015945068581236614, |
|
"mc2": 0.4623692353701492, |
|
"mc2_stderr": 0.015502597273810991 |
|
}, |
|
"harness|ko_commongen_v2|2": { |
|
"acc": 0.3860684769775679, |
|
"acc_stderr": 0.016738130760321743, |
|
"acc_norm": 0.4722550177095632, |
|
"acc_norm_stderr": 0.017163867979456012 |
|
} |
|
}, |
|
"versions": { |
|
"all": 0, |
|
"harness|ko_arc_challenge|25": 0, |
|
"harness|ko_hellaswag|10": 0, |
|
"harness|ko_mmlu_world_religions|5": 1, |
|
"harness|ko_mmlu_management|5": 1, |
|
"harness|ko_mmlu_miscellaneous|5": 1, |
|
"harness|ko_mmlu_anatomy|5": 1, |
|
"harness|ko_mmlu_abstract_algebra|5": 1, |
|
"harness|ko_mmlu_conceptual_physics|5": 1, |
|
"harness|ko_mmlu_virology|5": 1, |
|
"harness|ko_mmlu_philosophy|5": 1, |
|
"harness|ko_mmlu_human_aging|5": 1, |
|
"harness|ko_mmlu_human_sexuality|5": 1, |
|
"harness|ko_mmlu_medical_genetics|5": 1, |
|
"harness|ko_mmlu_high_school_geography|5": 1, |
|
"harness|ko_mmlu_electrical_engineering|5": 1, |
|
"harness|ko_mmlu_college_physics|5": 1, |
|
"harness|ko_mmlu_high_school_microeconomics|5": 1, |
|
"harness|ko_mmlu_high_school_macroeconomics|5": 1, |
|
"harness|ko_mmlu_computer_security|5": 1, |
|
"harness|ko_mmlu_global_facts|5": 1, |
|
"harness|ko_mmlu_jurisprudence|5": 1, |
|
"harness|ko_mmlu_high_school_chemistry|5": 1, |
|
"harness|ko_mmlu_high_school_biology|5": 1, |
|
"harness|ko_mmlu_marketing|5": 1, |
|
"harness|ko_mmlu_clinical_knowledge|5": 1, |
|
"harness|ko_mmlu_public_relations|5": 1, |
|
"harness|ko_mmlu_high_school_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_physics|5": 1, |
|
"harness|ko_mmlu_sociology|5": 1, |
|
"harness|ko_mmlu_college_medicine|5": 1, |
|
"harness|ko_mmlu_elementary_mathematics|5": 1, |
|
"harness|ko_mmlu_college_biology|5": 1, |
|
"harness|ko_mmlu_college_chemistry|5": 1, |
|
"harness|ko_mmlu_us_foreign_policy|5": 1, |
|
"harness|ko_mmlu_moral_disputes|5": 1, |
|
"harness|ko_mmlu_logical_fallacies|5": 1, |
|
"harness|ko_mmlu_prehistory|5": 1, |
|
"harness|ko_mmlu_college_mathematics|5": 1, |
|
"harness|ko_mmlu_high_school_government_and_politics|5": 1, |
|
"harness|ko_mmlu_econometrics|5": 1, |
|
"harness|ko_mmlu_high_school_psychology|5": 1, |
|
"harness|ko_mmlu_formal_logic|5": 1, |
|
"harness|ko_mmlu_nutrition|5": 1, |
|
"harness|ko_mmlu_business_ethics|5": 1, |
|
"harness|ko_mmlu_international_law|5": 1, |
|
"harness|ko_mmlu_astronomy|5": 1, |
|
"harness|ko_mmlu_professional_psychology|5": 1, |
|
"harness|ko_mmlu_professional_accounting|5": 1, |
|
"harness|ko_mmlu_machine_learning|5": 1, |
|
"harness|ko_mmlu_high_school_statistics|5": 1, |
|
"harness|ko_mmlu_moral_scenarios|5": 1, |
|
"harness|ko_mmlu_college_computer_science|5": 1, |
|
"harness|ko_mmlu_high_school_computer_science|5": 1, |
|
"harness|ko_mmlu_professional_medicine|5": 1, |
|
"harness|ko_mmlu_security_studies|5": 1, |
|
"harness|ko_mmlu_high_school_world_history|5": 1, |
|
"harness|ko_mmlu_professional_law|5": 1, |
|
"harness|ko_mmlu_high_school_us_history|5": 1, |
|
"harness|ko_mmlu_high_school_european_history|5": 1, |
|
"harness|ko_truthfulqa_mc|0": 0, |
|
"harness|ko_commongen_v2|2": 1 |
|
}, |
|
"config_general": { |
|
"model_name": "Ja3ck/Mistral-instruct-Y24-DPO", |
|
"model_sha": "5cadddfbeeac1dd2be25ea036d8623968e987f3c", |
|
"model_dtype": "torch.float16", |
|
"lighteval_sha": "", |
|
"num_few_shot_default": 0, |
|
"num_fewshot_seeds": 1, |
|
"override_batch_size": 1, |
|
"max_samples": null |
|
} |
|
} |