{ "config_general": { "lighteval_sha": "", "num_few_shot_default": 0, "num_fewshot_seeds": 1, "override_batch_size": "auto:6", "max_samples": "null", "job_id": "", "model_name": "galatolo/cerbero-7b", "model_sha": "", "model_dtype": "torch.bfloat16", "model_size": "" }, "results": { "harness|hellaswag|10": { "acc,none": 0.6113324039036049, "acc_stderr,none": 0.0048645132621943105, "acc_norm,none": 0.8103963353913562, "acc_norm_stderr,none": 0.003911862797736137, "alias": "hellaswag" }, "harness|truthfulqa_mc2_m_es|0": { "acc,none": 0.27122940430925224, "acc_stderr,none": 0.015838013071561328, "alias": "truthfulqa_mc2_m_es" }, "harness|belebele_deu_Latn|5": { "acc,none": 0.7133333333333334, "acc_stderr,none": 0.015081863703787807, "acc_norm,none": 0.7133333333333334, "acc_norm_stderr,none": 0.015081863703787807, "alias": "belebele_deu_Latn" }, "harness|belebele_eng_Latn|5": { "acc,none": 0.8277777777777777, "acc_stderr,none": 0.012592780405781533, "acc_norm,none": 0.8277777777777777, "acc_norm_stderr,none": 0.012592780405781533, "alias": "belebele_eng_Latn" }, "harness|hellaswag_it|10": { "acc,none": 0.4778635918633743, "acc_stderr,none": 0.005210014744449426, "acc_norm,none": 0.6315674970085935, "acc_norm_stderr,none": 0.005031342082128735, "alias": "hellaswag_it" }, "harness|belebele_fra_Latn|5": { "acc,none": 0.7355555555555555, "acc_stderr,none": 0.014709405413413123, "acc_norm,none": 0.7355555555555555, "acc_norm_stderr,none": 0.014709405413413123, "alias": "belebele_fra_Latn" }, "harness|mmlu_m_de|5": { "acc,none": 0.5061849449389048, "acc_stderr,none": 0.004342243233169411, "alias": "mmlu_m_de" }, "harness|belebele_spa_Latn|5": { "acc,none": 0.7211111111111111, "acc_stderr,none": 0.014956736888683087, "acc_norm,none": 0.7211111111111111, "acc_norm_stderr,none": 0.014956736888683087, "alias": "belebele_spa_Latn" }, "harness|hellaswag_es|10": { "acc,none": 0.47151696180925967, "acc_stderr,none": 0.005156142104462169, "acc_norm,none": 0.6366545764881587, "acc_norm_stderr,none": 0.004967895723013549, "alias": "hellaswag_es" }, "harness|truthfulqa_mc2_m_it|0": { "acc,none": 0.30268199233716475, "acc_stderr,none": 0.016428781581749367, "alias": "truthfulqa_mc2_m_it" }, "harness|arc_challenge|25": { "acc,none": 0.5699658703071673, "acc_stderr,none": 0.014467631559137991, "acc_norm,none": 0.613481228668942, "acc_norm_stderr,none": 0.014230084761910474, "alias": "arc_challenge" }, "harness|arc_challenge_m_fr|25": { "acc,none": 0.42172797262617623, "acc_stderr,none": 0.014449768025902365, "acc_norm,none": 0.4627887082976903, "acc_norm_stderr,none": 0.014589571001051864, "alias": "arc_challenge_m_fr" }, "harness|hellaswag_fr|10": { "acc,none": 0.43842364532019706, "acc_stderr,none": 0.005135085885550969, "acc_norm,none": 0.6244377811094453, "acc_norm_stderr,none": 0.0050116629385795605, "alias": "hellaswag_fr" }, "harness|hendrycksTest|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-humanities|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-formal_logic|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_european_history|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_us_history|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_world_history|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-international_law|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-jurisprudence|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-logical_fallacies|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-moral_disputes|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-moral_scenarios|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-philosophy|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-prehistory|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-professional_law|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-world_religions|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-other|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-business_ethics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-clinical_knowledge|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_medicine|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-global_facts|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-human_aging|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-management|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-marketing|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-medical_genetics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-miscellaneous|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-nutrition|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-professional_accounting|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-professional_medicine|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-virology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-social_sciences|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-econometrics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_geography|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_government_and_politics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_macroeconomics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_microeconomics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_psychology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-human_sexuality|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-professional_psychology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-public_relations|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-security_studies|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-sociology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-us_foreign_policy|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-stem|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-abstract_algebra|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-anatomy|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-astronomy|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_biology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_chemistry|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_computer_science|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_mathematics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-college_physics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-computer_security|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-conceptual_physics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-electrical_engineering|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-elementary_mathematics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_biology|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_chemistry|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_computer_science|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_mathematics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_physics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-high_school_statistics|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hendrycksTest-machine_learning|5": { "acc,none": 0.6004842615012107, "acc_stderr,none": 0.13776064162347684, "alias": "mmlu" }, "harness|hellaswag_de|10": { "acc,none": 0.41417591801878734, "acc_stderr,none": 0.005089507655638241, "acc_norm,none": 0.5672502134927413, "acc_norm_stderr,none": 0.005119240170124385, "alias": "hellaswag_de" }, "harness|truthfulqa_mc2|0": { "acc,none": 0.4809105966670407, "acc_stderr,none": 0.015040316417034152, "alias": "truthfulqa_mc2" }, "harness|arc_challenge_m_es|25": { "acc,none": 0.45384615384615384, "acc_stderr,none": 0.014561448289640611, "acc_norm,none": 0.5008547008547009, "acc_norm_stderr,none": 0.014623863148445852, "alias": "arc_challenge_m_es" }, "harness|mmlu_m_fr|5": { "acc,none": 0.5164616912382552, "acc_stderr,none": 0.004367819439444167, "alias": "mmlu_m_fr" }, "harness|arc_challenge_m_de|25": { "acc,none": 0.41488451668092385, "acc_stderr,none": 0.014416604608618246, "acc_norm,none": 0.46877673224978617, "acc_norm_stderr,none": 0.014601589903081696, "alias": "arc_challenge_m_de" }, "harness|mmlu_m_es|5": { "acc,none": 0.5146992650367481, "acc_stderr,none": 0.004328309511475916, "alias": "mmlu_m_es" }, "harness|truthfulqa_mc2_m_fr|0": { "acc,none": 0.2909783989834816, "acc_stderr,none": 0.01620126162541297, "alias": "truthfulqa_mc2_m_fr" }, "harness|mmlu_m_it|5": { "acc,none": 0.5152224824355972, "acc_stderr,none": 0.0043440044239749604, "alias": "mmlu_m_it" }, "harness|arc_challenge_m_it|25": { "acc,none": 0.4833190761334474, "acc_stderr,none": 0.01462199936373081, "acc_norm,none": 0.5226689478186484, "acc_norm_stderr,none": 0.014615099353534883, "alias": "arc_challenge_m_it" }, "harness|belebele_ita_Latn|5": { "acc,none": 0.7177777777777777, "acc_stderr,none": 0.015011039301019264, "acc_norm,none": 0.7177777777777777, "acc_norm_stderr,none": 0.015011039301019264, "alias": "belebele_ita_Latn" }, "harness|truthfulqa_mc2_m_de|0": { "acc,none": 0.2715736040609137, "acc_stderr,none": 0.0158543794886016, "alias": "truthfulqa_mc2_m_de" } }, "versions": { "harness|hellaswag|10": 1.0, "harness|truthfulqa_mc2_m_es|0": "Yaml", "harness|belebele_deu_Latn|5": 0.0, "harness|belebele_eng_Latn|5": 0.0, "harness|hellaswag_it|10": 1.0, "harness|belebele_fra_Latn|5": 0.0, "harness|mmlu_m_de|5": "Yaml", "harness|belebele_spa_Latn|5": 0.0, "harness|hellaswag_es|10": 1.0, "harness|truthfulqa_mc2_m_it|0": "Yaml", "harness|arc_challenge|25": 1.0, "harness|arc_challenge_m_fr|25": 1.0, "harness|hellaswag_fr|10": 1.0, "harness|hendrycksTest|5": "N/A", "harness|hendrycksTest-humanities|5": "N/A", "harness|hendrycksTest-formal_logic|5": "N/A", "harness|hendrycksTest-high_school_european_history|5": "N/A", "harness|hendrycksTest-high_school_us_history|5": "N/A", "harness|hendrycksTest-high_school_world_history|5": "N/A", "harness|hendrycksTest-international_law|5": "N/A", "harness|hendrycksTest-jurisprudence|5": "N/A", "harness|hendrycksTest-logical_fallacies|5": "N/A", "harness|hendrycksTest-moral_disputes|5": "N/A", "harness|hendrycksTest-moral_scenarios|5": "N/A", "harness|hendrycksTest-philosophy|5": "N/A", "harness|hendrycksTest-prehistory|5": "N/A", "harness|hendrycksTest-professional_law|5": "N/A", "harness|hendrycksTest-world_religions|5": "N/A", "harness|hendrycksTest-other|5": "N/A", "harness|hendrycksTest-business_ethics|5": "N/A", "harness|hendrycksTest-clinical_knowledge|5": "N/A", "harness|hendrycksTest-college_medicine|5": "N/A", "harness|hendrycksTest-global_facts|5": "N/A", "harness|hendrycksTest-human_aging|5": "N/A", "harness|hendrycksTest-management|5": "N/A", "harness|hendrycksTest-marketing|5": "N/A", "harness|hendrycksTest-medical_genetics|5": "N/A", "harness|hendrycksTest-miscellaneous|5": "N/A", "harness|hendrycksTest-nutrition|5": "N/A", "harness|hendrycksTest-professional_accounting|5": "N/A", "harness|hendrycksTest-professional_medicine|5": "N/A", "harness|hendrycksTest-virology|5": "N/A", "harness|hendrycksTest-social_sciences|5": "N/A", "harness|hendrycksTest-econometrics|5": "N/A", "harness|hendrycksTest-high_school_geography|5": "N/A", "harness|hendrycksTest-high_school_government_and_politics|5": "N/A", "harness|hendrycksTest-high_school_macroeconomics|5": "N/A", "harness|hendrycksTest-high_school_microeconomics|5": "N/A", "harness|hendrycksTest-high_school_psychology|5": "N/A", "harness|hendrycksTest-human_sexuality|5": "N/A", "harness|hendrycksTest-professional_psychology|5": "N/A", "harness|hendrycksTest-public_relations|5": "N/A", "harness|hendrycksTest-security_studies|5": "N/A", "harness|hendrycksTest-sociology|5": "N/A", "harness|hendrycksTest-us_foreign_policy|5": "N/A", "harness|hendrycksTest-stem|5": "N/A", "harness|hendrycksTest-abstract_algebra|5": "N/A", "harness|hendrycksTest-anatomy|5": "N/A", "harness|hendrycksTest-astronomy|5": "N/A", "harness|hendrycksTest-college_biology|5": "N/A", "harness|hendrycksTest-college_chemistry|5": "N/A", "harness|hendrycksTest-college_computer_science|5": "N/A", "harness|hendrycksTest-college_mathematics|5": "N/A", "harness|hendrycksTest-college_physics|5": "N/A", "harness|hendrycksTest-computer_security|5": "N/A", "harness|hendrycksTest-conceptual_physics|5": "N/A", "harness|hendrycksTest-electrical_engineering|5": "N/A", "harness|hendrycksTest-elementary_mathematics|5": "N/A", "harness|hendrycksTest-high_school_biology|5": "N/A", "harness|hendrycksTest-high_school_chemistry|5": "N/A", "harness|hendrycksTest-high_school_computer_science|5": "N/A", "harness|hendrycksTest-high_school_mathematics|5": "N/A", "harness|hendrycksTest-high_school_physics|5": "N/A", "harness|hendrycksTest-high_school_statistics|5": "N/A", "harness|hendrycksTest-machine_learning|5": "N/A", "harness|hellaswag_de|10": 1.0, "harness|truthfulqa_mc2|0": 2.0, "harness|arc_challenge_m_es|25": 1.0, "harness|mmlu_m_fr|5": "Yaml", "harness|arc_challenge_m_de|25": 1.0, "harness|mmlu_m_es|5": "Yaml", "harness|truthfulqa_mc2_m_fr|0": "Yaml", "harness|mmlu_m_it|5": "Yaml", "harness|arc_challenge_m_it|25": 1.0, "harness|belebele_ita_Latn|5": 0.0, "harness|truthfulqa_mc2_m_de|0": "Yaml" } }