{ "config_general": { "lighteval_sha": "?", "num_fewshot_seeds": 1, "override_batch_size": 1, "max_samples": null, "job_id": "", "start_time": 601.718819721, "end_time": 119780.857050082, "total_evaluation_time_secondes": "119179.138230361", "model_name": "CombinHorizon/YiSM-blossom5.1-34B-SLERP", "model_sha": "ebd8d6507623008567a0548cd0ff9e28cbd6a656", "model_dtype": "torch.bfloat16", "model_size": "64.05 GB", "config": null }, "results": { "community|acva:Algeria|0": { "acc_norm": 0.6564102564102564, "acc_norm_stderr": 0.03409627301409855 }, "community|acva:Ancient_Egypt|0": { "acc_norm": 0.48253968253968255, "acc_norm_stderr": 0.02819942275430029 }, "community|acva:Arab_Empire|0": { "acc_norm": 0.3886792452830189, "acc_norm_stderr": 0.030000485448675986 }, "community|acva:Arabic_Architecture|0": { "acc_norm": 0.6564102564102564, "acc_norm_stderr": 0.03409627301409854 }, "community|acva:Arabic_Art|0": { "acc_norm": 0.517948717948718, "acc_norm_stderr": 0.03587477098773825 }, "community|acva:Arabic_Astronomy|0": { "acc_norm": 0.5128205128205128, "acc_norm_stderr": 0.03588610523192216 }, "community|acva:Arabic_Calligraphy|0": { "acc_norm": 0.7568627450980392, "acc_norm_stderr": 0.0269164301739633 }, "community|acva:Arabic_Ceremony|0": { "acc_norm": 0.6270270270270271, "acc_norm_stderr": 0.03565109718452138 }, "community|acva:Arabic_Clothing|0": { "acc_norm": 0.6, "acc_norm_stderr": 0.035172622905632896 }, "community|acva:Arabic_Culture|0": { "acc_norm": 0.7743589743589744, "acc_norm_stderr": 0.030010921825357008 }, "community|acva:Arabic_Food|0": { "acc_norm": 0.5692307692307692, "acc_norm_stderr": 0.0355521325205876 }, "community|acva:Arabic_Funeral|0": { "acc_norm": 0.5368421052631579, "acc_norm_stderr": 0.05143087276324537 }, "community|acva:Arabic_Geography|0": { "acc_norm": 0.6137931034482759, "acc_norm_stderr": 0.04057324734419035 }, "community|acva:Arabic_History|0": { "acc_norm": 0.3128205128205128, "acc_norm_stderr": 0.033287550657248546 }, "community|acva:Arabic_Language_Origin|0": { "acc_norm": 0.7368421052631579, "acc_norm_stderr": 0.045418364592773264 }, "community|acva:Arabic_Literature|0": { "acc_norm": 0.4206896551724138, "acc_norm_stderr": 0.0411391498118926 }, "community|acva:Arabic_Math|0": { "acc_norm": 0.5692307692307692, "acc_norm_stderr": 0.03555213252058761 }, "community|acva:Arabic_Medicine|0": { "acc_norm": 0.7103448275862069, "acc_norm_stderr": 0.03780019230438014 }, "community|acva:Arabic_Music|0": { "acc_norm": 0.5251798561151079, "acc_norm_stderr": 0.04250882046969544 }, "community|acva:Arabic_Ornament|0": { "acc_norm": 0.676923076923077, "acc_norm_stderr": 0.03357544396403132 }, "community|acva:Arabic_Philosophy|0": { "acc_norm": 0.503448275862069, "acc_norm_stderr": 0.041665675771015785 }, "community|acva:Arabic_Physics_and_Chemistry|0": { "acc_norm": 0.6717948717948717, "acc_norm_stderr": 0.033712437824137076 }, "community|acva:Arabic_Wedding|0": { "acc_norm": 0.5743589743589743, "acc_norm_stderr": 0.03549871080367707 }, "community|acva:Bahrain|0": { "acc_norm": 0.6888888888888889, "acc_norm_stderr": 0.06979205927323111 }, "community|acva:Comoros|0": { "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.0752101433090355 }, "community|acva:Egypt_modern|0": { "acc_norm": 0.6105263157894737, "acc_norm_stderr": 0.05029529117145395 }, "community|acva:InfluenceFromAncientEgypt|0": { "acc_norm": 0.841025641025641, "acc_norm_stderr": 0.026252296135516585 }, "community|acva:InfluenceFromByzantium|0": { "acc_norm": 0.7172413793103448, "acc_norm_stderr": 0.037528339580033376 }, "community|acva:InfluenceFromChina|0": { "acc_norm": 0.6871794871794872, "acc_norm_stderr": 0.033287550657248546 }, "community|acva:InfluenceFromGreece|0": { "acc_norm": 0.7692307692307693, "acc_norm_stderr": 0.0302493752938313 }, "community|acva:InfluenceFromIslam|0": { "acc_norm": 0.8344827586206897, "acc_norm_stderr": 0.03097055996622407 }, "community|acva:InfluenceFromPersia|0": { "acc_norm": 0.9542857142857143, "acc_norm_stderr": 0.015833996924252337 }, "community|acva:InfluenceFromRome|0": { "acc_norm": 0.6102564102564103, "acc_norm_stderr": 0.03501424776256371 }, "community|acva:Iraq|0": { "acc_norm": 0.611764705882353, "acc_norm_stderr": 0.053174090822038264 }, "community|acva:Islam_Education|0": { "acc_norm": 0.7538461538461538, "acc_norm_stderr": 0.030927428371225682 }, "community|acva:Islam_branches_and_schools|0": { "acc_norm": 0.4514285714285714, "acc_norm_stderr": 0.03772562898529836 }, "community|acva:Islamic_law_system|0": { "acc_norm": 0.7538461538461538, "acc_norm_stderr": 0.030927428371225647 }, "community|acva:Jordan|0": { "acc_norm": 0.6, "acc_norm_stderr": 0.07385489458759965 }, "community|acva:Kuwait|0": { "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.07106690545187012 }, "community|acva:Lebanon|0": { "acc_norm": 0.6888888888888889, "acc_norm_stderr": 0.06979205927323111 }, "community|acva:Libya|0": { "acc_norm": 0.5555555555555556, "acc_norm_stderr": 0.07491109582924914 }, "community|acva:Mauritania|0": { "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.0752101433090355 }, "community|acva:Mesopotamia_civilization|0": { "acc_norm": 0.5419354838709678, "acc_norm_stderr": 0.040149187308210765 }, "community|acva:Morocco|0": { "acc_norm": 0.7555555555555555, "acc_norm_stderr": 0.06478835438717 }, "community|acva:Oman|0": { "acc_norm": 0.6, "acc_norm_stderr": 0.07385489458759965 }, "community|acva:Palestine|0": { "acc_norm": 0.6235294117647059, "acc_norm_stderr": 0.052863310306265295 }, "community|acva:Qatar|0": { "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.0752101433090355 }, "community|acva:Saudi_Arabia|0": { "acc_norm": 0.7435897435897436, "acc_norm_stderr": 0.03134970994274491 }, "community|acva:Somalia|0": { "acc_norm": 0.6666666666666666, "acc_norm_stderr": 0.07106690545187012 }, "community|acva:Sudan|0": { "acc_norm": 0.6222222222222222, "acc_norm_stderr": 0.07309112127323451 }, "community|acva:Syria|0": { "acc_norm": 0.6222222222222222, "acc_norm_stderr": 0.07309112127323451 }, "community|acva:Tunisia|0": { "acc_norm": 0.6, "acc_norm_stderr": 0.07385489458759965 }, "community|acva:United_Arab_Emirates|0": { "acc_norm": 0.7647058823529411, "acc_norm_stderr": 0.04628210543937906 }, "community|acva:Yemen|0": { "acc_norm": 0.3, "acc_norm_stderr": 0.15275252316519466 }, "community|acva:communication|0": { "acc_norm": 0.5357142857142857, "acc_norm_stderr": 0.0261761615204498 }, "community|acva:computer_and_phone|0": { "acc_norm": 0.6711864406779661, "acc_norm_stderr": 0.02739824728293585 }, "community|acva:daily_life|0": { "acc_norm": 0.6023738872403561, "acc_norm_stderr": 0.02669936168374338 }, "community|acva:entertainment|0": { "acc_norm": 0.7457627118644068, "acc_norm_stderr": 0.02539488736489469 }, "community|alghafa:mcq_exams_test_ar|0": { "acc_norm": 0.27827648114901254, "acc_norm_stderr": 0.01900580568193068 }, "community|alghafa:meta_ar_dialects|0": { "acc_norm": 0.2978683966635774, "acc_norm_stderr": 0.006226815581230629 }, "community|alghafa:meta_ar_msa|0": { "acc_norm": 0.33631284916201115, "acc_norm_stderr": 0.0158010037291459 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { "acc_norm": 0.5333333333333333, "acc_norm_stderr": 0.05799451149344531 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { "acc_norm": 0.5533333333333333, "acc_norm_stderr": 0.04072790343023465 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { "acc_norm": 0.4, "acc_norm_stderr": 0.04013400372543905 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { "acc_norm": 0.6407754846779237, "acc_norm_stderr": 0.005366039381909045 }, "community|alghafa:multiple_choice_rating_sentiment_task|0": { "acc_norm": 0.42035029190992496, "acc_norm_stderr": 0.00637573294904214 }, "community|alghafa:multiple_choice_sentiment_task|0": { "acc_norm": 0.3540697674418605, "acc_norm_stderr": 0.011534513735394268 }, "community|arabic_exams|0": { "acc_norm": 0.2569832402234637, "acc_norm_stderr": 0.018874232932308742 }, "community|arabic_mmlu:abstract_algebra|0": { "acc_norm": 0.25, "acc_norm_stderr": 0.04351941398892446 }, "community|arabic_mmlu:anatomy|0": { "acc_norm": 0.34074074074074073, "acc_norm_stderr": 0.04094376269996794 }, "community|arabic_mmlu:astronomy|0": { "acc_norm": 0.3881578947368421, "acc_norm_stderr": 0.03965842097512744 }, "community|arabic_mmlu:business_ethics|0": { "acc_norm": 0.48, "acc_norm_stderr": 0.050211673156867795 }, "community|arabic_mmlu:clinical_knowledge|0": { "acc_norm": 0.33962264150943394, "acc_norm_stderr": 0.02914690474779834 }, "community|arabic_mmlu:college_biology|0": { "acc_norm": 0.2916666666666667, "acc_norm_stderr": 0.03800968060554858 }, "community|arabic_mmlu:college_chemistry|0": { "acc_norm": 0.28, "acc_norm_stderr": 0.04512608598542127 }, "community|arabic_mmlu:college_computer_science|0": { "acc_norm": 0.32, "acc_norm_stderr": 0.046882617226215034 }, "community|arabic_mmlu:college_mathematics|0": { "acc_norm": 0.35, "acc_norm_stderr": 0.0479372485441102 }, "community|arabic_mmlu:college_medicine|0": { "acc_norm": 0.2658959537572254, "acc_norm_stderr": 0.0336876293225943 }, "community|arabic_mmlu:college_physics|0": { "acc_norm": 0.2549019607843137, "acc_norm_stderr": 0.043364327079931785 }, "community|arabic_mmlu:computer_security|0": { "acc_norm": 0.47, "acc_norm_stderr": 0.05016135580465919 }, "community|arabic_mmlu:conceptual_physics|0": { "acc_norm": 0.2936170212765957, "acc_norm_stderr": 0.029771642712491227 }, "community|arabic_mmlu:econometrics|0": { "acc_norm": 0.2631578947368421, "acc_norm_stderr": 0.041424397194893624 }, "community|arabic_mmlu:electrical_engineering|0": { "acc_norm": 0.42758620689655175, "acc_norm_stderr": 0.041227371113703316 }, "community|arabic_mmlu:elementary_mathematics|0": { "acc_norm": 0.40476190476190477, "acc_norm_stderr": 0.025279850397404904 }, "community|arabic_mmlu:formal_logic|0": { "acc_norm": 0.25396825396825395, "acc_norm_stderr": 0.03893259610604674 }, "community|arabic_mmlu:global_facts|0": { "acc_norm": 0.29, "acc_norm_stderr": 0.04560480215720683 }, "community|arabic_mmlu:high_school_biology|0": { "acc_norm": 0.3225806451612903, "acc_norm_stderr": 0.02659308451657228 }, "community|arabic_mmlu:high_school_chemistry|0": { "acc_norm": 0.3054187192118227, "acc_norm_stderr": 0.03240661565868408 }, "community|arabic_mmlu:high_school_computer_science|0": { "acc_norm": 0.53, "acc_norm_stderr": 0.050161355804659205 }, "community|arabic_mmlu:high_school_european_history|0": { "acc_norm": 0.21212121212121213, "acc_norm_stderr": 0.031922715695482995 }, "community|arabic_mmlu:high_school_geography|0": { "acc_norm": 0.3434343434343434, "acc_norm_stderr": 0.03383201223244441 }, "community|arabic_mmlu:high_school_government_and_politics|0": { "acc_norm": 0.29533678756476683, "acc_norm_stderr": 0.032922966391551435 }, "community|arabic_mmlu:high_school_macroeconomics|0": { "acc_norm": 0.3153846153846154, "acc_norm_stderr": 0.02355964698318994 }, "community|arabic_mmlu:high_school_mathematics|0": { "acc_norm": 0.3296296296296296, "acc_norm_stderr": 0.028661201116524582 }, "community|arabic_mmlu:high_school_microeconomics|0": { "acc_norm": 0.2815126050420168, "acc_norm_stderr": 0.029213549414372167 }, "community|arabic_mmlu:high_school_physics|0": { "acc_norm": 0.271523178807947, "acc_norm_stderr": 0.036313298039696525 }, "community|arabic_mmlu:high_school_psychology|0": { "acc_norm": 0.27706422018348625, "acc_norm_stderr": 0.01918848259016953 }, "community|arabic_mmlu:high_school_statistics|0": { "acc_norm": 0.3194444444444444, "acc_norm_stderr": 0.03179876342176853 }, "community|arabic_mmlu:high_school_us_history|0": { "acc_norm": 0.22549019607843138, "acc_norm_stderr": 0.02933116229425172 }, "community|arabic_mmlu:high_school_world_history|0": { "acc_norm": 0.270042194092827, "acc_norm_stderr": 0.028900721906293433 }, "community|arabic_mmlu:human_aging|0": { "acc_norm": 0.33183856502242154, "acc_norm_stderr": 0.031602951437766785 }, "community|arabic_mmlu:human_sexuality|0": { "acc_norm": 0.2900763358778626, "acc_norm_stderr": 0.03980066246467765 }, "community|arabic_mmlu:international_law|0": { "acc_norm": 0.5041322314049587, "acc_norm_stderr": 0.045641987674327526 }, "community|arabic_mmlu:jurisprudence|0": { "acc_norm": 0.3888888888888889, "acc_norm_stderr": 0.0471282125742677 }, "community|arabic_mmlu:logical_fallacies|0": { "acc_norm": 0.37423312883435583, "acc_norm_stderr": 0.03802068102899615 }, "community|arabic_mmlu:machine_learning|0": { "acc_norm": 0.2857142857142857, "acc_norm_stderr": 0.04287858751340455 }, "community|arabic_mmlu:management|0": { "acc_norm": 0.34951456310679613, "acc_norm_stderr": 0.04721188506097173 }, "community|arabic_mmlu:marketing|0": { "acc_norm": 0.4017094017094017, "acc_norm_stderr": 0.03211693751051622 }, "community|arabic_mmlu:medical_genetics|0": { "acc_norm": 0.36, "acc_norm_stderr": 0.048241815132442176 }, "community|arabic_mmlu:miscellaneous|0": { "acc_norm": 0.3665389527458493, "acc_norm_stderr": 0.01723124462679703 }, "community|arabic_mmlu:moral_disputes|0": { "acc_norm": 0.37283236994219654, "acc_norm_stderr": 0.026033890613576284 }, "community|arabic_mmlu:moral_scenarios|0": { "acc_norm": 0.25251396648044694, "acc_norm_stderr": 0.014530330201468631 }, "community|arabic_mmlu:nutrition|0": { "acc_norm": 0.3660130718954248, "acc_norm_stderr": 0.027582811415159624 }, "community|arabic_mmlu:philosophy|0": { "acc_norm": 0.3536977491961415, "acc_norm_stderr": 0.02715520810320088 }, "community|arabic_mmlu:prehistory|0": { "acc_norm": 0.36728395061728397, "acc_norm_stderr": 0.026822801759507894 }, "community|arabic_mmlu:professional_accounting|0": { "acc_norm": 0.2872340425531915, "acc_norm_stderr": 0.02699219917306436 }, "community|arabic_mmlu:professional_law|0": { "acc_norm": 0.27249022164276404, "acc_norm_stderr": 0.011371658294311538 }, "community|arabic_mmlu:professional_medicine|0": { "acc_norm": 0.14705882352941177, "acc_norm_stderr": 0.021513964052859633 }, "community|arabic_mmlu:professional_psychology|0": { "acc_norm": 0.31699346405228757, "acc_norm_stderr": 0.018824219512706214 }, "community|arabic_mmlu:public_relations|0": { "acc_norm": 0.33636363636363636, "acc_norm_stderr": 0.04525393596302506 }, "community|arabic_mmlu:security_studies|0": { "acc_norm": 0.3795918367346939, "acc_norm_stderr": 0.03106721126287248 }, "community|arabic_mmlu:sociology|0": { "acc_norm": 0.3333333333333333, "acc_norm_stderr": 0.03333333333333334 }, "community|arabic_mmlu:us_foreign_policy|0": { "acc_norm": 0.58, "acc_norm_stderr": 0.049604496374885836 }, "community|arabic_mmlu:virology|0": { "acc_norm": 0.3253012048192771, "acc_norm_stderr": 0.03647168523683227 }, "community|arabic_mmlu:world_religions|0": { "acc_norm": 0.34502923976608185, "acc_norm_stderr": 0.03645981377388806 }, "community|arc_challenge_okapi_ar|0": { "acc_norm": 0.33706896551724136, "acc_norm_stderr": 0.013885190464608725 }, "community|arc_easy_ar|0": { "acc_norm": 0.31514382402707275, "acc_norm_stderr": 0.009557014121502784 }, "community|boolq_ar|0": { "acc_norm": 0.645398773006135, "acc_norm_stderr": 0.008379961135774872 }, "community|copa_ext_ar|0": { "acc_norm": 0.4777777777777778, "acc_norm_stderr": 0.05294752255076824 }, "community|hellaswag_okapi_ar|0": { "acc_norm": 0.26649220368553045, "acc_norm_stderr": 0.004617000837669436 }, "community|openbook_qa_ext_ar|0": { "acc_norm": 0.397979797979798, "acc_norm_stderr": 0.02202280187415247 }, "community|piqa_ar|0": { "acc_norm": 0.5351882160392799, "acc_norm_stderr": 0.011652768039224288 }, "community|race_ar|0": { "acc_norm": 0.34408602150537637, "acc_norm_stderr": 0.0067673962834595 }, "community|sciq_ar|0": { "acc_norm": 0.535678391959799, "acc_norm_stderr": 0.015818610017459517 }, "community|toxigen_ar|0": { "acc_norm": 0.4716577540106952, "acc_norm_stderr": 0.016334194407777027 }, "lighteval|xstory_cloze:ar|0": { "acc": 0.5698213103904699, "acc_stderr": 0.012741052817471078 }, "community|acva:_average|0": { "acc_norm": 0.6238816361277529, "acc_norm_stderr": 0.04637354479043959 }, "community|alghafa:_average|0": { "acc_norm": 0.42381332640788627, "acc_norm_stderr": 0.022574036634196853 }, "community|arabic_mmlu:_average|0": { "acc_norm": 0.3324814595653192, "acc_norm_stderr": 0.03488750666630585 }, "all": { "acc_norm": 0.4706248369928593, "acc_norm_stderr": 0.037498344445906724, "acc": 0.5698213103904699, "acc_stderr": 0.012741052817471078 } }, "versions": { "community|acva:Algeria|0": 0, "community|acva:Ancient_Egypt|0": 0, "community|acva:Arab_Empire|0": 0, "community|acva:Arabic_Architecture|0": 0, "community|acva:Arabic_Art|0": 0, "community|acva:Arabic_Astronomy|0": 0, "community|acva:Arabic_Calligraphy|0": 0, "community|acva:Arabic_Ceremony|0": 0, "community|acva:Arabic_Clothing|0": 0, "community|acva:Arabic_Culture|0": 0, "community|acva:Arabic_Food|0": 0, "community|acva:Arabic_Funeral|0": 0, "community|acva:Arabic_Geography|0": 0, "community|acva:Arabic_History|0": 0, "community|acva:Arabic_Language_Origin|0": 0, "community|acva:Arabic_Literature|0": 0, "community|acva:Arabic_Math|0": 0, "community|acva:Arabic_Medicine|0": 0, "community|acva:Arabic_Music|0": 0, "community|acva:Arabic_Ornament|0": 0, "community|acva:Arabic_Philosophy|0": 0, "community|acva:Arabic_Physics_and_Chemistry|0": 0, "community|acva:Arabic_Wedding|0": 0, "community|acva:Bahrain|0": 0, "community|acva:Comoros|0": 0, "community|acva:Egypt_modern|0": 0, "community|acva:InfluenceFromAncientEgypt|0": 0, "community|acva:InfluenceFromByzantium|0": 0, "community|acva:InfluenceFromChina|0": 0, "community|acva:InfluenceFromGreece|0": 0, "community|acva:InfluenceFromIslam|0": 0, "community|acva:InfluenceFromPersia|0": 0, "community|acva:InfluenceFromRome|0": 0, "community|acva:Iraq|0": 0, "community|acva:Islam_Education|0": 0, "community|acva:Islam_branches_and_schools|0": 0, "community|acva:Islamic_law_system|0": 0, "community|acva:Jordan|0": 0, "community|acva:Kuwait|0": 0, "community|acva:Lebanon|0": 0, "community|acva:Libya|0": 0, "community|acva:Mauritania|0": 0, "community|acva:Mesopotamia_civilization|0": 0, "community|acva:Morocco|0": 0, "community|acva:Oman|0": 0, "community|acva:Palestine|0": 0, "community|acva:Qatar|0": 0, "community|acva:Saudi_Arabia|0": 0, "community|acva:Somalia|0": 0, "community|acva:Sudan|0": 0, "community|acva:Syria|0": 0, "community|acva:Tunisia|0": 0, "community|acva:United_Arab_Emirates|0": 0, "community|acva:Yemen|0": 0, "community|acva:communication|0": 0, "community|acva:computer_and_phone|0": 0, "community|acva:daily_life|0": 0, "community|acva:entertainment|0": 0, "community|alghafa:mcq_exams_test_ar|0": 0, "community|alghafa:meta_ar_dialects|0": 0, "community|alghafa:meta_ar_msa|0": 0, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": 0, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": 0, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": 0, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": 0, "community|alghafa:multiple_choice_rating_sentiment_task|0": 0, "community|alghafa:multiple_choice_sentiment_task|0": 0, "community|arabic_exams|0": 0, "community|arabic_mmlu:abstract_algebra|0": 0, "community|arabic_mmlu:anatomy|0": 0, "community|arabic_mmlu:astronomy|0": 0, "community|arabic_mmlu:business_ethics|0": 0, "community|arabic_mmlu:clinical_knowledge|0": 0, "community|arabic_mmlu:college_biology|0": 0, "community|arabic_mmlu:college_chemistry|0": 0, "community|arabic_mmlu:college_computer_science|0": 0, "community|arabic_mmlu:college_mathematics|0": 0, "community|arabic_mmlu:college_medicine|0": 0, "community|arabic_mmlu:college_physics|0": 0, "community|arabic_mmlu:computer_security|0": 0, "community|arabic_mmlu:conceptual_physics|0": 0, "community|arabic_mmlu:econometrics|0": 0, "community|arabic_mmlu:electrical_engineering|0": 0, "community|arabic_mmlu:elementary_mathematics|0": 0, "community|arabic_mmlu:formal_logic|0": 0, "community|arabic_mmlu:global_facts|0": 0, "community|arabic_mmlu:high_school_biology|0": 0, "community|arabic_mmlu:high_school_chemistry|0": 0, "community|arabic_mmlu:high_school_computer_science|0": 0, "community|arabic_mmlu:high_school_european_history|0": 0, "community|arabic_mmlu:high_school_geography|0": 0, "community|arabic_mmlu:high_school_government_and_politics|0": 0, "community|arabic_mmlu:high_school_macroeconomics|0": 0, "community|arabic_mmlu:high_school_mathematics|0": 0, "community|arabic_mmlu:high_school_microeconomics|0": 0, "community|arabic_mmlu:high_school_physics|0": 0, "community|arabic_mmlu:high_school_psychology|0": 0, "community|arabic_mmlu:high_school_statistics|0": 0, "community|arabic_mmlu:high_school_us_history|0": 0, "community|arabic_mmlu:high_school_world_history|0": 0, "community|arabic_mmlu:human_aging|0": 0, "community|arabic_mmlu:human_sexuality|0": 0, "community|arabic_mmlu:international_law|0": 0, "community|arabic_mmlu:jurisprudence|0": 0, "community|arabic_mmlu:logical_fallacies|0": 0, "community|arabic_mmlu:machine_learning|0": 0, "community|arabic_mmlu:management|0": 0, "community|arabic_mmlu:marketing|0": 0, "community|arabic_mmlu:medical_genetics|0": 0, "community|arabic_mmlu:miscellaneous|0": 0, "community|arabic_mmlu:moral_disputes|0": 0, "community|arabic_mmlu:moral_scenarios|0": 0, "community|arabic_mmlu:nutrition|0": 0, "community|arabic_mmlu:philosophy|0": 0, "community|arabic_mmlu:prehistory|0": 0, "community|arabic_mmlu:professional_accounting|0": 0, "community|arabic_mmlu:professional_law|0": 0, "community|arabic_mmlu:professional_medicine|0": 0, "community|arabic_mmlu:professional_psychology|0": 0, "community|arabic_mmlu:public_relations|0": 0, "community|arabic_mmlu:security_studies|0": 0, "community|arabic_mmlu:sociology|0": 0, "community|arabic_mmlu:us_foreign_policy|0": 0, "community|arabic_mmlu:virology|0": 0, "community|arabic_mmlu:world_religions|0": 0, "community|arc_challenge_okapi_ar|0": 0, "community|arc_easy_ar|0": 0, "community|boolq_ar|0": 0, "community|copa_ext_ar|0": 0, "community|hellaswag_okapi_ar|0": 0, "community|openbook_qa_ext_ar|0": 0, "community|piqa_ar|0": 0, "community|race_ar|0": 0, "community|sciq_ar|0": 0, "community|toxigen_ar|0": 0, "lighteval|xstory_cloze:ar|0": 0 }, "config_tasks": { "community|acva:Algeria": { "name": "acva:Algeria", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Algeria", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Ancient_Egypt": { "name": "acva:Ancient_Egypt", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Ancient_Egypt", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 315, "effective_num_docs": 315, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arab_Empire": { "name": "acva:Arab_Empire", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arab_Empire", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Architecture": { "name": "acva:Arabic_Architecture", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Architecture", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Art": { "name": "acva:Arabic_Art", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Art", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Astronomy": { "name": "acva:Arabic_Astronomy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Astronomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Calligraphy": { "name": "acva:Arabic_Calligraphy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Calligraphy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 255, "effective_num_docs": 255, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Ceremony": { "name": "acva:Arabic_Ceremony", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Ceremony", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 185, "effective_num_docs": 185, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Clothing": { "name": "acva:Arabic_Clothing", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Clothing", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Culture": { "name": "acva:Arabic_Culture", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Culture", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Food": { "name": "acva:Arabic_Food", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Food", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Funeral": { "name": "acva:Arabic_Funeral", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Funeral", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Geography": { "name": "acva:Arabic_Geography", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Geography", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_History": { "name": "acva:Arabic_History", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_History", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Language_Origin": { "name": "acva:Arabic_Language_Origin", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Language_Origin", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Literature": { "name": "acva:Arabic_Literature", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Literature", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Math": { "name": "acva:Arabic_Math", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Math", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Medicine": { "name": "acva:Arabic_Medicine", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Music": { "name": "acva:Arabic_Music", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Music", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 139, "effective_num_docs": 139, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Ornament": { "name": "acva:Arabic_Ornament", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Ornament", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Philosophy": { "name": "acva:Arabic_Philosophy", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Philosophy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Physics_and_Chemistry": { "name": "acva:Arabic_Physics_and_Chemistry", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Physics_and_Chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Arabic_Wedding": { "name": "acva:Arabic_Wedding", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Arabic_Wedding", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Bahrain": { "name": "acva:Bahrain", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Bahrain", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Comoros": { "name": "acva:Comoros", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Comoros", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Egypt_modern": { "name": "acva:Egypt_modern", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Egypt_modern", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 95, "effective_num_docs": 95, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromAncientEgypt": { "name": "acva:InfluenceFromAncientEgypt", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromAncientEgypt", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromByzantium": { "name": "acva:InfluenceFromByzantium", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromByzantium", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromChina": { "name": "acva:InfluenceFromChina", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromChina", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromGreece": { "name": "acva:InfluenceFromGreece", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromGreece", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromIslam": { "name": "acva:InfluenceFromIslam", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromIslam", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromPersia": { "name": "acva:InfluenceFromPersia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromPersia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 175, "effective_num_docs": 175, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:InfluenceFromRome": { "name": "acva:InfluenceFromRome", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "InfluenceFromRome", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Iraq": { "name": "acva:Iraq", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Iraq", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islam_Education": { "name": "acva:Islam_Education", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islam_Education", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islam_branches_and_schools": { "name": "acva:Islam_branches_and_schools", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islam_branches_and_schools", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 175, "effective_num_docs": 175, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Islamic_law_system": { "name": "acva:Islamic_law_system", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Islamic_law_system", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Jordan": { "name": "acva:Jordan", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Jordan", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Kuwait": { "name": "acva:Kuwait", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Kuwait", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Lebanon": { "name": "acva:Lebanon", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Lebanon", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Libya": { "name": "acva:Libya", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Libya", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Mauritania": { "name": "acva:Mauritania", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Mauritania", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Mesopotamia_civilization": { "name": "acva:Mesopotamia_civilization", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Mesopotamia_civilization", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 155, "effective_num_docs": 155, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Morocco": { "name": "acva:Morocco", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Morocco", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Oman": { "name": "acva:Oman", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Oman", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Palestine": { "name": "acva:Palestine", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Palestine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Qatar": { "name": "acva:Qatar", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Qatar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Saudi_Arabia": { "name": "acva:Saudi_Arabia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Saudi_Arabia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 195, "effective_num_docs": 195, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Somalia": { "name": "acva:Somalia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Somalia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Sudan": { "name": "acva:Sudan", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Sudan", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Syria": { "name": "acva:Syria", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Syria", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Tunisia": { "name": "acva:Tunisia", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Tunisia", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 45, "effective_num_docs": 45, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:United_Arab_Emirates": { "name": "acva:United_Arab_Emirates", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "United_Arab_Emirates", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 85, "effective_num_docs": 85, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:Yemen": { "name": "acva:Yemen", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "Yemen", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 10, "effective_num_docs": 10, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:communication": { "name": "acva:communication", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "communication", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 364, "effective_num_docs": 364, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:computer_and_phone": { "name": "acva:computer_and_phone", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "computer_and_phone", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 295, "effective_num_docs": 295, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:daily_life": { "name": "acva:daily_life", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "daily_life", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 337, "effective_num_docs": 337, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|acva:entertainment": { "name": "acva:entertainment", "prompt_function": "acva", "hf_repo": "OALL/ACVA", "hf_subset": "entertainment", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 295, "effective_num_docs": 295, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:mcq_exams_test_ar": { "name": "alghafa:mcq_exams_test_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "mcq_exams_test_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 557, "effective_num_docs": 557, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:meta_ar_dialects": { "name": "alghafa:meta_ar_dialects", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "meta_ar_dialects", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 5395, "effective_num_docs": 5395, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:meta_ar_msa": { "name": "alghafa:meta_ar_msa", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "meta_ar_msa", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task": { "name": "alghafa:multiple_choice_facts_truefalse_balanced_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_facts_truefalse_balanced_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 75, "effective_num_docs": 75, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task": { "name": "alghafa:multiple_choice_grounded_statement_soqal_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_grounded_statement_soqal_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 150, "effective_num_docs": 150, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task": { "name": "alghafa:multiple_choice_grounded_statement_xglue_mlqa_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_grounded_statement_xglue_mlqa_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 150, "effective_num_docs": 150, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task": { "name": "alghafa:multiple_choice_rating_sentiment_no_neutral_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_rating_sentiment_no_neutral_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 7995, "effective_num_docs": 7995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_rating_sentiment_task": { "name": "alghafa:multiple_choice_rating_sentiment_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_rating_sentiment_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 5995, "effective_num_docs": 5995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|alghafa:multiple_choice_sentiment_task": { "name": "alghafa:multiple_choice_sentiment_task", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Native", "hf_subset": "multiple_choice_sentiment_task", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1720, "effective_num_docs": 1720, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_exams": { "name": "arabic_exams", "prompt_function": "arabic_exams", "hf_repo": "OALL/Arabic_EXAMS", "hf_subset": "default", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 537, "effective_num_docs": 537, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:abstract_algebra": { "name": "arabic_mmlu:abstract_algebra", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "abstract_algebra", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:anatomy": { "name": "arabic_mmlu:anatomy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "anatomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 135, "effective_num_docs": 135, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:astronomy": { "name": "arabic_mmlu:astronomy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "astronomy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 152, "effective_num_docs": 152, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:business_ethics": { "name": "arabic_mmlu:business_ethics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "business_ethics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:clinical_knowledge": { "name": "arabic_mmlu:clinical_knowledge", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "clinical_knowledge", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 265, "effective_num_docs": 265, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_biology": { "name": "arabic_mmlu:college_biology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_biology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 144, "effective_num_docs": 144, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_chemistry": { "name": "arabic_mmlu:college_chemistry", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_computer_science": { "name": "arabic_mmlu:college_computer_science", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_computer_science", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_mathematics": { "name": "arabic_mmlu:college_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_medicine": { "name": "arabic_mmlu:college_medicine", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 173, "effective_num_docs": 173, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:college_physics": { "name": "arabic_mmlu:college_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "college_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 102, "effective_num_docs": 102, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:computer_security": { "name": "arabic_mmlu:computer_security", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "computer_security", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:conceptual_physics": { "name": "arabic_mmlu:conceptual_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "conceptual_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 235, "effective_num_docs": 235, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:econometrics": { "name": "arabic_mmlu:econometrics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "econometrics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 114, "effective_num_docs": 114, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:electrical_engineering": { "name": "arabic_mmlu:electrical_engineering", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "electrical_engineering", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 145, "effective_num_docs": 145, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:elementary_mathematics": { "name": "arabic_mmlu:elementary_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "elementary_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 378, "effective_num_docs": 378, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:formal_logic": { "name": "arabic_mmlu:formal_logic", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "formal_logic", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 126, "effective_num_docs": 126, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:global_facts": { "name": "arabic_mmlu:global_facts", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "global_facts", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_biology": { "name": "arabic_mmlu:high_school_biology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_biology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 310, "effective_num_docs": 310, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_chemistry": { "name": "arabic_mmlu:high_school_chemistry", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_chemistry", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 203, "effective_num_docs": 203, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_computer_science": { "name": "arabic_mmlu:high_school_computer_science", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_computer_science", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_european_history": { "name": "arabic_mmlu:high_school_european_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_european_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 165, "effective_num_docs": 165, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_geography": { "name": "arabic_mmlu:high_school_geography", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_geography", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 198, "effective_num_docs": 198, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_government_and_politics": { "name": "arabic_mmlu:high_school_government_and_politics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_government_and_politics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 193, "effective_num_docs": 193, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_macroeconomics": { "name": "arabic_mmlu:high_school_macroeconomics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_macroeconomics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 390, "effective_num_docs": 390, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_mathematics": { "name": "arabic_mmlu:high_school_mathematics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_mathematics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 270, "effective_num_docs": 270, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_microeconomics": { "name": "arabic_mmlu:high_school_microeconomics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_microeconomics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 238, "effective_num_docs": 238, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_physics": { "name": "arabic_mmlu:high_school_physics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_physics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 151, "effective_num_docs": 151, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_psychology": { "name": "arabic_mmlu:high_school_psychology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_psychology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 545, "effective_num_docs": 545, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_statistics": { "name": "arabic_mmlu:high_school_statistics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_statistics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 216, "effective_num_docs": 216, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_us_history": { "name": "arabic_mmlu:high_school_us_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_us_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 204, "effective_num_docs": 204, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:high_school_world_history": { "name": "arabic_mmlu:high_school_world_history", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "high_school_world_history", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 237, "effective_num_docs": 237, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:human_aging": { "name": "arabic_mmlu:human_aging", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "human_aging", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 223, "effective_num_docs": 223, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:human_sexuality": { "name": "arabic_mmlu:human_sexuality", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "human_sexuality", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 131, "effective_num_docs": 131, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:international_law": { "name": "arabic_mmlu:international_law", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "international_law", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 121, "effective_num_docs": 121, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:jurisprudence": { "name": "arabic_mmlu:jurisprudence", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "jurisprudence", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 108, "effective_num_docs": 108, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:logical_fallacies": { "name": "arabic_mmlu:logical_fallacies", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "logical_fallacies", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 163, "effective_num_docs": 163, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:machine_learning": { "name": "arabic_mmlu:machine_learning", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "machine_learning", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 112, "effective_num_docs": 112, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:management": { "name": "arabic_mmlu:management", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "management", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 103, "effective_num_docs": 103, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:marketing": { "name": "arabic_mmlu:marketing", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "marketing", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 234, "effective_num_docs": 234, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:medical_genetics": { "name": "arabic_mmlu:medical_genetics", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "medical_genetics", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:miscellaneous": { "name": "arabic_mmlu:miscellaneous", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "miscellaneous", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 783, "effective_num_docs": 783, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:moral_disputes": { "name": "arabic_mmlu:moral_disputes", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "moral_disputes", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 346, "effective_num_docs": 346, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:moral_scenarios": { "name": "arabic_mmlu:moral_scenarios", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "moral_scenarios", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 895, "effective_num_docs": 895, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:nutrition": { "name": "arabic_mmlu:nutrition", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "nutrition", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 306, "effective_num_docs": 306, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:philosophy": { "name": "arabic_mmlu:philosophy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "philosophy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 311, "effective_num_docs": 311, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:prehistory": { "name": "arabic_mmlu:prehistory", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "prehistory", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 324, "effective_num_docs": 324, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_accounting": { "name": "arabic_mmlu:professional_accounting", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_accounting", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 282, "effective_num_docs": 282, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_law": { "name": "arabic_mmlu:professional_law", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_law", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1534, "effective_num_docs": 1534, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_medicine": { "name": "arabic_mmlu:professional_medicine", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_medicine", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 272, "effective_num_docs": 272, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:professional_psychology": { "name": "arabic_mmlu:professional_psychology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "professional_psychology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 612, "effective_num_docs": 612, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:public_relations": { "name": "arabic_mmlu:public_relations", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "public_relations", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 110, "effective_num_docs": 110, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:security_studies": { "name": "arabic_mmlu:security_studies", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "security_studies", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 245, "effective_num_docs": 245, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:sociology": { "name": "arabic_mmlu:sociology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "sociology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 201, "effective_num_docs": 201, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:us_foreign_policy": { "name": "arabic_mmlu:us_foreign_policy", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "us_foreign_policy", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 100, "effective_num_docs": 100, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:virology": { "name": "arabic_mmlu:virology", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "virology", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 166, "effective_num_docs": 166, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arabic_mmlu:world_religions": { "name": "arabic_mmlu:world_religions", "prompt_function": "mmlu_arabic", "hf_repo": "OALL/Arabic_MMLU", "hf_subset": "world_religions", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "dev" ], "evaluation_splits": [ "test" ], "few_shots_split": "dev", "few_shots_select": "sequential", "generation_size": -1, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 171, "effective_num_docs": 171, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arc_challenge_okapi_ar": { "name": "arc_challenge_okapi_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "arc_challenge_okapi_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1160, "effective_num_docs": 1160, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|arc_easy_ar": { "name": "arc_easy_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "arc_easy_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 2364, "effective_num_docs": 2364, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|boolq_ar": { "name": "boolq_ar", "prompt_function": "boolq_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "boolq_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 3260, "effective_num_docs": 3260, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|copa_ext_ar": { "name": "copa_ext_ar", "prompt_function": "copa_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "copa_ext_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 90, "effective_num_docs": 90, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|hellaswag_okapi_ar": { "name": "hellaswag_okapi_ar", "prompt_function": "hellaswag_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "hellaswag_okapi_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 9171, "effective_num_docs": 9171, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|openbook_qa_ext_ar": { "name": "openbook_qa_ext_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "openbook_qa_ext_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 495, "effective_num_docs": 495, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|piqa_ar": { "name": "piqa_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "piqa_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 1833, "effective_num_docs": 1833, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|race_ar": { "name": "race_ar", "prompt_function": "alghafa_prompt", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "race_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 4929, "effective_num_docs": 4929, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|sciq_ar": { "name": "sciq_ar", "prompt_function": "sciq_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "sciq_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 995, "effective_num_docs": 995, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "community|toxigen_ar": { "name": "toxigen_ar", "prompt_function": "toxigen_prompt_arabic", "hf_repo": "OALL/AlGhafa-Arabic-LLM-Benchmark-Translated", "hf_subset": "toxigen_ar", "metric": [ "loglikelihood_acc_norm" ], "hf_avail_splits": [ "test", "validation" ], "evaluation_splits": [ "test" ], "few_shots_split": "validation", "few_shots_select": "sequential", "generation_size": null, "stop_sequence": null, "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "community" ], "original_num_docs": 935, "effective_num_docs": 935, "trust_dataset": null, "must_remove_duplicate_docs": null, "version": 0 }, "lighteval|xstory_cloze:ar": { "name": "xstory_cloze:ar", "prompt_function": "storycloze", "hf_repo": "juletxara/xstory_cloze", "hf_subset": "ar", "metric": [ "loglikelihood_acc" ], "hf_avail_splits": [ "training", "eval" ], "evaluation_splits": [ "eval" ], "few_shots_split": null, "few_shots_select": null, "generation_size": -1, "stop_sequence": [ "\n" ], "output_regex": null, "num_samples": null, "frozen": false, "suite": [ "lighteval" ], "original_num_docs": 1511, "effective_num_docs": 1511, "trust_dataset": true, "must_remove_duplicate_docs": null, "version": 0 } }, "summary_tasks": { "community|acva:Algeria|0": { "hashes": { "hash_examples": "da5a3003cd46f6f9", "hash_full_prompts": "da5a3003cd46f6f9", "hash_input_tokens": "3fda880d984f4082", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Ancient_Egypt|0": { "hashes": { "hash_examples": "52d6f767fede195b", "hash_full_prompts": "52d6f767fede195b", "hash_input_tokens": "4be5ae51ada906a8", "hash_cont_tokens": "5f7d1751e6ad9399" }, "truncated": 0, "non_truncated": 315, "padded": 630, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arab_Empire|0": { "hashes": { "hash_examples": "8dacff6a79804a75", "hash_full_prompts": "8dacff6a79804a75", "hash_input_tokens": "b3d5feb6bf99253e", "hash_cont_tokens": "8783a9653c6992cc" }, "truncated": 0, "non_truncated": 265, "padded": 530, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Architecture|0": { "hashes": { "hash_examples": "df286cd862d9f6bb", "hash_full_prompts": "df286cd862d9f6bb", "hash_input_tokens": "e576ddd798f53203", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Art|0": { "hashes": { "hash_examples": "112883d764118a49", "hash_full_prompts": "112883d764118a49", "hash_input_tokens": "be306ec70280ed16", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Astronomy|0": { "hashes": { "hash_examples": "20dcdf2454bf8671", "hash_full_prompts": "20dcdf2454bf8671", "hash_input_tokens": "4c36847d6a113753", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Calligraphy|0": { "hashes": { "hash_examples": "3a9f9d1ebe868a15", "hash_full_prompts": "3a9f9d1ebe868a15", "hash_input_tokens": "ad8567ab36f9ee13", "hash_cont_tokens": "1e63d5b9bb8d45b9" }, "truncated": 0, "non_truncated": 255, "padded": 510, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Ceremony|0": { "hashes": { "hash_examples": "c927630f8d2f44da", "hash_full_prompts": "c927630f8d2f44da", "hash_input_tokens": "0cff9336020b25bf", "hash_cont_tokens": "587187a4b9ec6b9f" }, "truncated": 0, "non_truncated": 185, "padded": 370, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Clothing|0": { "hashes": { "hash_examples": "6ad0740c2ac6ac92", "hash_full_prompts": "6ad0740c2ac6ac92", "hash_input_tokens": "cb7ac62dc81e362f", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Culture|0": { "hashes": { "hash_examples": "2177bd857ad872ae", "hash_full_prompts": "2177bd857ad872ae", "hash_input_tokens": "0e7015608d35b78f", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Food|0": { "hashes": { "hash_examples": "a6ada65b71d7c9c5", "hash_full_prompts": "a6ada65b71d7c9c5", "hash_input_tokens": "176f9f86e09f22c1", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Funeral|0": { "hashes": { "hash_examples": "fcee39dc29eaae91", "hash_full_prompts": "fcee39dc29eaae91", "hash_input_tokens": "c80feef86920f08d", "hash_cont_tokens": "a53062899e4fc8e9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Geography|0": { "hashes": { "hash_examples": "d36eda7c89231c02", "hash_full_prompts": "d36eda7c89231c02", "hash_input_tokens": "504ae8b28699844a", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_History|0": { "hashes": { "hash_examples": "6354ac0d6db6a5fc", "hash_full_prompts": "6354ac0d6db6a5fc", "hash_input_tokens": "9e10d1bd3304f9e3", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Language_Origin|0": { "hashes": { "hash_examples": "ddc967c8aca34402", "hash_full_prompts": "ddc967c8aca34402", "hash_input_tokens": "4b081791e36fc2ba", "hash_cont_tokens": "a53062899e4fc8e9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Literature|0": { "hashes": { "hash_examples": "4305379fd46be5d8", "hash_full_prompts": "4305379fd46be5d8", "hash_input_tokens": "d642f44fc8d57341", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Math|0": { "hashes": { "hash_examples": "dec621144f4d28be", "hash_full_prompts": "dec621144f4d28be", "hash_input_tokens": "33e9d2e77da98db9", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Medicine|0": { "hashes": { "hash_examples": "2b344cdae9495ff2", "hash_full_prompts": "2b344cdae9495ff2", "hash_input_tokens": "7ad55962699f5b17", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Music|0": { "hashes": { "hash_examples": "0c54624d881944ce", "hash_full_prompts": "0c54624d881944ce", "hash_input_tokens": "b53b28e573c29a20", "hash_cont_tokens": "1859e6278efcea1b" }, "truncated": 0, "non_truncated": 139, "padded": 278, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Ornament|0": { "hashes": { "hash_examples": "251a4a84289d8bc1", "hash_full_prompts": "251a4a84289d8bc1", "hash_input_tokens": "b423e8d38bf1401e", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Philosophy|0": { "hashes": { "hash_examples": "3f86fb9c94c13d22", "hash_full_prompts": "3f86fb9c94c13d22", "hash_input_tokens": "423d7e9e5ac038c6", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Physics_and_Chemistry|0": { "hashes": { "hash_examples": "8fec65af3695b62a", "hash_full_prompts": "8fec65af3695b62a", "hash_input_tokens": "3e6062fcee5d7548", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Arabic_Wedding|0": { "hashes": { "hash_examples": "9cc3477184d7a4b8", "hash_full_prompts": "9cc3477184d7a4b8", "hash_input_tokens": "4a2c6818369d4914", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Bahrain|0": { "hashes": { "hash_examples": "c92e803a0fa8b9e2", "hash_full_prompts": "c92e803a0fa8b9e2", "hash_input_tokens": "3cf2201dc483d741", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Comoros|0": { "hashes": { "hash_examples": "06e5d4bba8e54cae", "hash_full_prompts": "06e5d4bba8e54cae", "hash_input_tokens": "3ce79340c7eb4e18", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Egypt_modern|0": { "hashes": { "hash_examples": "c6ec369164f93446", "hash_full_prompts": "c6ec369164f93446", "hash_input_tokens": "ab12b501c65eb974", "hash_cont_tokens": "a53062899e4fc8e9" }, "truncated": 0, "non_truncated": 95, "padded": 190, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromAncientEgypt|0": { "hashes": { "hash_examples": "b9d56d74818b9bd4", "hash_full_prompts": "b9d56d74818b9bd4", "hash_input_tokens": "60732064896b0b00", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromByzantium|0": { "hashes": { "hash_examples": "5316c9624e7e59b8", "hash_full_prompts": "5316c9624e7e59b8", "hash_input_tokens": "a10469d9efe8b860", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromChina|0": { "hashes": { "hash_examples": "87894bce95a56411", "hash_full_prompts": "87894bce95a56411", "hash_input_tokens": "e37ddf38156c8db3", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromGreece|0": { "hashes": { "hash_examples": "0baa78a27e469312", "hash_full_prompts": "0baa78a27e469312", "hash_input_tokens": "498a9a128e2b1dbd", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromIslam|0": { "hashes": { "hash_examples": "0c2532cde6541ff2", "hash_full_prompts": "0c2532cde6541ff2", "hash_input_tokens": "9fb099aed6a50c12", "hash_cont_tokens": "c15aaf8e70b82ff0" }, "truncated": 0, "non_truncated": 145, "padded": 290, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromPersia|0": { "hashes": { "hash_examples": "efcd8112dc53c6e5", "hash_full_prompts": "efcd8112dc53c6e5", "hash_input_tokens": "2f73621de0cbd6ec", "hash_cont_tokens": "7305f8930f27cd4e" }, "truncated": 0, "non_truncated": 175, "padded": 350, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:InfluenceFromRome|0": { "hashes": { "hash_examples": "9db61480e2e85fd3", "hash_full_prompts": "9db61480e2e85fd3", "hash_input_tokens": "78adc78c1a00aa49", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Iraq|0": { "hashes": { "hash_examples": "96dac3dfa8d2f41f", "hash_full_prompts": "96dac3dfa8d2f41f", "hash_input_tokens": "8604335e7393c0a4", "hash_cont_tokens": "79b22c57a64b25bf" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islam_Education|0": { "hashes": { "hash_examples": "0d80355f6a4cb51b", "hash_full_prompts": "0d80355f6a4cb51b", "hash_input_tokens": "7c9d80290354111e", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islam_branches_and_schools|0": { "hashes": { "hash_examples": "5cedce1be2c3ad50", "hash_full_prompts": "5cedce1be2c3ad50", "hash_input_tokens": "50431e94910ebbd2", "hash_cont_tokens": "7305f8930f27cd4e" }, "truncated": 0, "non_truncated": 175, "padded": 350, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Islamic_law_system|0": { "hashes": { "hash_examples": "c0e6db8bc84e105e", "hash_full_prompts": "c0e6db8bc84e105e", "hash_input_tokens": "b1a9a4f17160e249", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Jordan|0": { "hashes": { "hash_examples": "33deb5b4e5ddd6a1", "hash_full_prompts": "33deb5b4e5ddd6a1", "hash_input_tokens": "891619f2d990deaa", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Kuwait|0": { "hashes": { "hash_examples": "eb41773346d7c46c", "hash_full_prompts": "eb41773346d7c46c", "hash_input_tokens": "384d46606a3ddf4f", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Lebanon|0": { "hashes": { "hash_examples": "25932dbf4c13d34f", "hash_full_prompts": "25932dbf4c13d34f", "hash_input_tokens": "6d27674e405c03e0", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Libya|0": { "hashes": { "hash_examples": "f2c4db63cd402926", "hash_full_prompts": "f2c4db63cd402926", "hash_input_tokens": "36841f4d6ef3035a", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Mauritania|0": { "hashes": { "hash_examples": "8723ab5fdf286b54", "hash_full_prompts": "8723ab5fdf286b54", "hash_input_tokens": "7665ce9620c4b9b4", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Mesopotamia_civilization|0": { "hashes": { "hash_examples": "c33f5502a6130ca9", "hash_full_prompts": "c33f5502a6130ca9", "hash_input_tokens": "637fb92f14db5c7c", "hash_cont_tokens": "d6ed79461fc11fb7" }, "truncated": 0, "non_truncated": 155, "padded": 310, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Morocco|0": { "hashes": { "hash_examples": "588a5ed27904b1ae", "hash_full_prompts": "588a5ed27904b1ae", "hash_input_tokens": "80ca5c68d9cf7102", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Oman|0": { "hashes": { "hash_examples": "d447c52b94248b69", "hash_full_prompts": "d447c52b94248b69", "hash_input_tokens": "75ed77fe92ffacaf", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Palestine|0": { "hashes": { "hash_examples": "19197e076ad14ff5", "hash_full_prompts": "19197e076ad14ff5", "hash_input_tokens": "c3d6f52430735d79", "hash_cont_tokens": "79b22c57a64b25bf" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Qatar|0": { "hashes": { "hash_examples": "cf0736fa185b28f6", "hash_full_prompts": "cf0736fa185b28f6", "hash_input_tokens": "1eccbeb7657108aa", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Saudi_Arabia|0": { "hashes": { "hash_examples": "69beda6e1b85a08d", "hash_full_prompts": "69beda6e1b85a08d", "hash_input_tokens": "29baeafac0cc466f", "hash_cont_tokens": "b3ed374c07f6a1ba" }, "truncated": 0, "non_truncated": 195, "padded": 390, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Somalia|0": { "hashes": { "hash_examples": "b387940c65784fbf", "hash_full_prompts": "b387940c65784fbf", "hash_input_tokens": "5117233323f3c8b8", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Sudan|0": { "hashes": { "hash_examples": "e02c32b9d2dd0c3f", "hash_full_prompts": "e02c32b9d2dd0c3f", "hash_input_tokens": "c0e361cf9ef787e2", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Syria|0": { "hashes": { "hash_examples": "60a6f8fe73bda4bb", "hash_full_prompts": "60a6f8fe73bda4bb", "hash_input_tokens": "a904be81a7e61500", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Tunisia|0": { "hashes": { "hash_examples": "34bb15d3830c5649", "hash_full_prompts": "34bb15d3830c5649", "hash_input_tokens": "8f846e5cac5c2f43", "hash_cont_tokens": "e32fa30311c5caac" }, "truncated": 0, "non_truncated": 45, "padded": 90, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:United_Arab_Emirates|0": { "hashes": { "hash_examples": "98a0ba78172718ce", "hash_full_prompts": "98a0ba78172718ce", "hash_input_tokens": "d0270a16e3b08bc6", "hash_cont_tokens": "79b22c57a64b25bf" }, "truncated": 0, "non_truncated": 85, "padded": 170, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:Yemen|0": { "hashes": { "hash_examples": "18e9bcccbb4ced7a", "hash_full_prompts": "18e9bcccbb4ced7a", "hash_input_tokens": "8e6014e7551469dd", "hash_cont_tokens": "546600309db314db" }, "truncated": 0, "non_truncated": 10, "padded": 20, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:communication|0": { "hashes": { "hash_examples": "9ff28ab5eab5c97b", "hash_full_prompts": "9ff28ab5eab5c97b", "hash_input_tokens": "61ada0772437754d", "hash_cont_tokens": "922348efeac48904" }, "truncated": 0, "non_truncated": 364, "padded": 728, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:computer_and_phone|0": { "hashes": { "hash_examples": "37bac2f086aaf6c2", "hash_full_prompts": "37bac2f086aaf6c2", "hash_input_tokens": "305b63444bd678d2", "hash_cont_tokens": "21c566d247b85282" }, "truncated": 0, "non_truncated": 295, "padded": 590, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:daily_life|0": { "hashes": { "hash_examples": "bf07363c1c252e2f", "hash_full_prompts": "bf07363c1c252e2f", "hash_input_tokens": "9fbfbe318b4723c3", "hash_cont_tokens": "35a4aa65a4889d29" }, "truncated": 0, "non_truncated": 337, "padded": 674, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|acva:entertainment|0": { "hashes": { "hash_examples": "37077bc00f0ac56a", "hash_full_prompts": "37077bc00f0ac56a", "hash_input_tokens": "fcab9d3983de19b3", "hash_cont_tokens": "21c566d247b85282" }, "truncated": 0, "non_truncated": 295, "padded": 590, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:mcq_exams_test_ar|0": { "hashes": { "hash_examples": "c07a5e78c5c0b8fe", "hash_full_prompts": "c07a5e78c5c0b8fe", "hash_input_tokens": "74c9404cea5904fc", "hash_cont_tokens": "1c609c966761a1cb" }, "truncated": 0, "non_truncated": 557, "padded": 2228, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:meta_ar_dialects|0": { "hashes": { "hash_examples": "c0b6081f83e14064", "hash_full_prompts": "c0b6081f83e14064", "hash_input_tokens": "aa16305075e7b616", "hash_cont_tokens": "db8fa460c7021ecc" }, "truncated": 0, "non_truncated": 5395, "padded": 21544, "non_padded": 36, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:meta_ar_msa|0": { "hashes": { "hash_examples": "64eb78a7c5b7484b", "hash_full_prompts": "64eb78a7c5b7484b", "hash_input_tokens": "8d3ff22d1b8242ea", "hash_cont_tokens": "aa5db0b20bfebd00" }, "truncated": 0, "non_truncated": 895, "padded": 3576, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_facts_truefalse_balanced_task|0": { "hashes": { "hash_examples": "54fc3502c1c02c06", "hash_full_prompts": "54fc3502c1c02c06", "hash_input_tokens": "e9ec1cdbedd3ccb6", "hash_cont_tokens": "21a564f3caaf138d" }, "truncated": 0, "non_truncated": 75, "padded": 150, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_grounded_statement_soqal_task|0": { "hashes": { "hash_examples": "46572d83696552ae", "hash_full_prompts": "46572d83696552ae", "hash_input_tokens": "0ac1d8400a4a59fb", "hash_cont_tokens": "145bd6c153138cfe" }, "truncated": 0, "non_truncated": 150, "padded": 748, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_grounded_statement_xglue_mlqa_task|0": { "hashes": { "hash_examples": "f430d97ff715bc1c", "hash_full_prompts": "f430d97ff715bc1c", "hash_input_tokens": "83fb3699a557b00e", "hash_cont_tokens": "e4d3f8dc64e26917" }, "truncated": 0, "non_truncated": 150, "padded": 750, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_rating_sentiment_no_neutral_task|0": { "hashes": { "hash_examples": "6b70a7416584f98c", "hash_full_prompts": "6b70a7416584f98c", "hash_input_tokens": "a33991ade36f7c5d", "hash_cont_tokens": "b0cc8cf11c102265" }, "truncated": 0, "non_truncated": 7995, "padded": 15990, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_rating_sentiment_task|0": { "hashes": { "hash_examples": "bc2005cc9d2f436e", "hash_full_prompts": "bc2005cc9d2f436e", "hash_input_tokens": "5518b1f5b6f4616a", "hash_cont_tokens": "7c4c9000eafebe43" }, "truncated": 0, "non_truncated": 5995, "padded": 17861, "non_padded": 124, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|alghafa:multiple_choice_sentiment_task|0": { "hashes": { "hash_examples": "6fb0e254ea5945d8", "hash_full_prompts": "6fb0e254ea5945d8", "hash_input_tokens": "015e33a15ed1458f", "hash_cont_tokens": "02aed8bf71d25cfb" }, "truncated": 0, "non_truncated": 1720, "padded": 5140, "non_padded": 20, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_exams|0": { "hashes": { "hash_examples": "6d721df351722656", "hash_full_prompts": "6d721df351722656", "hash_input_tokens": "5ff81ca23675d49d", "hash_cont_tokens": "75d35d1d02a02179" }, "truncated": 0, "non_truncated": 537, "padded": 2134, "non_padded": 14, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:abstract_algebra|0": { "hashes": { "hash_examples": "f2ddca8f45c0a511", "hash_full_prompts": "f2ddca8f45c0a511", "hash_input_tokens": "fb34dcea34c4bfbb", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:anatomy|0": { "hashes": { "hash_examples": "dfdbc1b83107668d", "hash_full_prompts": "dfdbc1b83107668d", "hash_input_tokens": "19c9cb242d45d065", "hash_cont_tokens": "e7cfbda8199e7611" }, "truncated": 0, "non_truncated": 135, "padded": 538, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:astronomy|0": { "hashes": { "hash_examples": "9736a606002a848e", "hash_full_prompts": "9736a606002a848e", "hash_input_tokens": "9d1c33c929a72e91", "hash_cont_tokens": "d5464986c8d97559" }, "truncated": 0, "non_truncated": 152, "padded": 608, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:business_ethics|0": { "hashes": { "hash_examples": "735e452fbb6dc63d", "hash_full_prompts": "735e452fbb6dc63d", "hash_input_tokens": "494de2f99782e0ac", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:clinical_knowledge|0": { "hashes": { "hash_examples": "6ab0ca4da98aedcf", "hash_full_prompts": "6ab0ca4da98aedcf", "hash_input_tokens": "0f2134b5d5c12f33", "hash_cont_tokens": "048c2bb287cd81e3" }, "truncated": 0, "non_truncated": 265, "padded": 1056, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_biology|0": { "hashes": { "hash_examples": "17e4e390848018a4", "hash_full_prompts": "17e4e390848018a4", "hash_input_tokens": "7e2470c02c478aaa", "hash_cont_tokens": "f77b26b202ef9b2b" }, "truncated": 0, "non_truncated": 144, "padded": 574, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_chemistry|0": { "hashes": { "hash_examples": "4abb169f6dfd234b", "hash_full_prompts": "4abb169f6dfd234b", "hash_input_tokens": "c9623f461652329b", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_computer_science|0": { "hashes": { "hash_examples": "a369e2e941358a1e", "hash_full_prompts": "a369e2e941358a1e", "hash_input_tokens": "38f0ee128006b399", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_mathematics|0": { "hashes": { "hash_examples": "d7be03b8b6020bff", "hash_full_prompts": "d7be03b8b6020bff", "hash_input_tokens": "344cd523e7f65220", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_medicine|0": { "hashes": { "hash_examples": "0518a00f097346bf", "hash_full_prompts": "0518a00f097346bf", "hash_input_tokens": "92f6493f269b92b5", "hash_cont_tokens": "76fa38a751ac9bc2" }, "truncated": 0, "non_truncated": 173, "padded": 692, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:college_physics|0": { "hashes": { "hash_examples": "5d842cd49bc70e12", "hash_full_prompts": "5d842cd49bc70e12", "hash_input_tokens": "1a781538ab2ffcf2", "hash_cont_tokens": "c786ae78224e9572" }, "truncated": 0, "non_truncated": 102, "padded": 408, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:computer_security|0": { "hashes": { "hash_examples": "8e85d9f85be9b32f", "hash_full_prompts": "8e85d9f85be9b32f", "hash_input_tokens": "af95abb26e3c981c", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 400, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:conceptual_physics|0": { "hashes": { "hash_examples": "7964b55a0a49502b", "hash_full_prompts": "7964b55a0a49502b", "hash_input_tokens": "e91a579917bb3e94", "hash_cont_tokens": "cdabd8b4dc5070a7" }, "truncated": 0, "non_truncated": 235, "padded": 936, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:econometrics|0": { "hashes": { "hash_examples": "1e192eae38347257", "hash_full_prompts": "1e192eae38347257", "hash_input_tokens": "8aa1780f05d9dd09", "hash_cont_tokens": "79f7b7012bc5cae3" }, "truncated": 0, "non_truncated": 114, "padded": 456, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:electrical_engineering|0": { "hashes": { "hash_examples": "cf97671d5c441da1", "hash_full_prompts": "cf97671d5c441da1", "hash_input_tokens": "6ed28feef4c59f3c", "hash_cont_tokens": "98c8c04cfecc2e3a" }, "truncated": 0, "non_truncated": 145, "padded": 566, "non_padded": 14, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:elementary_mathematics|0": { "hashes": { "hash_examples": "6f49107ed43c40c5", "hash_full_prompts": "6f49107ed43c40c5", "hash_input_tokens": "534c9cbe68c54152", "hash_cont_tokens": "55f0c3be194d2e8f" }, "truncated": 0, "non_truncated": 378, "padded": 1506, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:formal_logic|0": { "hashes": { "hash_examples": "7922c376008ba77b", "hash_full_prompts": "7922c376008ba77b", "hash_input_tokens": "07db153ef937684e", "hash_cont_tokens": "e715b1eaed2453d1" }, "truncated": 0, "non_truncated": 126, "padded": 500, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:global_facts|0": { "hashes": { "hash_examples": "11f9813185047d5b", "hash_full_prompts": "11f9813185047d5b", "hash_input_tokens": "766202ff5cd4e636", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 392, "non_padded": 8, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_biology|0": { "hashes": { "hash_examples": "2a804b1d90cbe66e", "hash_full_prompts": "2a804b1d90cbe66e", "hash_input_tokens": "34455a7857b50046", "hash_cont_tokens": "b5a22be1545a5885" }, "truncated": 0, "non_truncated": 310, "padded": 1238, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_chemistry|0": { "hashes": { "hash_examples": "0032168adabc53b4", "hash_full_prompts": "0032168adabc53b4", "hash_input_tokens": "7623206a301d6446", "hash_cont_tokens": "41f59f267986db24" }, "truncated": 0, "non_truncated": 203, "padded": 808, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_computer_science|0": { "hashes": { "hash_examples": "f2fb8740f9df980f", "hash_full_prompts": "f2fb8740f9df980f", "hash_input_tokens": "5222590c759fd0ed", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 398, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_european_history|0": { "hashes": { "hash_examples": "73509021e7e66435", "hash_full_prompts": "73509021e7e66435", "hash_input_tokens": "d3eeb4ec025d4817", "hash_cont_tokens": "ba925766deaa3c15" }, "truncated": 0, "non_truncated": 165, "padded": 660, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_geography|0": { "hashes": { "hash_examples": "9e08d1894940ff42", "hash_full_prompts": "9e08d1894940ff42", "hash_input_tokens": "27ea5b8c7af8d8ce", "hash_cont_tokens": "ba892eb0674f32f0" }, "truncated": 0, "non_truncated": 198, "padded": 786, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_government_and_politics|0": { "hashes": { "hash_examples": "64b7e97817ca6c76", "hash_full_prompts": "64b7e97817ca6c76", "hash_input_tokens": "149272b7d904eaaa", "hash_cont_tokens": "da6ef64a42a0438e" }, "truncated": 0, "non_truncated": 193, "padded": 771, "non_padded": 1, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_macroeconomics|0": { "hashes": { "hash_examples": "9f582da8534bd2ef", "hash_full_prompts": "9f582da8534bd2ef", "hash_input_tokens": "6e8f43b6c114f14c", "hash_cont_tokens": "6660aeaa8368b295" }, "truncated": 0, "non_truncated": 390, "padded": 1556, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_mathematics|0": { "hashes": { "hash_examples": "fd54f1c10d423c51", "hash_full_prompts": "fd54f1c10d423c51", "hash_input_tokens": "1d00aebef57b4512", "hash_cont_tokens": "07575f7140327432" }, "truncated": 0, "non_truncated": 270, "padded": 1074, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_microeconomics|0": { "hashes": { "hash_examples": "7037896925aaf42f", "hash_full_prompts": "7037896925aaf42f", "hash_input_tokens": "cac0df68663c1dae", "hash_cont_tokens": "19bfd23905b3bec3" }, "truncated": 0, "non_truncated": 238, "padded": 946, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_physics|0": { "hashes": { "hash_examples": "60c3776215167dae", "hash_full_prompts": "60c3776215167dae", "hash_input_tokens": "08328936e9eac5df", "hash_cont_tokens": "0562b521128fbe6b" }, "truncated": 0, "non_truncated": 151, "padded": 604, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_psychology|0": { "hashes": { "hash_examples": "61176bfd5da1298f", "hash_full_prompts": "61176bfd5da1298f", "hash_input_tokens": "0bde10a2cf4f41a7", "hash_cont_tokens": "cfa877943c39a466" }, "truncated": 0, "non_truncated": 545, "padded": 2164, "non_padded": 16, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_statistics|0": { "hashes": { "hash_examples": "40dfeebd1ea10f76", "hash_full_prompts": "40dfeebd1ea10f76", "hash_input_tokens": "91cc9ad90604cfab", "hash_cont_tokens": "e58c5a09e16fdd84" }, "truncated": 0, "non_truncated": 216, "padded": 862, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_us_history|0": { "hashes": { "hash_examples": "03daa510ba917f4d", "hash_full_prompts": "03daa510ba917f4d", "hash_input_tokens": "97cb25f55d24d0bf", "hash_cont_tokens": "b93a8c4f2a2af9f0" }, "truncated": 0, "non_truncated": 204, "padded": 806, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:high_school_world_history|0": { "hashes": { "hash_examples": "be075ffd579f43c2", "hash_full_prompts": "be075ffd579f43c2", "hash_input_tokens": "388181fcdc17cac3", "hash_cont_tokens": "54a49424057f2e6b" }, "truncated": 0, "non_truncated": 237, "padded": 922, "non_padded": 26, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:human_aging|0": { "hashes": { "hash_examples": "caa5b69f640bd1ef", "hash_full_prompts": "caa5b69f640bd1ef", "hash_input_tokens": "52af470aa11b69a6", "hash_cont_tokens": "eec7417389927586" }, "truncated": 0, "non_truncated": 223, "padded": 888, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:human_sexuality|0": { "hashes": { "hash_examples": "5ed2e38fb25a3767", "hash_full_prompts": "5ed2e38fb25a3767", "hash_input_tokens": "35fb3fcc4cac5ea0", "hash_cont_tokens": "9dc6bb3c4ecb3178" }, "truncated": 0, "non_truncated": 131, "padded": 518, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:international_law|0": { "hashes": { "hash_examples": "4e3e9e28d1b96484", "hash_full_prompts": "4e3e9e28d1b96484", "hash_input_tokens": "6592ce058a08a3d3", "hash_cont_tokens": "192d0de6ed8ebc9b" }, "truncated": 0, "non_truncated": 121, "padded": 482, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:jurisprudence|0": { "hashes": { "hash_examples": "e264b755366310b3", "hash_full_prompts": "e264b755366310b3", "hash_input_tokens": "c00da155a9d241c5", "hash_cont_tokens": "75dcdf05908542a5" }, "truncated": 0, "non_truncated": 108, "padded": 430, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:logical_fallacies|0": { "hashes": { "hash_examples": "a4ab6965a3e38071", "hash_full_prompts": "a4ab6965a3e38071", "hash_input_tokens": "494e6acac2c1e1cb", "hash_cont_tokens": "ffd60b4e1fa30f63" }, "truncated": 0, "non_truncated": 163, "padded": 652, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:machine_learning|0": { "hashes": { "hash_examples": "b92320efa6636b40", "hash_full_prompts": "b92320efa6636b40", "hash_input_tokens": "e12dc096d068b228", "hash_cont_tokens": "4f0e7389d8fa87d9" }, "truncated": 0, "non_truncated": 112, "padded": 444, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:management|0": { "hashes": { "hash_examples": "c9ee4872a850fe20", "hash_full_prompts": "c9ee4872a850fe20", "hash_input_tokens": "6a420f80b29dc0eb", "hash_cont_tokens": "75a8c1e4c452cb1d" }, "truncated": 0, "non_truncated": 103, "padded": 408, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:marketing|0": { "hashes": { "hash_examples": "0c151b70f6a047e3", "hash_full_prompts": "0c151b70f6a047e3", "hash_input_tokens": "84a6e40173f3305f", "hash_cont_tokens": "727f87587fa3a95b" }, "truncated": 0, "non_truncated": 234, "padded": 930, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:medical_genetics|0": { "hashes": { "hash_examples": "513f6cb8fca3a24e", "hash_full_prompts": "513f6cb8fca3a24e", "hash_input_tokens": "ba044e03d9dab114", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 394, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:miscellaneous|0": { "hashes": { "hash_examples": "259a190d635331db", "hash_full_prompts": "259a190d635331db", "hash_input_tokens": "11b5b4e2ca9ec999", "hash_cont_tokens": "73f36e25a6fef508" }, "truncated": 0, "non_truncated": 783, "padded": 3100, "non_padded": 32, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:moral_disputes|0": { "hashes": { "hash_examples": "b85052c48a0b7bc3", "hash_full_prompts": "b85052c48a0b7bc3", "hash_input_tokens": "6113413bf25f1998", "hash_cont_tokens": "6b2c2a581ad40b69" }, "truncated": 0, "non_truncated": 346, "padded": 1370, "non_padded": 14, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:moral_scenarios|0": { "hashes": { "hash_examples": "28d0b069ef00dd00", "hash_full_prompts": "28d0b069ef00dd00", "hash_input_tokens": "169b6a2f69dea239", "hash_cont_tokens": "af64a7a018654c30" }, "truncated": 0, "non_truncated": 895, "padded": 3580, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:nutrition|0": { "hashes": { "hash_examples": "00c9bc5f1d305b2f", "hash_full_prompts": "00c9bc5f1d305b2f", "hash_input_tokens": "c6f3c9967ab495ed", "hash_cont_tokens": "6e20cc0d504ecac3" }, "truncated": 0, "non_truncated": 306, "padded": 1212, "non_padded": 12, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:philosophy|0": { "hashes": { "hash_examples": "a458c08454a3fd5f", "hash_full_prompts": "a458c08454a3fd5f", "hash_input_tokens": "2bc29c155959e8f1", "hash_cont_tokens": "aa4ac747c265c6ee" }, "truncated": 0, "non_truncated": 311, "padded": 1234, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:prehistory|0": { "hashes": { "hash_examples": "d6a0ecbdbb670e9c", "hash_full_prompts": "d6a0ecbdbb670e9c", "hash_input_tokens": "434b994dd18d95d6", "hash_cont_tokens": "77d8e266e29e78db" }, "truncated": 0, "non_truncated": 324, "padded": 1286, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_accounting|0": { "hashes": { "hash_examples": "b4a95fe480b6540e", "hash_full_prompts": "b4a95fe480b6540e", "hash_input_tokens": "714986665ea8c367", "hash_cont_tokens": "e25888a5ba75c974" }, "truncated": 0, "non_truncated": 282, "padded": 1128, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_law|0": { "hashes": { "hash_examples": "c2be9651cdbdde3b", "hash_full_prompts": "c2be9651cdbdde3b", "hash_input_tokens": "e636b884c800f26c", "hash_cont_tokens": "d1bfd61f4ba42a8b" }, "truncated": 0, "non_truncated": 1534, "padded": 6126, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_medicine|0": { "hashes": { "hash_examples": "26ce92416288f273", "hash_full_prompts": "26ce92416288f273", "hash_input_tokens": "b313dc05e50e31aa", "hash_cont_tokens": "6606c8af95c09c78" }, "truncated": 0, "non_truncated": 272, "padded": 1082, "non_padded": 6, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:professional_psychology|0": { "hashes": { "hash_examples": "71ea5f182ea9a641", "hash_full_prompts": "71ea5f182ea9a641", "hash_input_tokens": "548cbe1a67b9340c", "hash_cont_tokens": "b061897bffe334ec" }, "truncated": 0, "non_truncated": 612, "padded": 2438, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:public_relations|0": { "hashes": { "hash_examples": "125adc21f91f8d77", "hash_full_prompts": "125adc21f91f8d77", "hash_input_tokens": "0a3b879795cea1d4", "hash_cont_tokens": "5c3107c12bceb18e" }, "truncated": 0, "non_truncated": 110, "padded": 438, "non_padded": 2, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:security_studies|0": { "hashes": { "hash_examples": "3c18b216c099fb26", "hash_full_prompts": "3c18b216c099fb26", "hash_input_tokens": "9a3525d9fb39eaeb", "hash_cont_tokens": "ce4361df75a6e6a7" }, "truncated": 0, "non_truncated": 245, "padded": 980, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:sociology|0": { "hashes": { "hash_examples": "3f2a9634cef7417d", "hash_full_prompts": "3f2a9634cef7417d", "hash_input_tokens": "03689972cc765336", "hash_cont_tokens": "cbbeed1c4c5a128e" }, "truncated": 0, "non_truncated": 201, "padded": 800, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:us_foreign_policy|0": { "hashes": { "hash_examples": "22249da54056475e", "hash_full_prompts": "22249da54056475e", "hash_input_tokens": "9cf7257c481425c0", "hash_cont_tokens": "00a223315c15a9ce" }, "truncated": 0, "non_truncated": 100, "padded": 396, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:virology|0": { "hashes": { "hash_examples": "9d194b9471dc624e", "hash_full_prompts": "9d194b9471dc624e", "hash_input_tokens": "9cb93eaa1e3a2444", "hash_cont_tokens": "84e4fe7f10383133" }, "truncated": 0, "non_truncated": 166, "padded": 654, "non_padded": 10, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arabic_mmlu:world_religions|0": { "hashes": { "hash_examples": "229e5fe50082b064", "hash_full_prompts": "229e5fe50082b064", "hash_input_tokens": "427506f7b3220a25", "hash_cont_tokens": "a0fac287dd015964" }, "truncated": 0, "non_truncated": 171, "padded": 680, "non_padded": 4, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arc_challenge_okapi_ar|0": { "hashes": { "hash_examples": "ab893807673bc355", "hash_full_prompts": "ab893807673bc355", "hash_input_tokens": "aba9faeb29515a4e", "hash_cont_tokens": "3b08380afd4003dd" }, "truncated": 0, "non_truncated": 1160, "padded": 4607, "non_padded": 33, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|arc_easy_ar|0": { "hashes": { "hash_examples": "acb688624acc3d04", "hash_full_prompts": "acb688624acc3d04", "hash_input_tokens": "300ec1091f36041e", "hash_cont_tokens": "a5ff8db30710c8eb" }, "truncated": 0, "non_truncated": 2364, "padded": 9394, "non_padded": 62, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|boolq_ar|0": { "hashes": { "hash_examples": "48355a67867e0c32", "hash_full_prompts": "48355a67867e0c32", "hash_input_tokens": "d6da88c86df9601e", "hash_cont_tokens": "647569b4ef9142b6" }, "truncated": 2, "non_truncated": 3258, "padded": 6501, "non_padded": 19, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|copa_ext_ar|0": { "hashes": { "hash_examples": "9bb83301bb72eecf", "hash_full_prompts": "9bb83301bb72eecf", "hash_input_tokens": "d703db6f7b6a3745", "hash_cont_tokens": "6240ddc895d662c5" }, "truncated": 0, "non_truncated": 90, "padded": 180, "non_padded": 0, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|hellaswag_okapi_ar|0": { "hashes": { "hash_examples": "6e8cf57a322dfadd", "hash_full_prompts": "6e8cf57a322dfadd", "hash_input_tokens": "cc431d860b0d78d4", "hash_cont_tokens": "c62d1b0bf12dfd31" }, "truncated": 0, "non_truncated": 9171, "padded": 36637, "non_padded": 47, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|openbook_qa_ext_ar|0": { "hashes": { "hash_examples": "923d41eb0aca93eb", "hash_full_prompts": "923d41eb0aca93eb", "hash_input_tokens": "7d2e58d70c04e83d", "hash_cont_tokens": "5400680776484572" }, "truncated": 0, "non_truncated": 495, "padded": 1971, "non_padded": 9, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|piqa_ar|0": { "hashes": { "hash_examples": "94bc205a520d3ea0", "hash_full_prompts": "94bc205a520d3ea0", "hash_input_tokens": "a76a81f80ef1f500", "hash_cont_tokens": "5c629f9c59ced4fa" }, "truncated": 0, "non_truncated": 1833, "padded": 3653, "non_padded": 13, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|race_ar|0": { "hashes": { "hash_examples": "de65130bae647516", "hash_full_prompts": "de65130bae647516", "hash_input_tokens": "9612747fa401e66e", "hash_cont_tokens": "3a36d0b695f01a74" }, "truncated": 124, "non_truncated": 4805, "padded": 19587, "non_padded": 129, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|sciq_ar|0": { "hashes": { "hash_examples": "0117e422da72544e", "hash_full_prompts": "0117e422da72544e", "hash_input_tokens": "db33019489c5a405", "hash_cont_tokens": "884827cc3b5659ad" }, "truncated": 0, "non_truncated": 995, "padded": 3975, "non_padded": 5, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "community|toxigen_ar|0": { "hashes": { "hash_examples": "1e139513004a9a2e", "hash_full_prompts": "1e139513004a9a2e", "hash_input_tokens": "c10a12207615347d", "hash_cont_tokens": "083c19f8d59268d0" }, "truncated": 0, "non_truncated": 935, "padded": 1859, "non_padded": 11, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 }, "lighteval|xstory_cloze:ar|0": { "hashes": { "hash_examples": "865426a22c787481", "hash_full_prompts": "865426a22c787481", "hash_input_tokens": "fdf843f8148316ae", "hash_cont_tokens": "203e48ea9c289f4d" }, "truncated": 0, "non_truncated": 1511, "padded": 3003, "non_padded": 19, "effective_few_shots": 0.0, "num_truncated_few_shots": 0 } }, "summary_general": { "hashes": { "hash_examples": "333e6e8097d9cc72", "hash_full_prompts": "333e6e8097d9cc72", "hash_input_tokens": "db58d045a453a3be", "hash_cont_tokens": "947428e37bb1074a" }, "truncated": 126, "non_truncated": 72838, "padded": 234779, "non_padded": 844, "num_truncated_few_shots": 0 } }