djstrong's picture
test
ae845e4
raw
history blame
27.8 kB
{
"results": {
"polish": {
"acc,none": 0.5485088323770747,
"acc_stderr,none": 0.006173394316567496,
"acc_norm,none": 0.5613679808033012,
"acc_norm_stderr,none": 0.008184643816451457,
"f1,score-first": 0.5560581187634298,
"f1_stderr,score-first": 0.10012844040270957,
"accuracy,score-first": 0.5560581187634298,
"accuracy_stderr,score-first": 0.10012844040270957,
"alias": "polish"
},
"belebele_pol_Latn": {
"acc,none": 0.5077777777777778,
"acc_stderr,none": 0.016673915944964764,
"acc_norm,none": 0.5077777777777778,
"acc_norm_stderr,none": 0.016673915944964764,
"alias": " - belebele_pol_Latn"
},
"polemo2_in": {
"f1,score-first": 0.554016620498615,
"f1_stderr,score-first": 0.018511989768822235,
"accuracy,score-first": 0.554016620498615,
"accuracy_stderr,score-first": 0.018511989768822235,
"alias": " - polemo2_in"
},
"polemo2_in_multiple_choice": {
"acc,none": 0.27285318559556787,
"acc_stderr,none": 0.016588531234686118,
"acc_norm,none": 0.20221606648199447,
"acc_norm_stderr,none": 0.01495832068619682,
"alias": " - polemo2_in_multiple_choice"
},
"polemo2_out": {
"f1,score-first": 0.5566801619433198,
"f1_stderr,score-first": 0.0223737095446236,
"accuracy,score-first": 0.5566801619433198,
"accuracy_stderr,score-first": 0.0223737095446236,
"alias": " - polemo2_out"
},
"polemo2_out_multiple_choice": {
"acc,none": 0.3299595141700405,
"acc_stderr,none": 0.02117665642255094,
"acc_norm,none": 0.3340080971659919,
"acc_norm_stderr,none": 0.021241711816127367,
"alias": " - polemo2_out_multiple_choice"
},
"polish_8tags_multiple_choice": {
"acc,none": 0.6365507776761208,
"acc_stderr,none": 0.007275251233570105,
"acc_norm,none": 0.6779505946935042,
"acc_norm_stderr,none": 0.0070675659629489225,
"alias": " - polish_8tags_multiple_choice"
},
"polish_8tags_regex": {
"f1,score-first": 0.6564501372369624,
"f1_stderr,score-first": 0.007182991593741683,
"accuracy,score-first": 0.6564501372369624,
"accuracy_stderr,score-first": 0.007182991593741683,
"alias": " - polish_8tags_regex"
},
"polish_belebele_regex": {
"f1,score-first": 0.5433333333333333,
"f1_stderr,score-first": 0.016613188209796736,
"accuracy,score-first": 0.5433333333333333,
"accuracy_stderr,score-first": 0.016613188209796736,
"alias": " - polish_belebele_regex"
},
"polish_dyk_multiple_choice": {
"acc,none": 0.5481049562682215,
"acc_stderr,none": 0.01552222935866704,
"acc_norm,none": 0.5481049562682215,
"acc_norm_stderr,none": 0.01552222935866704,
"alias": " - polish_dyk_multiple_choice"
},
"polish_dyk_regex": {
"f1,score-first": 0.1749271137026239,
"f1_stderr,score-first": 0.011848903583973948,
"accuracy,score-first": 0.1749271137026239,
"accuracy_stderr,score-first": 0.011848903583973948,
"alias": " - polish_dyk_regex"
},
"polish_ppc_multiple_choice": {
"acc,none": 0.449,
"acc_stderr,none": 0.015736792768752016,
"acc_norm,none": 0.449,
"acc_norm_stderr,none": 0.015736792768752016,
"alias": " - polish_ppc_multiple_choice"
},
"polish_ppc_regex": {
"f1,score-first": 0.46,
"f1_stderr,score-first": 0.01576859691439438,
"accuracy,score-first": 0.46,
"accuracy_stderr,score-first": 0.01576859691439438,
"alias": " - polish_ppc_regex"
},
"polish_psc_multiple_choice": {
"acc,none": 0.725417439703154,
"acc_stderr,none": 0.01359948992290204,
"acc_norm,none": 0.725417439703154,
"acc_norm_stderr,none": 0.01359948992290204,
"alias": " - polish_psc_multiple_choice"
},
"polish_psc_regex": {
"f1,score-first": 0.5918367346938775,
"f1_stderr,score-first": 0.014976490836384216,
"accuracy,score-first": 0.5918367346938775,
"accuracy_stderr,score-first": 0.014976490836384216,
"alias": " - polish_psc_regex"
}
},
"groups": {
"polish": {
"acc,none": 0.5485088323770747,
"acc_stderr,none": 0.006173394316567496,
"acc_norm,none": 0.5613679808033012,
"acc_norm_stderr,none": 0.008184643816451457,
"f1,score-first": 0.5560581187634298,
"f1_stderr,score-first": 0.10012844040270957,
"accuracy,score-first": 0.5560581187634298,
"accuracy_stderr,score-first": 0.10012844040270957,
"alias": "polish"
}
},
"configs": {
"belebele_pol_Latn": {
"task": "belebele_pol_Latn",
"group": "belebele",
"dataset_path": "facebook/belebele",
"test_split": "pol_Latn",
"fewshot_split": "pol_Latn",
"doc_to_text": "P: {{flores_passage}}\nQ: {{question.strip()}}\nA: {{mc_answer1}}\nB: {{mc_answer2}}\nC: {{mc_answer3}}\nD: {{mc_answer4}}\nAnswer:",
"doc_to_target": "{{['1', '2', '3', '4'].index(correct_answer_num)}}",
"doc_to_choice": [
"A",
"B",
"C",
"D"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"fewshot_config": {
"sampler": "first_n"
},
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{question}}",
"metadata": {
"version": 0.0
}
},
"polemo2_in": {
"task": "polemo2_in",
"group": [
"polemo2"
],
"dataset_path": "allegro/klej-polemo2-in",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii. Możliwe odpowiedzi:\nA - Neutralny\nB - Negatywny\nC - Pozytywny\nD - Niejednoznaczny\nPrawidłowa odpowiedź:",
"doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}",
"metadata": {
"version": 1.0
}
},
"polemo2_in_multiple_choice": {
"task": "polemo2_in_multiple_choice",
"group": [
"polemo2_mc"
],
"dataset_path": "allegro/klej-polemo2-in",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii: Neutralny, Negatywny, Pozytywny, Niejednoznaczny.\nSentyment:",
"doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}",
"doc_to_choice": [
"Neutralny",
"Negatywny",
"Pozytywny",
"Niejednoznaczny"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}"
},
"polemo2_out": {
"task": "polemo2_out",
"group": [
"polemo2"
],
"dataset_path": "allegro/klej-polemo2-out",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii. Możliwe odpowiedzi:\nA - Neutralny\nB - Negatywny\nC - Pozytywny\nD - Niejednoznaczny\nPrawidłowa odpowiedź:",
"doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}",
"metadata": {
"version": 1.0
}
},
"polemo2_out_multiple_choice": {
"task": "polemo2_out_multiple_choice",
"group": [
"polemo2_mc"
],
"dataset_path": "allegro/klej-polemo2-out",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Opinia: \"{{sentence}}\"\nOkreśl sentyment podanej opinii: Neutralny, Negatywny, Pozytywny, Niejednoznaczny.\nSentyment:",
"doc_to_target": "{{['__label__meta_zero', '__label__meta_minus_m', '__label__meta_plus_m', '__label__meta_amb'].index(target)}}",
"doc_to_choice": [
"Neutralny",
"Negatywny",
"Pozytywny",
"Niejednoznaczny"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}"
},
"polish_8tags_multiple_choice": {
"task": "polish_8tags_multiple_choice",
"dataset_path": "djstrong/8tags",
"training_split": "train",
"test_split": "test",
"fewshot_split": "train",
"doc_to_text": "Tytuł: \"{{sentence}}\"\nDo podanego tytułu przyporządkuj jedną najlepiej pasującą kategorię z podanych: Film, Historia, Jedzenie, Medycyna, Motoryzacja, Praca, Sport, Technologie.\nKategoria:",
"doc_to_target": "{{label|int}}",
"doc_to_choice": [
"Film",
"Historia",
"Jedzenie",
"Medycyna",
"Motoryzacja",
"Praca",
"Sport",
"Technologie"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}"
},
"polish_8tags_regex": {
"task": "polish_8tags_regex",
"dataset_path": "djstrong/8tags",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Tytuł: \"{{sentence}}\"\nPytanie: jaka kategoria najlepiej pasuje do podanego tytułu?\nMożliwe odpowiedzi:\nA - film\nB - historia\nC - jedzenie\nD - medycyna\nE - motoryzacja\nF - praca\nG - sport\nH - technologie\nPrawidłowa odpowiedź:",
"doc_to_target": "{{label|int}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCDEFGH]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3,
"E": 4,
"F": 5,
"G": 6,
"H": 7
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence}}"
},
"polish_belebele_regex": {
"task": "polish_belebele_regex",
"dataset_path": "facebook/belebele",
"test_split": "pol_Latn",
"doc_to_text": "Fragment: \"{{flores_passage}}\"\nPytanie: \"{{question}}\"\nMożliwe odpowiedzi:\nA - {{mc_answer1}}\nB - {{mc_answer2}}\nC - {{mc_answer3}}\nD - {{mc_answer4}}\nPrawidłowa odpowiedź:",
"doc_to_target": "{{correct_answer_num|int - 1}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{flores_passage}} {{question}} {{mc_answer1}} {{mc_answer2}} {{mc_answer3}} {{mc_answer4}}"
},
"polish_dyk_multiple_choice": {
"task": "polish_dyk_multiple_choice",
"dataset_path": "allegro/klej-dyk",
"training_split": "train",
"test_split": "test",
"doc_to_text": "Pytanie: \"{{question}}\"\nSugerowana odpowiedź: \"{{answer}}\"\nPytanie: Czy sugerowana odpowiedź na zadane pytanie jest poprawna?\nOdpowiedz krótko \"Tak\" lub \"Nie\". Prawidłowa odpowiedź:",
"doc_to_target": "{{target|int}}",
"doc_to_choice": [
"Nie",
"Tak"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{question}} {{answer}}"
},
"polish_dyk_regex": {
"task": "polish_dyk_regex",
"dataset_path": "allegro/klej-dyk",
"training_split": "train",
"test_split": "test",
"doc_to_text": "Pytanie: \"{{question}}\"\nSugerowana odpowiedź: \"{{answer}}\"\nCzy sugerowana odpowiedź na zadane pytanie jest poprawna? Możliwe opcje:\nA - brakuje sugerowanej odpowiedzi\nB - nie, sugerowana odpowiedź nie jest poprawna\nC - tak, sugerowana odpowiedź jest poprawna\nD - brakuje pytania\nPrawidłowa opcja:",
"doc_to_target": "{{target|int+1}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{question}} {{answer}}"
},
"polish_ppc_multiple_choice": {
"task": "polish_ppc_multiple_choice",
"dataset_path": "djstrong/ppc",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Zdanie A: \"{{sentence_A}}\"\nZdanie B: \"{{sentence_B}}\"\nPytanie: jaka jest zależność między zdaniami A i B? Możliwe odpowiedzi:\nA - znaczą dokładnie to samo\nB - mają podobne znaczenie\nC - mają różne znaczenie\nPrawidłowa odpowiedź:",
"doc_to_target": "{{label|int - 1}}",
"doc_to_choice": [
"A",
"B",
"C"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence_A}} {{sentence_B}}"
},
"polish_ppc_regex": {
"task": "polish_ppc_regex",
"dataset_path": "djstrong/ppc",
"training_split": "train",
"validation_split": "validation",
"test_split": "test",
"doc_to_text": "Zdanie A: \"{{sentence_A}}\"\nZdanie B: \"{{sentence_B}}\"\nPytanie: jaka jest zależność między zdaniami A i B? Możliwe odpowiedzi:\nA - wszystkie odpowiedzi poprawne\nB - znaczą dokładnie to samo\nC - mają podobne znaczenie\nD - mają różne znaczenie\nPrawidłowa odpowiedź:",
"doc_to_target": "{{label|int}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{sentence_A}} {{sentence_B}}"
},
"polish_psc_multiple_choice": {
"task": "polish_psc_multiple_choice",
"dataset_path": "allegro/klej-psc",
"training_split": "train",
"test_split": "test",
"doc_to_text": "Tekst: \"{{extract_text}}\"\nPodsumowanie: \"{{summary_text}}\"\nPytanie: Czy podsumowanie dla podanego tekstu jest poprawne?\nOdpowiedz krótko \"Tak\" lub \"Nie\". Prawidłowa odpowiedź:",
"doc_to_target": "{{label|int}}",
"doc_to_choice": [
"Nie",
"Tak"
],
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "acc",
"aggregation": "mean",
"higher_is_better": true
},
{
"metric": "acc_norm",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": true,
"doc_to_decontamination_query": "{{extract_text}} {{summary_text}}"
},
"polish_psc_regex": {
"task": "polish_psc_regex",
"dataset_path": "allegro/klej-psc",
"training_split": "train",
"test_split": "test",
"doc_to_text": "Fragment 1: \"{{extract_text}}\"\nFragment 2: \"{{summary_text}}\"\nPytanie: jaka jest zależność między fragmentami 1 i 2?\nMożliwe odpowiedzi:\nA - wszystkie odpowiedzi poprawne\nB - dotyczą tego samego artykułu\nC - dotyczą różnych artykułów\nD - brak poprawnej odpowiedzi\nPrawidłowa odpowiedź:",
"doc_to_target": "{{label|int + 1}}",
"description": "",
"target_delimiter": " ",
"fewshot_delimiter": "\n\n",
"num_fewshot": 0,
"metric_list": [
{
"metric": "f1",
"aggregation": "mean",
"higher_is_better": true,
"hf_evaluate": true,
"average": "micro"
},
{
"metric": "accuracy",
"aggregation": "mean",
"higher_is_better": true
}
],
"output_type": "generate_until",
"generation_kwargs": {
"until": [
".",
","
],
"do_sample": false,
"temperature": 0.0,
"max_gen_toks": 50
},
"repeats": 1,
"filter_list": [
{
"name": "score-first",
"filter": [
{
"function": "regex",
"regex_pattern": "(\\b[ABCD]\\b)"
},
{
"function": "take_first"
},
{
"function": "map",
"mapping_dict": {
"A": 0,
"B": 1,
"C": 2,
"D": 3
},
"default_value": -1
},
{
"function": "take_first"
}
]
}
],
"should_decontaminate": true,
"doc_to_decontamination_query": "{{extract_text}} {{summary_text}}"
}
},
"versions": {
"belebele_pol_Latn": 0.0,
"polemo2_in": 1.0,
"polemo2_in_multiple_choice": "Yaml",
"polemo2_out": 1.0,
"polemo2_out_multiple_choice": "Yaml",
"polish": "N/A",
"polish_8tags_multiple_choice": "Yaml",
"polish_8tags_regex": "Yaml",
"polish_belebele_regex": "Yaml",
"polish_dyk_multiple_choice": "Yaml",
"polish_dyk_regex": "Yaml",
"polish_ppc_multiple_choice": "Yaml",
"polish_ppc_regex": "Yaml",
"polish_psc_multiple_choice": "Yaml",
"polish_psc_regex": "Yaml"
},
"n-shot": {
"belebele_pol_Latn": 0,
"polemo2_in": 0,
"polemo2_in_multiple_choice": 0,
"polemo2_out": 0,
"polemo2_out_multiple_choice": 0,
"polish": 0,
"polish_8tags_multiple_choice": 0,
"polish_8tags_regex": 0,
"polish_belebele_regex": 0,
"polish_dyk_multiple_choice": 0,
"polish_dyk_regex": 0,
"polish_ppc_multiple_choice": 0,
"polish_ppc_regex": 0,
"polish_psc_multiple_choice": 0,
"polish_psc_regex": 0
},
"config": {
"model": "hf",
"model_args": "pretrained=Voicelab/trurl-2-13b-academic",
"batch_size": "1",
"batch_sizes": [],
"device": "cuda:0",
"use_cache": null,
"limit": null,
"bootstrap_iters": 100000,
"gen_kwargs": null
},
"git_hash": null
}