Spaces:
Running
Running
{ | |
"benczechmark_propaganda_argumentace": { | |
"name": "Propaganda β Argumentace", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_argumentace", | |
"short_name": "P-Argumentace", | |
"category": "NLI", | |
"abbreviation": "P-ARG" | |
}, | |
"benczechmark_propaganda_fabulace": { | |
"name": "Propaganda β Fabulace", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_fabulace", | |
"short_name": "P-Fabulace", | |
"category": "NLI", | |
"abbreviation": "P-FAB" | |
}, | |
"benczechmark_propaganda_nazor": { | |
"name": "Propaganda β NΓ‘zor", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_nazor", | |
"short_name": "P-NΓ‘zor", | |
"category": "NLI", | |
"abbreviation": "P-NAZOR" | |
}, | |
"benczechmark_propaganda_strach": { | |
"name": "Propaganda β Strach", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_strach", | |
"short_name": "P-Strach", | |
"category": "NLI", | |
"abbreviation": "P-STCH" | |
}, | |
"benczechmark_propaganda_zamereni": { | |
"name": "Propaganda β ZamΔΕenΓ", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_zamereni", | |
"short_name": "P-ZamΔΕenΓ", | |
"category": "NLI", | |
"abbreviation": "P-MER" | |
}, | |
"benczechmark_propaganda_demonizace": { | |
"name": "Propaganda β DΓ©monizace", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_demonizace", | |
"short_name": "P-DΓ©monizace", | |
"category": "NLI", | |
"abbreviation": "P-DEMON" | |
}, | |
"benczechmark_propaganda_lokace": { | |
"name": "Propaganda β Lokace", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_lokace", | |
"short_name": "P-Lokace", | |
"category": "NLI", | |
"abbreviation": "P-LOK" | |
}, | |
"benczechmark_propaganda_relativizace": { | |
"name": "Propaganda β Relativizace", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_relativizace", | |
"short_name": "P-Relativizace", | |
"category": "NLI", | |
"abbreviation": "P-REL" | |
}, | |
"benczechmark_propaganda_vina": { | |
"name": "Propaganda β Vina", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_vina", | |
"short_name": "P-Vina", | |
"category": "NLI", | |
"abbreviation": "P-VINA" | |
}, | |
"benczechmark_propaganda_zanr": { | |
"name": "Propaganda β Ε½Γ‘nr", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_zanr", | |
"short_name": "P-Ε½Γ‘nr", | |
"category": "NLI", | |
"abbreviation": "P-ZANR" | |
}, | |
"benczechmark_propaganda_emoce": { | |
"name": "Propaganda β Emoce", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_emoce", | |
"short_name": "P-Emoce", | |
"category": "NLI", | |
"abbreviation": "P-EMOCE" | |
}, | |
"benczechmark_propaganda_nalepkovani": { | |
"name": "Propaganda β NΓ‘lepkovΓ‘nΓ", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_nalepkovani", | |
"short_name": "P-NΓ‘lepkovΓ‘nΓ", | |
"category": "NLI", | |
"abbreviation": "P-LEP" | |
}, | |
"benczechmark_propaganda_rusko": { | |
"name": "Propaganda β Rusko", | |
"source_url": "https://huggingface.co/datasets/CZLC/propaganda_rusko", | |
"short_name": "P-Rusko", | |
"category": "NLI", | |
"abbreviation": "P-RUS" | |
}, | |
"benczechmark_sentiment_mall": { | |
"name": "CzechSentiment MALL", | |
"source_url": "https://huggingface.co/datasets/CZLC/mall_sentiment_balanced", | |
"short_name": "S-Mall", | |
"category": "Sentiment", | |
"abbreviation": "S-MALL" | |
}, | |
"benczechmark_sentiment_fb": { | |
"name": "CzechSentiment FB", | |
"source_url": "https://huggingface.co/datasets/CZLC/fb_sentiment_balanced", | |
"short_name": "S-FB", | |
"category": "Sentiment", | |
"abbreviation": "S-FB" | |
}, | |
"benczechmark_sentiment_csfd": { | |
"name": "CzechSentiment CSFD", | |
"source_url": "https://huggingface.co/datasets/CZLC/csfd_sentiment_balanced", | |
"short_name": "S-CSFD", | |
"category": "Sentiment", | |
"abbreviation": "S-CSFD" | |
}, | |
"benczechmark_summarization": { | |
"name": "SUMECZECH", | |
"source_url": "https://huggingface.co/datasets/CZLC/sumeczech_downsampled", | |
"short_name": "Summarization", | |
"category": "Summarization", | |
"abbreviation": "SUM" | |
}, | |
"benczechmark_grammarerrorcorrection": { | |
"name": "GrammarErrorCorrection", | |
"source_url": "https://huggingface.co/datasets/CZLC/cs_gec", | |
"short_name": "Grammar Error Correction", | |
"category": "Syntactical Reasoning", | |
"abbreviation": "GEC" | |
}, | |
"benczechmark_cs_naturalquestions": { | |
"name": "NaturalQuestions-CZ", | |
"source_url": "https://huggingface.co/datasets/CZLC/cs_naturalquestions", | |
"short_name": "CS Natural Questions", | |
"category": "Knowledge", | |
"abbreviation": "NQ" | |
}, | |
"benczechmark_cs_sqad32": { | |
"name": "SQAD3.2", | |
"source_url": "https://huggingface.co/datasets/CZLC/SQAD_3.2", | |
"short_name": "CS SQAD 3.2", | |
"category": "Knowledge", | |
"abbreviation": "SQAD32" | |
}, | |
"benczechmark_cs_triviaQA": { | |
"name": "TriviaQA-CZ", | |
"source_url": "https://huggingface.co/datasets/CZLC/cs_triviaqa", | |
"short_name": "CS TriviaQA", | |
"category": "Knowledge", | |
"abbreviation": "TQA" | |
}, | |
"benczechmark_csfever_nli": { | |
"name": "CSFEVER", | |
"source_url": "https://huggingface.co/datasets/CZLC/ctu-aic/csfever_nli", | |
"short_name": "CSFever NLI", | |
"category": "NLI", | |
"abbreviation": "CFR" | |
}, | |
"benczechmark_ctkfacts_nli": { | |
"name": "CTKFACTS", | |
"source_url": "https://huggingface.co/datasets/CZLC/ctu-aic/ctkfacts_nli", | |
"short_name": "CTKFacts NLI", | |
"category": "NLI", | |
"abbreviation": "CTK" | |
}, | |
"benczechmark_cs_ner": { | |
"name": "CZECH NER CORPUS 2.0", | |
"source_url": "https://huggingface.co/datasets/CZLC/fewshot-goes-multilingual/cs_czech-named-entity-corpus_2.0", | |
"short_name": "CS NER", | |
"category": "NER", | |
"abbreviation": "CZNERC" | |
}, | |
"benczechmark_hellaswag": { | |
"name": "HellaSwag-CZ", | |
"source_url": "https://huggingface.co/datasets/CZLC/cs_hellaswag", | |
"short_name": "HellaSwag", | |
"category": "Language Modeling", | |
"abbreviation": "HASG" | |
}, | |
"benczechmark_histcorpus": { | |
"name": "Historical Corpus", | |
"source_url": "https://huggingface.co/datasets/CZLC/benczechmark_histcorpus", | |
"short_name": "HistCorpus", | |
"category": "Language Modeling", | |
"abbreviation": "HIST" | |
}, | |
"benczechmark_klokan_qa": { | |
"name": "Klokan QA", | |
"source_url": "https://huggingface.co/datasets/hynky/klokan-qa", | |
"short_name": "Klokan QA", | |
"category": "Czech Math Reasoning", | |
"abbreviation": "KQA" | |
}, | |
"benczechmark_cs_court_decisions_ner": { | |
"name": "Czech Court Decisions", | |
"source_url": "https://huggingface.co/datasets/CZLC/fewshot-goes-multilingual/cs_czech-court-decisions-ner", | |
"short_name": "CS Court Decisions NER", | |
"category": "NER", | |
"abbreviation": "CCDNER" | |
}, | |
"benczechmark_umimeto_biology": { | |
"name": "Umimeto.cz β Biology", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-biology", | |
"short_name": "Umimeto.cz β Biology", | |
"category": "General Reasoning", | |
"abbreviation": "UT-BIO" | |
}, | |
"benczechmark_umimeto_chemistry": { | |
"name": "Umimeto.cz β Chemistry", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-chemistry", | |
"short_name": "Umimeto.cz β Chemistry", | |
"category": "General Reasoning", | |
"abbreviation": "UT-CHEM" | |
}, | |
"benczechmark_umimeto_czech": { | |
"name": "Umimeto.cz β Czech Language", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-czech", | |
"short_name": "Umimeto.cz β Czech", | |
"category": "General Reasoning", | |
"abbreviation": "UT-CZEL" | |
}, | |
"benczechmark_umimeto_history": { | |
"name": "Umimeto.cz β History", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-history", | |
"short_name": "Umimeto.cz β History", | |
"category": "General Reasoning", | |
"abbreviation": "UT-HIST" | |
}, | |
"benczechmark_umimeto_informatics": { | |
"name": "Umimeto.cz β Informatics", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-informatics", | |
"short_name": "Umimeto.cz β Informatics", | |
"category": "General Reasoning", | |
"abbreviation": "UT-IT" | |
}, | |
"benczechmark_umimeto_math": { | |
"name": "Umimeto.cz β Math", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-math", | |
"short_name": "Umimeto.cz β Math", | |
"category": "Czech Math Reasoning", | |
"abbreviation": "UT-MATH" | |
}, | |
"benczechmark_umimeto_physics": { | |
"name": "Umimeto.cz β Physics", | |
"source_url": "https://huggingface.co/datasets/CZLC/umimeto-physics", | |
"short_name": "Umimeto.cz β Physics", | |
"category": "General Reasoning", | |
"abbreviation": "UT-PHYS" | |
}, | |
"benczechmark_cermat_czmath_mc": { | |
"name": "CERMAT β Czech Math β MC", | |
"source_url": "https://huggingface.co/datasets/CZLC/cermat_math_mc", | |
"short_name": "Cermat Czech Math MC", | |
"category": "Czech Math Reasoning", | |
"abbreviation": "CCM-MC" | |
}, | |
"benczechmark_cermat_czmath_open": { | |
"name": "CERMAT β Czech Math β OPEN", | |
"source_url": "https://huggingface.co/datasets/CZLC/cermat_math_open", | |
"short_name": "Cermat Czech Math Open", | |
"category": "Czech Math Reasoning", | |
"abbreviation": "CCM-OPEN" | |
}, | |
"benczechmark_cermat_czech_tf": { | |
"name": "CERMAT β Czech Language β TF", | |
"source_url": "https://huggingface.co/datasets/CZLC/cermat_czech_tf", | |
"short_name": "Cermat Czech Language TF", | |
"category": "General Reasoning", | |
"abbreviation": "CCL-TF" | |
}, | |
"benczechmark_cermat_czech_mc": { | |
"name": "CERMAT β Czech Language β MC", | |
"source_url": "https://huggingface.co/datasets/CZLC/cermat_czech_mc", | |
"short_name": "Cermat Czech Language MC", | |
"category": "General Reasoning", | |
"abbreviation": "CCL-MC" | |
}, | |
"benczechmark_cermat_czech_open": { | |
"name": "CERMAT β Czech Language β OPEN", | |
"source_url": "https://huggingface.co/datasets/CZLC/cermat_czech_open", | |
"short_name": "Cermat Czech Language Open", | |
"category": "General Reasoning", | |
"abbreviation": "CCL-OPEN" | |
}, | |
"benczechmark_history_ir": { | |
"name": "Historical Relelvance Grading", | |
"source_url": "https://huggingface.co/datasets/CZLC/history_retrieval", | |
"abbreviation": "HIST-IR", | |
"category": "Historical", | |
"short_name": "Czech History IR" | |
}, | |
"benczechmark_agree": { | |
"name": "Agree", | |
"source_url": "https://huggingface.co/datasets/davidadamczyk/czechbench_agree", | |
"abbreviation": "Agree", | |
"category": "Syntactical Reasoning", | |
"short_name": "Agree" | |
}, | |
"benczechmark_belebele": { | |
"name": "Belebele", | |
"source_url": "https://huggingface.co/datasets/davidadamczyk/czechbench_belebele", | |
"abbreviation": "BB", | |
"category": "Knowledge", | |
"short_name": "Belebele" | |
}, | |
"benczechmark_czechnews": { | |
"name": "CZ-NEWS", | |
"source_url": "https://huggingface.co/datasets/davidadamczyk/czechbench_czech_news", | |
"abbreviation": "CZ-NEWS", | |
"category": "Topic Classification", | |
"short_name": "Czech News" | |
}, | |
"benczechmark_subjectivity": { | |
"name": "Subjectivity", | |
"source_url": "https://huggingface.co/datasets/davidadamczyk/czechbench_subjectivity", | |
"abbreviation": "SUBJ", | |
"category": "Sentiment", | |
"short_name": "Subjectivity" | |
}, | |
"benczechmark_essay": { | |
"name": "Czech National Corpus β Essays", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_skript12", | |
"abbreviation": "CNC-E", | |
"category": "Language Modeling", | |
"short_name": "CNC β Essays" | |
}, | |
"benczechmark_fiction": { | |
"name": "Czech National Corpus β Fiction", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_fictree", | |
"abbreviation": "CNC-F", | |
"category": "Language Modeling", | |
"short_name": "CNC β Fiction" | |
}, | |
"benczechmark_capek": { | |
"name": "Czech National Corpus β Karel Δapek", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_Capek", | |
"abbreviation": "CNC-CAP", | |
"category": "Language Modeling", | |
"short_name": "CNC β Δapek" | |
}, | |
"benczechmark_correspondence": { | |
"name": "Czech National Corpus β Correspondence", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_KSK", | |
"abbreviation": "CNC-KSK", | |
"category": "Language Modeling", | |
"short_name": "CNC β Correspondence" | |
}, | |
"benczechmark_havlicek": { | |
"name": "Czech National Corpus β Karel HavlΓΔek β Noviny", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_KHavlicek_HistNews", | |
"abbreviation": "CNC-KH", | |
"category": "Language Modeling", | |
"short_name": "CNC β KHavlicek β HistNews" | |
}, | |
"benczechmark_speeches": { | |
"name": "Czech National Corpus β Speeches", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_PrezPrejavy", | |
"abbreviation": "CNC-SPCH", | |
"category": "Language Modeling", | |
"short_name": "CNC β Speeches" | |
}, | |
"benczechmark_spoken": { | |
"name": "Czech National Corpus β Spoken", | |
"source_url": "https://huggingface.co/datasets/CZLC/CNC_oral_ortofon", | |
"abbreviation": "CNC-SPK", | |
"category": "Language Modeling", | |
"short_name": "CNC β Spoken" | |
}, | |
"benczechmark_snli": { | |
"name": "Czech SNLI", | |
"source_url": "https://huggingface.co/datasets/CZLC/snli", | |
"abbreviation": "SNLI", | |
"category": "NLI", | |
"short_name": "Czech SNLI" | |
} | |
} |