Spaces:
Building
Building
# constants | |
OVERALL_INFO = ["Model", "Overall"] | |
risk_topic_1_columns = [ | |
"Crimes and Illegal Activities", | |
"Cybersecurity", | |
"Data Privacy", | |
"Ethics and Morality", | |
"Physical and Mental Health", | |
"Hate Speech", | |
"Extremism", | |
"Inappropriate Suggestions" | |
] | |
risk_topic_1_columns = [item.lower() for item in risk_topic_1_columns] | |
attack_columns = [ | |
"Adaptive Attack", | |
"Positive Induction", | |
"Reverse Induction", | |
"Code Injection", | |
"Instruction Jailbreak", | |
"Goal Hijacking", | |
"Instruction Encryption", | |
"DeepInception", | |
"In-Context Attack", | |
"Chain of Utterances", | |
"Compositional Instructions" | |
] | |
attack_columns = [item.lower() for item in attack_columns] | |
XLSX_DIR = "./file//results.xlsx" | |
LEADERBOARD_INTRODUCTION = """# π S-Eval Leaderboard | |
You can get more detailed information from our [project](https://github.com/IS2Lab/S-Eval) and [Paper](https://arxiv.org/abs/2405.14191). | |
""" | |
SELECT_SET_INTRO = ( | |
"Select whether Chinese or English results should be shown." | |
) | |
TABLE_INTRODUCTION_1 = """In the table below, we summarize the safety scores (%) of differnet models on Base Risk Prompt Set.""" | |
TABLE_INTRODUCTION_2 = """In the table below, we summarize the attack success rate (%) of the instruction attacks in Attack Prompt Set on different models""" | |
LEADERBORAD_INFO = """ | |
S-Eval is a new comprehensive, multi-dimensional and open-ended safety evaluation benchmark. It consists of 220,000 evaluation prompts, including 20,000 base risk prompts (10,000 in Chinese and 10,000 in English) and 200, 000 corresponding attack prompts derived from 10 popular adversarial instruction attacks. These test prompts are generated based on a comprehensive and unified risk taxonomy. The risk taxonomy has a structured hierarchy with four levels, comprising 8 risk dimensions, 25 risk categories, 56 risk subcategories, and 52 risk sub-subcategories, specifically designed to encompass all crucial dimensions of safety evaluation and accurately reflect the varied safety levels of LLMs across different risk dimensions. | |
""" | |
CITATION_BUTTON_LABEL = "If our work is useful for your own, you can cite us with the following BibTex entry:" | |
CITATION_BUTTON_TEXT = r""" | |
@article{yuan2024seval, | |
title={S-Eval: Automatic and Adaptive Test Generation for Benchmarking Safety Evaluation of Large Language Models}, | |
author={Xiaohan Yuan and Jinfeng Li and Dongxia Wang and Yuefeng Chen and Xiaofeng Mao and Longtao Huang and Hui Xue and Wenhai Wang and Kui Ren and Jingyi Wang}, | |
journal={arXiv preprint arXiv:2405.14191}, | |
year={2024} | |
} | |
""" | |