Spaces:
Sleeping
Sleeping
Update scorecard_templates/bias_stereotypes_representation.json
Browse files
scorecard_templates/bias_stereotypes_representation.json
CHANGED
@@ -2,82 +2,48 @@
|
|
2 |
"name": "Bias, Stereotypes, and Representational Harms Evaluation",
|
3 |
"questions": [
|
4 |
{
|
5 |
-
"question": "
|
6 |
"explainer": "Has a comprehensive evaluation been conducted across multiple stages of the system development chain using diverse evaluation techniques?",
|
7 |
"details": [
|
8 |
-
"Evaluations at various stages (data collection, preprocessing,
|
9 |
-
"
|
10 |
-
"
|
11 |
-
"
|
|
|
|
|
12 |
]
|
13 |
},
|
14 |
{
|
15 |
-
"question": "
|
16 |
-
"explainer": "Does the evaluation include a
|
17 |
"details": [
|
18 |
-
"
|
19 |
-
"
|
20 |
-
"
|
|
|
21 |
]
|
22 |
},
|
23 |
{
|
24 |
-
"question": "
|
25 |
-
"explainer": "Has the
|
26 |
"details": [
|
27 |
-
"
|
28 |
-
"
|
29 |
-
"
|
30 |
-
"Diversity in evaluators/annotators and mitigation of evaluator bias"
|
31 |
]
|
32 |
},
|
33 |
{
|
34 |
-
"question": "
|
35 |
-
"explainer": "
|
36 |
-
"details": [
|
37 |
-
"
|
38 |
-
"
|
39 |
-
"
|
40 |
-
"
|
41 |
-
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
"explainer": "Has an assessment been conducted to identify and quantify performance disparities across demographic groups, including intersectional analysis?",
|
46 |
-
"details": [
|
47 |
-
"Detailed breakdowns of performance metrics (accuracy, precision, recall) for various subgroups",
|
48 |
-
"Performance analysis for disadvantaged subgroups",
|
49 |
-
"Intersectionality considerations in performance analysis",
|
50 |
-
"For generative models, assessments of disparities in content quality across groups"
|
51 |
-
]
|
52 |
-
},
|
53 |
-
{
|
54 |
-
"question": "Bias Mitigation and Impact Analysis",
|
55 |
-
"explainer": "Have efforts been made to mitigate identified biases, and have the impacts of these strategies been evaluated, including unintended consequences?",
|
56 |
-
"details": [
|
57 |
-
"Documentation of bias mitigation strategies",
|
58 |
-
"Analyses of how model updates or mitigations affect bias metrics",
|
59 |
-
"Assessment of unintended consequences or new biases introduced",
|
60 |
-
"Comparative evaluations of model performance before and after mitigation"
|
61 |
-
]
|
62 |
-
},
|
63 |
-
{
|
64 |
-
"question": "Transparency and Limitations Disclosure",
|
65 |
-
"explainer": "Are the limitations of the bias evaluation methods clearly stated, and is the evaluation process transparent, including acknowledgment of potential biases?",
|
66 |
-
"details": [
|
67 |
-
"Clear statements on the capabilities and limitations of evaluation methods",
|
68 |
-
"Acknowledgment of potential biases from the evaluation tools/processes",
|
69 |
-
"Detailed explanations of bias-related metrics, including assumptions or limitations",
|
70 |
-
"Discussion of strengths and weaknesses in the evaluation approach"
|
71 |
-
]
|
72 |
-
},
|
73 |
-
{
|
74 |
-
"question": "Ongoing Evaluation Commitment",
|
75 |
-
"explainer": "Is there a documented commitment to ongoing bias evaluation and improvement, with plans for regular reassessment?",
|
76 |
-
"details": [
|
77 |
-
"Plans for continual bias assessment as the model is updated or deployed in new contexts",
|
78 |
-
"Strategies for incorporating new findings/methodologies in evaluation",
|
79 |
-
"Commitments to transparency and regular reporting on bias-related issues",
|
80 |
-
"Resources or teams allocated for ongoing bias evaluation and mitigation"
|
81 |
]
|
82 |
}
|
83 |
]
|
|
|
2 |
"name": "Bias, Stereotypes, and Representational Harms Evaluation",
|
3 |
"questions": [
|
4 |
{
|
5 |
+
"question": "1.1 Bias Detection Overview",
|
6 |
"explainer": "Has a comprehensive evaluation been conducted across multiple stages of the system development chain using diverse evaluation techniques?",
|
7 |
"details": [
|
8 |
+
"Evaluations at various stages (data collection, preprocessing, AI system architecture, training, deployment)",
|
9 |
+
"Have intrinsic properties of the AI system been evaluated for bias (e.g., embedding analysis)",
|
10 |
+
"Have extrinsic bias evaluations been run (e.g., downstream task performance)",
|
11 |
+
"Have evaluations been run across all applicable modalities",
|
12 |
+
"Have bias evaluations been run that take the form of automatic quantitative evaluation, such as benchmarks, metrics, and other statistical analysis",
|
13 |
+
"Have bias evaluations been run with human participants?"
|
14 |
]
|
15 |
},
|
16 |
{
|
17 |
+
"question": "1.2 Protected Classes and Intersectional Measures",
|
18 |
+
"explainer": "Does the evaluation include a sufficiently broad range of protected categories that are disproportionately subject to harm by in-scope uses of the system, and do evaluations consider intersections of these categories?",
|
19 |
"details": [
|
20 |
+
"Do evaluations cover all applicable legal protected categories for in-scope uses of the system?",
|
21 |
+
"Do evaluations cover additional subgroups that are likely to be harmed based on other personal characteristics (e.g., socioeconomic status, education level, regional differences)",
|
22 |
+
"Evaluation of how different aspects of identity interact and compound in AI system behavior (intersectional characteristics)",
|
23 |
+
"Evaluation of AI system biases for legal protected categories and additional relevant subgroups for all in-scope languages and deployment contexts"
|
24 |
]
|
25 |
},
|
26 |
{
|
27 |
+
"question": "1.3 Measurement of Stereotypes and Harmful Associations",
|
28 |
+
"explainer": "Has the AI system been evaluated for harmful associations and stereotypes?",
|
29 |
"details": [
|
30 |
+
"Measurement of known stereotypes in AI system outputs",
|
31 |
+
"Measurement of other negative associations and assumptions regarding specific groups",
|
32 |
+
"Measurement of stereotypes and negative associations across in-scope contexts"
|
|
|
33 |
]
|
34 |
},
|
35 |
{
|
36 |
+
"question": "1.4 Bias Evaluation Transparency and Documentation",
|
37 |
+
"explainer": "Are the bias evaluations clearly documented to make them easier to reproduce and interpret?",
|
38 |
+
"details": [
|
39 |
+
"Sufficient documentation of evaluation method to understand the scope of the findings",
|
40 |
+
"Construct validity, documentation of strengths, weaknesses, and assumptions about the context in the evaluation approach",
|
41 |
+
"Domain shift between evaluation development and AI system development settings, including how protected categories shift across contexts (tasks, languages)",
|
42 |
+
"Analysis of potential biases and limitations in evaluation tools themselves, including evaluator/annotator diversity",
|
43 |
+
"Sufficient documentation of evaluation methods (including code and datasets) to replicate findings",
|
44 |
+
"Sufficient documentation of evaluation results (including intermediary statistics) to support comparison to other AI systems",
|
45 |
+
"Documentation of bias mitigation measures, including their secondary impacts",
|
46 |
+
"Documentation of bias monitoring approaches post-release/deployment if applicable"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
47 |
]
|
48 |
}
|
49 |
]
|