|
--- |
|
license: mit |
|
datasets: |
|
- ethical-spectacle/biased-corpus |
|
language: |
|
- en |
|
metrics: |
|
- f1(0.8998) |
|
- precision |
|
- recall() |
|
library_name: transformers |
|
co2_eq_emissions: |
|
emissions: 10 |
|
source: Code Carbon |
|
training_type: fine-tuning |
|
geographical_location: Albany, New York |
|
hardware_used: T4 |
|
base_model: |
|
- google-bert/bert-base-uncased |
|
pipeline_tag: text-classification |
|
tags: |
|
- Social Bias |
|
--- |
|
|
|
## How to Use |
|
``` |
|
classifier = pipeline("text-classification", model="maximuspowers/bias-type-classifier") // pass in return_all_scores=True for multi-label |
|
result = classifier("Tall people are so clumsy") |
|
|
|
// Example Result |
|
// [ |
|
// { |
|
// "label": "physical", |
|
// "score": 0.9972801208496094 |
|
// } |
|
// ] |
|
``` |
|
|
|
This model was trained on a [synthetic dataset](https://huggingface.co/datasets/ethical-spectacle/biased-corpus) of biased statements and questions, generated by Mistal 7B as part of the [GUS-Net paper](https://arxiv.org/abs/2410.08388). |
|
|
|
### Model Performance: |
|
| Label | F1 Score | Precision | Recall | |
|
|-----------------|----------|-----------|--------| |
|
| **Macro Average** | **0.8998** | **0.9213** | **0.8807** | |
|
| racial | 0.8613 | 0.9262 | 0.8049 | |
|
| religious | 0.9655 | 0.9716 | 0.9595 | |
|
| gender | 0.9160 | 0.9099 | 0.9223 | |
|
| age | 0.9185 | 0.9683 | 0.8737 | |
|
| nationality | 0.9083 | 0.9053 | 0.9113 | |
|
| sexuality | 0.9304 | 0.9484 | 0.9131 | |
|
| socioeconomic | 0.8273 | 0.8727 | 0.7864 | |
|
| educational | 0.8791 | 0.9091 | 0.8511 | |
|
| disability | 0.8713 | 0.8762 | 0.8665 | |
|
| political | 0.9127 | 0.8914 | 0.9351 | |
|
| physical | 0.9069 | 0.9547 | 0.8635 | |
|
|
|
### Training Params: |
|
**Learning Rate:** 5e-5 |
|
**Batch Size:** 16 |
|
**Epochs:** 3 |