|
--- |
|
datasets: |
|
- hatexplain |
|
language: |
|
- en |
|
pipeline_tag: text-classification |
|
metrics: |
|
- accuracy |
|
- f1 |
|
- precision |
|
- recall |
|
--- |
|
# BERT for hate speech classification |
|
The model is based on BERT and used for classifying a text as **toxic** and **non-toxic**. It achieved an **F1** score of **0.81** and an **Accuracy** of **0.77**. |
|
|
|
The model was fine-tuned on the HateXplain dataset found here: https://huggingface.co/datasets/hatexplain |
|
|
|
## How to use |
|
|
|
```python |
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline |
|
|
|
# Load model and tokenizer |
|
tokenizer = AutoTokenizer.from_pretrained('tum-nlp/bert-hateXplain') |
|
model = AutoModelForSequenceClassification.from_pretrained('tum-nlp/bert-hateXplain') |
|
|
|
# Create the pipeline for classification |
|
hate_classifier = pipeline("text-classification", model=model, tokenizer=tokenizer) |
|
|
|
# Predict |
|
hate_classifier("I like you. I love you") |
|
|
|
``` |