Edit model card

Model Card for POLLCHECK/RoBERTa-classifier

Model Description

This RoBERTa model has been fine-tuned for a binary classification task to determine whether statements are 0 OR "biased/ fake" or 1 OR "unbiased/ real". The model is based on the RoBERTa architecture, a robustly optimized BERT pretraining approach.

Intended Use

Primary Use: This model is intended for the classification of textual statements into two categories: biased and unbiased. It is suitable for analyzing news articles, editorials, and opinion pieces. Users: This model can be used by data scientists, journalists, content moderators, and social media platforms to detect bias in text.

Model Details

Architecture: The model uses the RoBERTa-base architecture. Training Data: The model was trained on a curated dataset comprising news articles, editorials, and opinion pieces labeled as biased or unbiased by domain experts. Performance Metrics

Usage

import torch

model_name = "POLLCHECK/RoBERTa-classifier"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSequenceClassification.from_pretrained(model_name)

texts = [
    "Religious Extremists Threaten Our Way of Life.",
    "Public Health Officials are working."
]
for text in texts:
    inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True)
    with torch.no_grad():
        outputs = model(**inputs)
        probabilities = torch.softmax(outputs.logits, dim=-1)
        predicted_label = "biased" if probabilities[0][0] > 0.5 else "unbiased"
        print(f"Text: {text}\nPredicted label: {predicted_label}")

Results

The following table presents the evaluation metrics for each class along with macro averages:

Class Precision Recall F1-Score
Biased/ fake (0) 0.93 0.96 0.94
Unbiased/ real (1) 0.96 0.92 0.94
Macro Avg 0.94 0.94 0.94
Downloads last month
2
Safetensors
Model size
125M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.