Roberta for Justification analyst
This model is a fine-tuned version of the Roberta architecture that has been trained specifically for sequence classification. The fine-tuning process involved using the PyTorch deep learning framework and specific hyperparameters (2-4e, 1-8 epsilon) with Adagrad optimizer.
Example Usage
To use the model, first load it in PyTorch:
import torch
from transformers import RobertaForSequenceClassification, RobertaTokenizer
# Load the fine-tuned model
model = RobertaForSequenceClassification.from_pretrained('Dzeniks/justification-analyst')
# Load the tokenizer
tokenizer = RobertaTokenizer.from_pretrained('Dzeniks/justification-analyst')
# Tokenize the input sequence
input_text = "This is a sample input sequence"
input = tokenizer.encode_plus(claim, evidence, return_tensors="pt")
# Use the model to make a prediction
model.eval()
with torch.no_grad():
prediction = model(**x)
predictions = torch.argmax(outputs[0], dim=1).item()
Classification Labels
The model was trained on a dataset consisting of claims and evidence, where the goal was to classify each claim as either supporting, refuting, or not having enough information to make a decision. The labels used for this task are as follows:
- Label 0: Supports
- Label 1: Refutes
- Label 2: Not enough information
- Downloads last month
- 11
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.