metadata
language: en
license: mit
library_name: transformers
tags:
- xlnet
- automatic-short-answer-grading
- regression
- education
- short-answer
- assessment
- grading
- transformers
pipeline_tag: text-classification
datasets:
- Meyerger/ASAG2024
metrics:
- mse
- rmse
- mae
- pearson correlation
model-index:
- name: XLENT_ASAG
results:
- task:
type: regression
name: automatic short answer grading
metrics:
- type: mse
value: 0.035
- type: rmse
value: 0.187
- type: mae
value: 0.142
- type: pearson correlation
value: 0.912
ASAG XLNet Regression Model
This model evaluates student answers by comparing them to reference answers and predicting a grade (regression).
Model Details
- Model Type: XLNet for Regression
- Task: Automatic Short Answer Grading (ASAG)
- Framework: PyTorch/Transformers
- Base Model: xlnet-base-cased
- Library: Transformers
Usage
from transformers import XLNetTokenizer, XLNetForSequenceClassification
import torch
# Load model and tokenizer
tokenizer = XLNetTokenizer.from_pretrained("kenzykhaled/XLENT_ASAG")
model = XLNetForSequenceClassification.from_pretrained("kenzykhaled/XLENT_ASAG")
# Prepare inputs
student_answer = "It is vision."
reference_answer = "The stimulus is seeing or hearing the cup fall."
inputs = tokenizer(
text=student_answer,
text_pair=reference_answer,
return_tensors="pt",
padding=True,
truncation=True
)
# Get prediction
with torch.no_grad():
outputs = model(**inputs)
# Get predicted grade (normalized between 0-1)
predicted_grade = outputs.logits.item()
predicted_grade = max(0, min(1, predicted_grade))
print(f"Predicted grade: {predicted_grade:.4f}")
Inference API Usage
This model can be used directly with the Hugging Face Inference API:
import requests
API_URL = "https://api-inference.huggingface.co/models/kenzykhaled/XLENT_ASAG"
headers = {"Authorization": "Bearer YOUR_HUGGING_FACE_TOKEN"}
def query(payload):
response = requests.post(API_URL, headers=headers, json=payload)
return response.json()
data = {
"inputs": {
"source_sentence": "It is vision.",
"sentences": ["The stimulus is seeing or hearing the cup fall."]
}
}
result = query(data)
print(result)
Training Data
This model was trained on the Meyerger/ASAG2024 dataset.
Use Cases
- Automated grading of student short-answer responses
- Educational technology platforms
- Learning management systems
- Assessment tools
- Teacher assistance for grading
Limitations
- The model is trained on specific educational domains and may not generalize well to all subjects
- Performance depends on the similarity of input data to the training data
- Should be used as an assistive tool for grading rather than a complete replacement for human evaluation
Ethical Considerations
When using this model for automated grading:
- Be transparent with students about the use of AI for grading
- Consider potential biases in evaluation
- Provide human review of edge cases
- Allow students to appeal automated grades