MT Sentinel Metrics
Collection
Machine Translation (MT) metrics designed explicitly to scrutinize the MT meta-evaluation process’s accuracy, robustness, and fairness.
•
7 items
•
Updated
•
6
This repository contains the SENTINELSRC metric model pre-trained on Direct Assessments (DA) annotations and further fine-tuned on Multidimensional Quality Metrics (MQM) data. For details on how to use our sentinel metric models, check our GitHub repository.
After having installed our repository package, you can use this model within Python in the following way:
from sentinel_metric import download_model, load_from_checkpoint
model_path = download_model("sapienzanlp/sentinel-src-mqm")
model = load_from_checkpoint(model_path)
data = [
{"src": "E.T.打电话回家"},
{"src": "托托,我感觉我们已经不在堪萨斯了"}
]
output = model.predict(data, batch_size=8, gpus=1)
Output:
# Segment scores
>>> output.scores
[0.6295096278190613, 0.46640336513519287]
# System score
>>> output.system_score
0.5479564964771271
This work has been published at ACL 2024 (Main Conference). If you use any part, please consider citing our paper as follows:
@inproceedings{perrella-etal-2024-guardians,
title = "Guardians of the Machine Translation Meta-Evaluation: Sentinel Metrics Fall In!",
author = "Perrella, Stefano and Proietti, Lorenzo and Scir{\`e}, Alessandro and Barba, Edoardo and Navigli, Roberto",
editor = "Ku, Lun-Wei and Martins, Andre and Srikumar, Vivek",
booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = aug,
year = "2024",
address = "Bangkok, Thailand", publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2024.acl-long.856",
pages = "16216--16244",
}
This work is licensed under Creative Commons Attribution-ShareAlike-NonCommercial 4.0.