mbart-large-cc25-ruquad-qag / eval /metric.first.answer.paragraph.questions_answers.lmqg_qag_ruquad.default.json
asahi417's picture
model update
0a5a177
raw
history blame
1.17 kB
{"validation": {"Bleu_1": 0.13012441026569896, "Bleu_2": 0.08195993433114934, "Bleu_3": 0.0580470215436035, "Bleu_4": 0.043074391952444474, "METEOR": 0.23451788039746743, "ROUGE_L": 0.1854271551591915, "BERTScore": 0.6946748925003534, "MoverScore": 0.5225861498803709, "QAAlignedF1Score (BERTScore)": 0.7735665088118578, "QAAlignedRecall (BERTScore)": 0.8000333043546412, "QAAlignedPrecision (BERTScore)": 0.74998383768986, "QAAlignedF1Score (MoverScore)": 0.5616079950472169, "QAAlignedRecall (MoverScore)": 0.5810802673098598, "QAAlignedPrecision (MoverScore)": 0.5450698084042794}, "test": {"Bleu_1": 0.13128563485015673, "Bleu_2": 0.08287119857965321, "Bleu_3": 0.058526142878210216, "Bleu_4": 0.04325857709320956, "METEOR": 0.23520885191169122, "ROUGE_L": 0.18588244856631136, "BERTScore": 0.6957883092191999, "MoverScore": 0.5228671786592183, "QAAlignedF1Score (BERTScore)": 0.7736153369991706, "QAAlignedRecall (BERTScore)": 0.8005435576998596, "QAAlignedPrecision (BERTScore)": 0.7496743080623613, "QAAlignedF1Score (MoverScore)": 0.5609644794916299, "QAAlignedRecall (MoverScore)": 0.5810727037427397, "QAAlignedPrecision (MoverScore)": 0.5439870886271599}}