mt5-small-dequad-qg / eval /metric.long.sentence.sentence_answer.question.lmqg_qg_dequad.default.json
asahi417's picture
model update
c07208a
raw
history blame
537 Bytes
{"validation": {"Bleu_1": 0.10517833004054772, "Bleu_2": 0.04414234005912018, "Bleu_3": 0.018582377750217826, "Bleu_4": 8.43042343985824e-07, "METEOR": 0.10547673712340604, "ROUGE_L": 0.10280309247610869, "BERTScore": 0.7910826379244186, "MoverScore": 0.5422143723483254}, "test": {"Bleu_1": 0.09782721696641554, "Bleu_2": 0.041332169064365216, "Bleu_3": 0.017238838008138573, "Bleu_4": 0.005876220390273766, "METEOR": 0.10752287923132721, "ROUGE_L": 0.09906508180177784, "BERTScore": 0.78457647542309, "MoverScore": 0.5403259843868893}}