|
|
|
|
|
"""Model evaluation""" |
|
|
|
from megatron import get_args |
|
from megatron import print_rank_0 |
|
from tasks.msdp.metrics import F1Metric |
|
from tqdm import tqdm |
|
|
|
|
|
def evaluate_f1(guess_file, answer_file): |
|
"""Evaluating F1 Score""" |
|
|
|
guess_list = [] |
|
print_rank_0('reading %s' % guess_file) |
|
with open(guess_file, "r") as f: |
|
for i, line in enumerate(tqdm(f)): |
|
line = line.strip() |
|
if "<|endoftext|>" in line: |
|
line = line.replace("<|endoftext|>", "") |
|
guess_list.append(line) |
|
|
|
answer_list = [] |
|
print_rank_0('reading %s' % answer_file) |
|
with open(answer_file, "r") as f: |
|
for i, line in enumerate(tqdm(f)): |
|
line = line.strip() |
|
if line == "no_passages_used": |
|
line = "" |
|
answer_list.append(line) |
|
|
|
assert len(guess_list) == len(answer_list), \ |
|
"lengths of guess and answer are different!" |
|
|
|
precision, recall, f1 = F1Metric.compute_all_pairs(guess_list, answer_list) |
|
print_rank_0('Precision: %.4f; recall: %.4f; f1: %.4f' % (precision, recall, f1)) |
|
|
|
print_rank_0('done :-)') |
|
|
|
|
|
def main(): |
|
args = get_args() |
|
|
|
evaluate_f1(args.guess_file, args.answer_file) |
|
|
|
|