Spaces:
Runtime error
Runtime error
import os | |
import argparse | |
import json | |
from llava.eval.m4c_evaluator import EvalAIAnswerProcessor | |
def parse_args(): | |
parser = argparse.ArgumentParser() | |
parser.add_argument('--annotation-file', type=str, required=True) | |
parser.add_argument('--result-file', type=str, required=True) | |
parser.add_argument('--result-upload-file', type=str, required=True) | |
return parser.parse_args() | |
if __name__ == '__main__': | |
args = parse_args() | |
os.makedirs(os.path.dirname(args.result_upload_file), exist_ok=True) | |
results = [] | |
error_line = 0 | |
for line_idx, line in enumerate(open(args.result_file)): | |
try: | |
results.append(json.loads(line)) | |
except: | |
error_line += 1 | |
results = {x['question_id']: x['text'] for x in results} | |
test_split = [json.loads(line) for line in open(args.annotation_file)] | |
split_ids = set([x['question_id'] for x in test_split]) | |
print(f'total results: {len(results)}, total split: {len(test_split)}, error_line: {error_line}') | |
all_answers = [] | |
answer_processor = EvalAIAnswerProcessor() | |
for x in test_split: | |
assert x['question_id'] in results | |
all_answers.append({ | |
'image': x['image'], | |
'answer': answer_processor(results[x['question_id']]) | |
}) | |
with open(args.result_upload_file, 'w') as f: | |
json.dump(all_answers, f) | |