import json import tqdm import numpy as np import multiprocessing as mp import random from collections import Counter random.seed(13) def _norm(x): return ' '.join(x.strip().split()) strategies = json.load(open('./strategy.json')) strategies = [e[1:-1] for e in strategies] strat2id = {strat: i for i, strat in enumerate(strategies)} original = json.load(open('./ESConv.json')) def process_data(d): emotion = d['emotion_type'] problem = d["problem_type"] situation = d['situation'] #init_intensity = int(d['score']['speaker']['begin_intensity']) #final_intensity = int(d['score']['speaker']['end_intensity']) d = d['dialog'] dial = [] for uttr in d: text = _norm(uttr['content']) role = uttr['speaker'] if role == 'seeker': dial.append({ 'text': text, 'speaker': 'usr', }) else: dial.append({ 'text': text, 'speaker': 'sys', 'strategy': uttr['annotation']['strategy'], }) res = { 'emotion_type': emotion, 'problem_type': problem, 'situation': situation, #'init_intensity': init_intensity, #'final_intensity': final_intensity, 'dialog': dial, } return res data = [] for e in map(process_data, tqdm.tqdm(original, total=len(original))): data.append(e) emotions = Counter([e['emotion_type'] for e in data]) problems = Counter([e['problem_type'] for e in data]) print('emotion', emotions) print('problem', problems) random.shuffle(data) dev_size = int(0.15 * len(data)) test_size = int(0.15 * len(data)) valid = data[:dev_size] test = data[dev_size: dev_size + test_size] train = data[dev_size + test_size:] print('train', len(train)) with open('./train.txt', 'w') as f: for e in train: f.write(json.dumps(e) + '\n') with open('./sample.json', 'w') as f: json.dump(train[:10], f, ensure_ascii=False, indent=2) print('valid', len(valid)) with open('./valid.txt', 'w') as f: for e in valid: f.write(json.dumps(e) + '\n') print('test', len(test)) with open('./test.txt', 'w') as f: for e in test: f.write(json.dumps(e) + '\n')