|
from Spider import dump_db_json_schema |
|
import json |
|
import os |
|
import copy |
|
from utils import write_jsonl_file, parse |
|
import shutil |
|
|
|
|
|
def preprocess(args, split): |
|
data_filepaths = [ |
|
os.path.join( |
|
os.path.join(args.input_dir, "sql_state_tracking"), f"cosql_{split}.json" |
|
) |
|
] |
|
db_path = os.path.join(args.input_dir, "database") |
|
out_db_path = os.path.join(args.output_dir, "database") |
|
|
|
if not os.path.exists(out_db_path): |
|
shutil.copytree(db_path, out_db_path) |
|
|
|
schema_cache = dict() |
|
processed_data = [] |
|
for data_filepath in data_filepaths: |
|
with open(data_filepath, encoding="utf-8") as f: |
|
cosql = json.load(f) |
|
|
|
for sample in cosql: |
|
dialog = { |
|
"locale": "en", |
|
"dialog": [], |
|
} |
|
db_id = sample["database_id"] |
|
if db_id not in schema_cache: |
|
schema_cache[db_id] = dump_db_json_schema( |
|
db_path + "/" + db_id + "/" + db_id + ".sqlite", db_id |
|
) |
|
schema = schema_cache[db_id] |
|
|
|
dialog["knowledge"] = { |
|
"type": "dict", |
|
"value": {"db_id": db_id, "schema": schema, "db_path": out_db_path}, |
|
} |
|
|
|
final_dialog = copy.deepcopy(dialog) |
|
final_dialog["turn"] = "single" |
|
final_dialog["dialog"].append( |
|
{ |
|
"roles": ["USER"], |
|
"utterance": sample["final"]["utterance"] |
|
.replace("``", '"') |
|
.replace("''", '"') |
|
.strip(), |
|
"sql": sample["final"]["query"], |
|
} |
|
) |
|
|
|
if split == "train": |
|
processed_data.append(final_dialog) |
|
dialog["turn"] = "multi" |
|
roles = ["USER", "SYSTEM"] |
|
|
|
for _, turn in enumerate(sample["interaction"]): |
|
for idx, utterance in enumerate( |
|
turn["utterance"] |
|
.replace("``", '"') |
|
.replace("''", '"') |
|
.strip() |
|
.split("|") |
|
): |
|
utterance = utterance.strip() |
|
|
|
dialog["dialog"].append( |
|
{"roles": [roles[idx % 2]], "utterance": utterance} |
|
) |
|
|
|
assert dialog["dialog"][-1]["roles"] == ["USER"] |
|
dialog["dialog"][-1]["sql"] = turn["query"] |
|
|
|
processed_data.append(dialog) |
|
|
|
write_jsonl_file(processed_data, os.path.join(args.output_dir, f"{split}.jsonl")) |
|
|
|
|
|
if __name__ == "__main__": |
|
args = parse() |
|
preprocess(args, "train") |
|
preprocess(args, "dev") |
|
|
|
shutil.copyfile( |
|
os.path.join(args.input_dir, "tables.json"), |
|
os.path.join(args.output_dir, "tables.json"), |
|
) |
|
|