|
|
|
|
|
import argparse |
|
import json |
|
import os |
|
import sys |
|
|
|
pwd = os.path.abspath(os.path.dirname(__file__)) |
|
sys.path.append(os.path.join(pwd, '../../')) |
|
|
|
from datasets import load_dataset |
|
from tqdm import tqdm |
|
|
|
from project_settings import project_path |
|
|
|
|
|
def get_args(): |
|
parser = argparse.ArgumentParser() |
|
|
|
parser.add_argument("--dataset_path", default="tiansz/ChineseSTS", type=str) |
|
parser.add_argument( |
|
"--dataset_cache_dir", |
|
default=(project_path / "hub_datasets").as_posix(), |
|
type=str |
|
) |
|
parser.add_argument( |
|
"--output_file", |
|
default=(project_path / "data/chinese_sts.jsonl"), |
|
type=str |
|
) |
|
|
|
args = parser.parse_args() |
|
return args |
|
|
|
|
|
def main(): |
|
args = get_args() |
|
|
|
dataset_dict = load_dataset( |
|
path=args.dataset_path, |
|
cache_dir=args.dataset_cache_dir, |
|
) |
|
print(dataset_dict) |
|
with open(args.output_file, "w", encoding="utf-8") as f: |
|
for sample in tqdm(dataset_dict["train"]): |
|
text = sample["text"] |
|
|
|
if text == "句子1\t句子2\t相似度": |
|
continue |
|
|
|
splits = text.split("\t") |
|
if len(splits) != 3: |
|
raise AssertionError |
|
|
|
sentence1 = splits[0] |
|
sentence2 = splits[1] |
|
label = splits[2] |
|
label = str(int(float(label))) |
|
|
|
if sentence1 == sentence2: |
|
continue |
|
|
|
if label not in ("0", "1", None): |
|
raise AssertionError |
|
|
|
row = { |
|
"sentence1": sentence1, |
|
"sentence2": sentence2, |
|
"label": label, |
|
"category": None, |
|
"data_source": "ChineseSTS", |
|
"split": "train" |
|
} |
|
|
|
row = json.dumps(row, ensure_ascii=False) |
|
f.write("{}\n".format(row)) |
|
|
|
return |
|
|
|
|
|
if __name__ == '__main__': |
|
main() |
|
|