File size: 3,169 Bytes
bfae27f
 
 
 
 
 
 
 
 
 
 
6d97d06
bfae27f
 
 
 
 
 
 
 
 
 
 
 
6d97d06
bfae27f
 
 
 
 
 
 
 
 
 
 
6d97d06
 
bfae27f
 
 
 
 
6d97d06
 
bfae27f
 
6d97d06
bfae27f
 
6d97d06
 
bfae27f
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
import os
import sys
import json
import tqdm
import argparse
import transformers

def parse_args():
    parser = argparse.ArgumentParser(description="Merge the boards into a single train set.")
    parser.add_argument("-t", "--tokenizer-name-or-path", default="meta-llama/Llama-3.1-8B", help="The name or path for the tokenizer")
    parser.add_argument("-l", "--limit", type=int, default=4096, help="Length limit in tokens for each post")
    parser.add_argument("-m", "--min", type=int, default=5, help="Minimum amount of message in each post")
    parser.add_argument("-i", "--id", default="<|start_header_id|>", help="Prefix token for message IDs")
    parser.add_argument("-c", "--content", default="<|end_header_id|>", help="Prefix token for message contents")
    return parser.parse_args()

def main():
    args = parse_args()
    tokenizer = transformers.AutoTokenizer.from_pretrained(args.tokenizer_name_or_path, legacy=True)
    if args.id not in tokenizer.vocab:
        print(f"The message ID prefix token \"{args.id}\" is not a token in \"{args.tokenizer_name_or_path}\", it will work but it's better to be a token in the tokenizer.")
    if args.content not in tokenizer.vocab:
        print(f"The message content prefix token \"{args.content}\" is not a token in \"{args.tokenizer_name_or_path}\", it will work but it's better to be a token in the tokenizer.")
    boards_dir = "boards"
    total_token_count = 0
    with open("merged_strings_train.jsonl", "w", encoding="utf8") as output:
        for board_path in tqdm.tqdm(os.listdir(boards_dir), desc="Boards"):
            board_name, ext = os.path.splitext(board_path)
            if ext != ".json":
                continue
            board_path = os.path.join(boards_dir, board_path)
            if not os.path.isfile(board_path):
                continue
            with open(board_path, "r", encoding="utf8") as f:
                board = json.load(f)
            for post in tqdm.tqdm(board, desc="Posts"):
                if len(post) < args.min:
                    continue
                post_content = board_name
                post_token_count = len(tokenizer.encode(post_content, add_special_tokens=False)) + 2 # Add 2 for the start of string and end of string tokens.
                for message in post:
                    formatted = f"{args.id}{message["id"]}{args.content}{message["content"]}"
                    formatted_token_count = len(tokenizer.encode(formatted, add_special_tokens=False))
                    added_token_count = post_token_count + formatted_token_count
                    if added_token_count > args.limit:
                        break
                    post_content += formatted
                    post_token_count = added_token_count
                json.dump({"input": "", "output": post_content}, output, ensure_ascii=False)
                output.write("\n")
                total_token_count += post_token_count
    print("Merge finished, total token count:", total_token_count)

if __name__ == "__main__":
    try:
        main()
    except KeyboardInterrupt:
        print("\nScript interrupted by user, exiting...")
        sys.exit(1)