from datasets import Dataset, load_dataset from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained('models/RedPajama-INCITE-Instruct-7B') max_seq = 2048 def make_prompt(code): return f'Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{code}\n\n### Response:\n' def is_not_too_long(data): encoded = tokenizer.encode(make_prompt(data['content'])) return len(encoded) < max_seq def deduplicate_dicts(dicts): seen = {} result = [] for d in dicts: content = d.get('content') if content not in seen: seen[content] = True result.append(d) return result dataset = load_dataset('json', data_files='ts_parser/ts-chunks.jsonl') data_short = dataset.filter(is_not_too_long) dedup = deduplicate_dicts(data_short['train']) data_short_dedup = Dataset.from_list(dedup) print(data_short_dedup) data_short_dedup.to_json('typescript-chunks.json')