--- language: - en license: cc-by-sa-4.0 dataset_info: - config_name: max_len-1024 features: - name: text dtype: string splits: - name: train num_bytes: 270688438 num_examples: 449722 download_size: 148512885 dataset_size: 270688438 - config_name: max_len-448 features: - name: text dtype: string splits: - name: train num_bytes: 272285366 num_examples: 930858 download_size: 151141635 dataset_size: 272285366 configs: - config_name: max_len-1024 data_files: - split: train path: max_len-1024/train-* - config_name: max_len-448 data_files: - split: train path: max_len-448/train-* --- # Wikipedia simple splitted Wikipedia simple data splitted using Langchain's RecursiveCharacterTextSplitter ## Usage - This dataset is meant to be an ultra high quality dataset. - This can be used for Annealing LLMs. ## Why its different - This dataset is split with a max length of 448 (128*3.5) and 1024 (256*4) characters - And rather than splitting by length, it is split using RecursiveCharacterTextSplitter. So, the chunks don't end randomly. - Can use very large batch sizes. ## License [CC-BY-SA](https://creativecommons.org/licenses/by-sa/4.0/deed.en)