diff --git a/.gitattributes b/.gitattributes index af56652d4c915a7bf5498f85361d7a88c0f0736c..980b52be1dbcea50056adeee231619a5d0a31596 100644 --- a/.gitattributes +++ b/.gitattributes @@ -23250,3 +23250,84 @@ train/tulu_flan/tulu_flan_0016-tokenized-chunked-1024-512-128-backfill-nodups/sh train/tulu_flan/tulu_flan_0016-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text train/tulu_flan/tulu_flan_0022-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text train/tulu_flan/tulu_flan_0016-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00062.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00063.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00058.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00072.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00051.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00069.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00054.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00057.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00065.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00071.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00064.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00068.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00056.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00042.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00070.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text +train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds filter=lfs diff=lfs merge=lfs -text diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..02cb8482a97adfb1f3711d05b6789d1ab46f8bbe --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108725, "hashes": {}}, "samples": 44128, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47872276, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9455084, "hashes": {}}, "samples": 6299, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6740146, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..320060a5b255c7390edfa6aa89dd6f3501d7b7ba --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_14215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36721199, + "num_truncated_tokens": 36693634 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..92ea564077bd6b3e71e06b262423a16a84845db3 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108713, "hashes": {}}, "samples": 43172, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48003662, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16484068, "hashes": {}}, "samples": 10641, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11747059, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..191d9c39d98054771d5eab7e2ff34525b0defb3d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_15967-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 40130649, + "num_truncated_tokens": 40097178 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..313823ba9812c4ac1398bf7a155ae9572f5d5def --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107649, "hashes": {}}, "samples": 42868, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47258633, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19881258, "hashes": {}}, "samples": 12636, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14073457, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..6cc4a60828aaacbf0124e2fbe98eac9745a369a4 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_16294-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 41776342, + "num_truncated_tokens": 41740254 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..c855e5a1ee118c657488d78a84729d83bc02429e --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108731, "hashes": {}}, "samples": 43437, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47698834, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16223002, "hashes": {}}, "samples": 10537, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11536023, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..03269da9944694166fb06524ad1b8014edbea6f5 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20157-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39995294, + "num_truncated_tokens": 39962065 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20479-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20479-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..8dcf0592a1d0d146296d07641523d29c504b7c22 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20479-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108377, "hashes": {}}, "samples": 42362, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47693193, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22123595, "hashes": {}}, "samples": 14151, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15652228, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..64910b46ac37c4a0e34681a4ca5a644b33c3ec8c --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108584, "hashes": {}}, "samples": 44670, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47893595, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8789604, "hashes": {}}, "samples": 5838, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6295797, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..4678ab96d8cc3f0b10e76506c91997b858b2f5ce --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_2232-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36386128, + "num_truncated_tokens": 36358275 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..8e1d166e74b3747053c10cb8f78ae2869370dc60 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108535, "hashes": {}}, "samples": 44332, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48013221, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8718413, "hashes": {}}, "samples": 5997, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6254804, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..0bdc063ceebe5df3c0016b708284b11e016bf69d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30083-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36355863, + "num_truncated_tokens": 36327748 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..769225d4e3355e1a674a91613756d830f1af1060 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107550, "hashes": {}}, "samples": 44401, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47735382, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9056848, "hashes": {}}, "samples": 5963, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6451989, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..558448ff0c2c6892cb9bf3303f8322c65c5293f3 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_31005-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36523487, + "num_truncated_tokens": 36496246 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..aef6f3c6c520c3f9f0b9a1e12329f9d57441df68 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107346, "hashes": {}}, "samples": 43081, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47725629, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18056520, "hashes": {}}, "samples": 11619, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12845249, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..665ef5d0f34a41ad7b1a2797080c9835907118df --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32107-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 40888826, + "num_truncated_tokens": 40854200 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..479f61c2594cae27fa850fe0f9281950af59efde --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108707, "hashes": {}}, "samples": 43553, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47842310, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15887096, "hashes": {}}, "samples": 10150, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11341346, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..9ec74342ff666238e908480ecdc437539cbb735b --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34368-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39835666, + "num_truncated_tokens": 39802447 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..73ea39b92452108e1844cfb1aba24def7c665990 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107712, "hashes": {}}, "samples": 43208, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47918427, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16948902, "hashes": {}}, "samples": 10963, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12121309, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..02acc3a0c738a7a763b28f90521f799773332e2d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35642-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 40351545, + "num_truncated_tokens": 40317300 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..bd0be973fc775c70ebb705f05a4b62c731893bc9 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108309, "hashes": {}}, "samples": 43546, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47458723, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15316201, "hashes": {}}, "samples": 10011, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10872005, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..4027410397bebaa31c7375e35f3c6cd6bd5c7841 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_35859-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39554651, + "num_truncated_tokens": 39522362 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..0b77cfe06ee371213b602ec29c0b2200d49f97a5 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107456, "hashes": {}}, "samples": 44076, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47592362, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11844849, "hashes": {}}, "samples": 7479, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8370864, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..373a7628836f9b47cb1c20810f06300f4db7120d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_38999-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 37880444, + "num_truncated_tokens": 37851015 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..51c08d667d71acbf26e1906c5d74694dee347f8d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106915, "hashes": {}}, "samples": 42537, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47657887, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19973766, "hashes": {}}, "samples": 12991, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14215136, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..ec7ce4b63fdc025d4032a3f32c77d449e877372b --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39030-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 41821525, + "num_truncated_tokens": 41785466 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..63a35134ce4bafb5808b3d126d09b6b9a1405bd9 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107800, "hashes": {}}, "samples": 42476, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47576226, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23529924, "hashes": {}}, "samples": 14792, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16702384, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..f523dfb79cdbf6c5c508ef19a26c4b24e65df1bd --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39208-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 43546050, + "num_truncated_tokens": 43507882 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..9521347db92d6965447b2978c9ac31f94b79564a --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108136, "hashes": {}}, "samples": 43580, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47714872, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15213151, "hashes": {}}, "samples": 9864, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10801620, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..72524e377760f4d6099f086a8d29e943430c1572 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_39477-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39506610, + "num_truncated_tokens": 39473768 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..c65bcc50cf8315b95cb730b945acc6200738b03d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108343, "hashes": {}}, "samples": 43391, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47650209, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15674157, "hashes": {}}, "samples": 10212, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11199852, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..d3885f5ed639ce7242b582b0088d5d1916bbec9a --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_43764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39732190, + "num_truncated_tokens": 39698964 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..1d0c35d1d1dc0a193e3fd34728bef1129e90729d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107293, "hashes": {}}, "samples": 44643, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47884153, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 7912209, "hashes": {}}, "samples": 5276, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5641556, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..c44158c0f155b79ee2744be7d2752a88be0f6820 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_44556-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 35964936, + "num_truncated_tokens": 35938367 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..420e77bcd799b2706bf3de5fbac6cc7758824e0e --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107943, "hashes": {}}, "samples": 43250, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47487803, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18000696, "hashes": {}}, "samples": 11454, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12743221, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..e572062c560774ed4b7f58b1da375a310dae5848 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_48755-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 40861066, + "num_truncated_tokens": 40827031 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..32f6eaf4592af29c1fcbe2a783bf0f9b62cf0654 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107807, "hashes": {}}, "samples": 43294, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47566095, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18157875, "hashes": {}}, "samples": 11774, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12891107, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..b9d0f3c56cb39abc22d7704aac254830e668b691 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_52517-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 40928599, + "num_truncated_tokens": 40893592 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..1c9a37a4778969a00f6e1ceff608157df1edd5d7 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107577, "hashes": {}}, "samples": 44093, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47486788, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13408828, "hashes": {}}, "samples": 8587, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9472090, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..779dc59d8494abcfecf9bc74e096af48d0721eb2 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_54460-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 38627885, + "num_truncated_tokens": 38596538 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..c101ae658b46dab9053f31c645edacb8878fe1e1 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108283, "hashes": {}}, "samples": 44628, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47895851, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8664593, "hashes": {}}, "samples": 5821, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6209343, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..f8ba5a64b49eb014b22e6bd17119fafbd3dec83b --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_61215-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36325193, + "num_truncated_tokens": 36296925 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..2bbc7929133e382744abc09fb40323cea9627d12 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108650, "hashes": {}}, "samples": 43807, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47678796, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14991766, "hashes": {}}, "samples": 9352, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10675387, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..b2e6ae7b4527760849065ab70a3ad596fbc1936f --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_64217-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 39404797, + "num_truncated_tokens": 39373353 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..d3434dd2d5fcd9052a624236c6b7f4fcc50afbfe --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106945, "hashes": {}}, "samples": 42986, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47623149, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19085838, "hashes": {}}, "samples": 12173, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13487985, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..9297ebc61fdc89a2e777947ab689841f7b98fd71 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_66216-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 41389045, + "num_truncated_tokens": 41354161 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..de5da3d7a10dfd27df8f0ae2307252b2d95fcdfc --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108199, "hashes": {}}, "samples": 44130, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47823216, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10589033, "hashes": {}}, "samples": 7058, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7583402, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..5fdb9fcc78bf8568875fd263fff9d35244b9fefb --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_69362-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 37264443, + "num_truncated_tokens": 37235966 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..e5be5c2c014f2722ed93b317345a369b4cae6858 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106881, "hashes": {}}, "samples": 43876, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47771847, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12385820, "hashes": {}}, "samples": 8228, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8863577, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..951e13ed104e206eee4608c79fd8012f3b82e823 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_70028-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 38133788, + "num_truncated_tokens": 38104454 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..daf11fdd7ce8a8500c6be3bbd2326a0531e3e379 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108039, "hashes": {}}, "samples": 42284, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47593943, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22245244, "hashes": {}}, "samples": 14330, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15923634, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..d03438d93a6e08e19f9b74900a2e960f1c36ac8d --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_73087-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 42924078, + "num_truncated_tokens": 42886773 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..eb5b342d89c863bf52288925f9cb363b8a96b7c8 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108660, "hashes": {}}, "samples": 43872, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47732858, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11700270, "hashes": {}}, "samples": 7918, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8386558, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..bc5e79045fa8ecd00f017b4c49befa671c26c3cd --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_74796-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 37801697, + "num_truncated_tokens": 37771931 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..894f8b36d17a8dadcd3fd9a87a1fd5cbe162d7d9 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107664, "hashes": {}}, "samples": 42773, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47535166, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21289372, "hashes": {}}, "samples": 13318, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15142951, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..76545eabdfcfb0a193121fd84649dd07e2c875dc --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_77764-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 42462157, + "num_truncated_tokens": 42425785 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..38ee55f3a79f9b4d4824e62b52556f8480de6247 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108340, "hashes": {}}, "samples": 43766, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47565532, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12464014, "hashes": {}}, "samples": 8189, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8892218, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..5e2c686d07873faf6ec891160994a448a8ea01c7 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_79661-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 38178140, + "num_truncated_tokens": 38147980 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..2c8c06b0a80f3cfdaa9a3f081f3670d7bfa63d90 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107353, "hashes": {}}, "samples": 42850, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47495627, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20041455, "hashes": {}}, "samples": 12737, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14297190, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..8ebf58ce3821accb1214848bf4d87f0412633d56 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84354-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 41853690, + "num_truncated_tokens": 41817960 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..71c6881e0aeb08e36af617f786d57457808b4ace --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108133, "hashes": {}}, "samples": 42525, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47627700, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22832435, "hashes": {}}, "samples": 14522, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16187019, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..0a263b98666b71047724ccfd1823186710ec8ae0 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_84770-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 43204494, + "num_truncated_tokens": 43165966 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..731614d80e32607c8088898a2ec4edb107eabf39 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108784, "hashes": {}}, "samples": 44583, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47850093, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9357258, "hashes": {}}, "samples": 6168, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6661383, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..99f232fc3e94bf2655f94b45925ed2df058a1676 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_87845-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 36662406, + "num_truncated_tokens": 36635358 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..56808fc4e01330ec8528e43a3bbca405cf4789f4 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107220, "hashes": {}}, "samples": 44913, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47920578, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 6823533, "hashes": {}}, "samples": 4452, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4862978, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..d9fb387b6c0a81979d7784001241d71092d12b4b --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_88182-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 35437690, + "num_truncated_tokens": 35412191 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..932b89e972302b9f1961c86f65ac903bd5ee6dc8 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108733, "hashes": {}}, "samples": 43541, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47970925, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13386148, "hashes": {}}, "samples": 8895, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9618804, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..e34e15d6deee71e40bd188a4035c06ae3a838d7a --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9059-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 38624419, + "num_truncated_tokens": 38593293 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..c963873da43ecd7a9055def453718ee99bda4cd7 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108712, "hashes": {}}, "samples": 43982, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47691292, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11951110, "hashes": {}}, "samples": 7723, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8494211, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..a39b293dc34d937413db55143a00c0f2c9d790f3 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_93359-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 37929640, + "num_truncated_tokens": 37900234 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..ff4cd29f2e6f755e26ceff294ef9bf1a2928c736 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107436, "hashes": {}}, "samples": 43922, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47405977, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11992919, "hashes": {}}, "samples": 7751, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8472678, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..392e30e0c2600211e517ad3b28c6c63e75ad3d46 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_9466-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 37950799, + "num_truncated_tokens": 37922043 +} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/index.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/index.json new file mode 100644 index 0000000000000000000000000000000000000000..376837596ed225f7a5a010091d23aa41dc2dfdba --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/index.json @@ -0,0 +1 @@ +{"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108797, "hashes": {}}, "samples": 42348, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47316454, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23569345, "hashes": {}}, "samples": 14917, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16721799, "hashes": {}}}], "version": 2} \ No newline at end of file diff --git a/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json new file mode 100644 index 0000000000000000000000000000000000000000..76d7450391fe15e8b25188c863253998288be079 --- /dev/null +++ b/train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_96947-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json @@ -0,0 +1,4 @@ +{ + "num_tokens": 43566460, + "num_truncated_tokens": 43527775 +} \ No newline at end of file diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds new file mode 100644 index 0000000000000000000000000000000000000000..05c9dbb236212e0eea311c3594050e7ae96bc5bf --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b707fb9be41536b92f962979a95756b0ac98994a66d95155fa107a3ace47362 +size 67108642 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds new file mode 100644 index 0000000000000000000000000000000000000000..d2804536f93afc60508ed73c05be2450fce7e00e --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20a7d3f9bf38103bf8d6e3ab95c6653230069b8908e17d7bf8a275bc94e97f6c +size 67108364 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds new file mode 100644 index 0000000000000000000000000000000000000000..3941052bac144f3bbd6e13f5b038c7976820fb4d --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49579238c0c79b808a9b818e3faed245b07138378d192532b0e8475a477e9add +size 67108179 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds new file mode 100644 index 0000000000000000000000000000000000000000..a609493f98e4900fa467faf245f48390e5a9f757 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d00f88f852f3070f0edd0480dfc8ae63c61f65bb8d417309bd93e0d6e134a6d +size 67108191 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds new file mode 100644 index 0000000000000000000000000000000000000000..bd7a35207012e0696a3d4b10bfcdb7ff5d9c8e7d --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:379b8f0b2cf8910fc6d4ba88b0ac70d5380723faeebd4087ef91eaf625225fb0 +size 67108792 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds new file mode 100644 index 0000000000000000000000000000000000000000..acacb33b174c7256e0f724ef74c544e3d454b01b --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:650c534cd1ad09ae99022740f44fc32bc17b9ccc43edecdf8030f5ceb9f9a3e6 +size 67107690 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds new file mode 100644 index 0000000000000000000000000000000000000000..4e1356e4431d264950cbb8203e7467cd9ac5df57 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:def9a12bb3aa27d9efe149d78b7a3c1627885f1d19bcf1d612131d18ac30201a +size 67107041 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds new file mode 100644 index 0000000000000000000000000000000000000000..5607b2898ccc494ee878567255cd28092bb0e331 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcc6b6dd6a97eb04c55cb179d676135bf5cd1d640f2afe84f756af822d8feb24 +size 67108858 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds new file mode 100644 index 0000000000000000000000000000000000000000..e9a6eb09f8860b7c6661d2acd0c2c4df13dfe9b6 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:657cd979b3179987f98c831f7b2c8e481c002dc9850a8a0d47d61db047dae6ac +size 67108287 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds new file mode 100644 index 0000000000000000000000000000000000000000..a94f1f08041c20a860c2e985290752c76358fa81 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b2db0b2ade1f698907a785a27a1e9ea6d168408c14dfa141c25f6c3f57b6a01 +size 67107742 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds new file mode 100644 index 0000000000000000000000000000000000000000..b2d2536d1030bfa8323beeeac2ac3c7e04fc3b58 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6345d6e9112b4dadd2093948913533c9bb6224b78e70a4b382093b75c0ac00bc +size 67107254 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds new file mode 100644 index 0000000000000000000000000000000000000000..4ee33285d63ad6b80b04f529e60ae3ed24371ba6 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f55fb3008d16e75ec4dd8196c51eb99b6bf2098c9e0acc4e39cf07513c3e4e8 +size 67108307 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds new file mode 100644 index 0000000000000000000000000000000000000000..d7bdeaa236ac0368f9c36a928abe8a33f980d33f --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1910fe9f72716f801b02e52cefc7e387a0e552bf4a15ae7662c4d3567875e6ca +size 67108374 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds new file mode 100644 index 0000000000000000000000000000000000000000..6ea3e8e87212636323c3026d3157b4154396ec83 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc98755699ec4520d18d7a9cbe26057e7f26b3cd8e7e53cac9287e41929ec454 +size 67107526 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds new file mode 100644 index 0000000000000000000000000000000000000000..5163186f238722c2b229a5f79a68df103f7e813a --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb56d14a1a8abb840c27b4fbebf53e14c8172637ec5657d4cbfc4a3c00fabb94 +size 67107828 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds new file mode 100644 index 0000000000000000000000000000000000000000..5fef81f790f91ba8c62afaba7ebf445b2f203465 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e121d8b85e93f10735da5aea20ec06781e18214ac7fce1212fcee244f5a4ea60 +size 67108024 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds new file mode 100644 index 0000000000000000000000000000000000000000..48c9c42100a1d7b6603479b92fffe72afd6e64b5 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4b1a1736e84c9202a94b44f2a3bd1800c865213c7b53b2bf0f8ce77f27574be +size 67108711 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds new file mode 100644 index 0000000000000000000000000000000000000000..6e2080b3fec2c7a775ecd54451edd48c1b7aa27a --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7dcba73a766a36b3ae35dacf0820fdb11f4b3660053203499cffe1251286622 +size 67108837 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds new file mode 100644 index 0000000000000000000000000000000000000000..b49743df66c76e273514d0615e2d13cad3c0558b --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b6f706bf025f1571d2e3ce82e8da38f9349a9f19a79fb53723079f372ba6760 +size 67107681 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds new file mode 100644 index 0000000000000000000000000000000000000000..778c989a1738f18b6d31b6952e8254f33c50c6c0 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f15e1be4592708c04d79256e56b414611b42a708fabe317356a8f9ce3182b05 +size 67108780 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds new file mode 100644 index 0000000000000000000000000000000000000000..d69f9d5ac18862daa5fccc15474df2439255cf52 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ed239cab44b067155ed03da87847b051cdbea85a7b42f673f804a13fbfa3722 +size 67107170 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds new file mode 100644 index 0000000000000000000000000000000000000000..d7f9a5d2f49e1679cc60fb63490680136df3e6b9 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82d1fd1b66fe2e2ab8789fd46efabc05fdee544a95f84f8b4589670b6cc10819 +size 67107784 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds new file mode 100644 index 0000000000000000000000000000000000000000..b701afaa6bf51cb5474a89caad092c53f74ac944 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e2e09aeccade41b28513fd76960970742fc7ab6d0dc00e394d4c4e4581ef0ec +size 67108532 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds new file mode 100644 index 0000000000000000000000000000000000000000..ccf008a9ab8f545bf110bd65d42c0508e6c60bc0 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:deb5e93197c6d8b3bfc3213fbc96abd58a741830bc480d8d47a4aa218f73a95c +size 67108173 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds new file mode 100644 index 0000000000000000000000000000000000000000..bfabb7fffdb623b0851f31b3678773dd26021e22 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5cf97441d735ec21d66bb70a1ec94d31f85ecb0049fb5a1004ed12b23205cfd +size 67106966 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00042.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00042.mds new file mode 100644 index 0000000000000000000000000000000000000000..81540d5d420cc8dafe7a65753a4f6d519fc5f162 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00042.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:621253817caa961ba62d17276e47daf9b462381a9dbec28604e36f10011bb091 +size 67107282 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds new file mode 100644 index 0000000000000000000000000000000000000000..cf1e1013b007a0359bc2ecc93e4c0fffc2cfe7fc --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00043.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd3fb38b59800b6cefd020813273f9773c4d10b50bd529e8389de132e9de84c4 +size 67108811 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds new file mode 100644 index 0000000000000000000000000000000000000000..3d772934ead03173002eda435dffc67308060db3 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00044.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d59322f68d84c71f83fce2d89b8f054f7d4285bb3ebe048b9d0ee7d5625457f +size 67107479 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00051.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00051.mds new file mode 100644 index 0000000000000000000000000000000000000000..b1f80f982e6afbf66d4fa902f9879bbb0bdd3179 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00051.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:975d4828a94bc728ac93f068a52f595a3490b703e06a408a00d52e4e9654d184 +size 67107979 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00054.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00054.mds new file mode 100644 index 0000000000000000000000000000000000000000..409ecdb0bd7f9485d498c1525157a84addfeb1da --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00054.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:657ebbfc9fa5eef23c87ba8a72978cdd1bc689acbeab33266081a24e02984784 +size 67107432 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00056.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00056.mds new file mode 100644 index 0000000000000000000000000000000000000000..097a5cae5d2752b8d4637e1d47dd628dccce72f4 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00056.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3c9c10985e047c16665eb2adcbd4a9ccbfcb8d53edd7ff1f7411113239f2d3d +size 67108146 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00057.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00057.mds new file mode 100644 index 0000000000000000000000000000000000000000..42ec590d1e4d01031d6d8b71b4614fa84f4d2342 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00057.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbec941086909c80d145e8ee0612f84312c6b60bd91be2692b16bbea6db426f6 +size 67107840 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00058.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00058.mds new file mode 100644 index 0000000000000000000000000000000000000000..4dba383eada187436c80c81fe037e0722df0122b --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00058.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:284f695bd47f7fd02eb187b3d4400656773a6eb5b61299c742f1c04dd58998c1 +size 67108806 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00062.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00062.mds new file mode 100644 index 0000000000000000000000000000000000000000..cd8163a91d65b75c5cd5d9988130ff76023028d9 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00062.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71170309de2dda91684e4ead67a5fbcdde1e9a23a86adef1edc440127e1e6ad5 +size 67108600 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00063.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00063.mds new file mode 100644 index 0000000000000000000000000000000000000000..ae29d070edf108bc8047dc6cd3f4d93dbe65bc00 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00063.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ee6d9d2cd5490d4a71d4b0c1ed498beff2d4eb1cd9b02ede12e9465bcb0f56a +size 67108776 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00064.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00064.mds new file mode 100644 index 0000000000000000000000000000000000000000..a1196ae49c15fadfa014dda07cc05a7afdd48b61 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00064.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:550454443d592c2e1dc88cd163d135066b6cbef1b1c4c56b314a388431cd4612 +size 67107577 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00065.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00065.mds new file mode 100644 index 0000000000000000000000000000000000000000..9bd3b2354055df48aeca7cbd138d3d35f548ba7c --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00065.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0213814167aa2ef18b7fe3b8e3fb6e6d7d6b75403b7c0da84151315f8ca4bd5 +size 67106947 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00068.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00068.mds new file mode 100644 index 0000000000000000000000000000000000000000..ee29e9262a74ef52f3b42e49062b254e2b59fb40 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00068.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08d0d3114f59637d1c58e229e6bb7fb99ea34df369de7185acaa4c34b3107939 +size 67108817 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00069.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00069.mds new file mode 100644 index 0000000000000000000000000000000000000000..828c181b724f17d9a74640c9faab006eb01b8c82 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00069.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7be66af0b8496a68fd1ed40396ebf98119c043a1db0c356702db7beff200819 +size 67108468 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00070.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00070.mds new file mode 100644 index 0000000000000000000000000000000000000000..2caa3b8610394a7410351d431a01ec08a967e1c8 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00070.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8243bff196d4cccc71d5b1828fce8e3527295bd153c1d993b2442522702108e +size 67107232 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00071.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00071.mds new file mode 100644 index 0000000000000000000000000000000000000000..12b1876d7cf0986b63efa0a8776632dde49c60ba --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00071.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1a34dc4df392b1688066c52ffca9ddcf382d438e90a688f5adc155af9d2b8ed +size 67107460 diff --git a/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00072.mds b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00072.mds new file mode 100644 index 0000000000000000000000000000000000000000..86fdd274d26b8bbf6d9668e0f1dba26db4c4f370 --- /dev/null +++ b/train/wikipedia_dup/wiki_0000-tokenized-chunked-1024-512-128-backfill-nodups/shard.00072.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bdda5e78b36dba6a93aee33dfefa2376a31d0e98336830b4de5ba611824b6bf +size 67108592 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds new file mode 100644 index 0000000000000000000000000000000000000000..17a7acb15b0e47c16b25efae880ee6b054eb0114 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ac6cd5df329ecffa30194acfa340355a275aad89202cd8c5de331f26e4d9c7a +size 67108552 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds new file mode 100644 index 0000000000000000000000000000000000000000..b6dabb8f8d33eec101adae3ebb2c17f17dbdfaa9 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59dc101432da63dbb0a790dffbe7d98945f274290356954c2b03382f8292f18e +size 67107701 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds new file mode 100644 index 0000000000000000000000000000000000000000..35b3f1b5cbf34119a71af99112012af4ff89a5cb --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbb9e6fbd94e0a5db7c7c63fd9312430c254c44bb04728ff407188f66d64c117 +size 67107657 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds new file mode 100644 index 0000000000000000000000000000000000000000..f8acde53669cf3a5c9ec0e9a734dad9c356d34f6 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5983ba49684be5a4c5bac3be5a2c0e3d294e56539d7dfada15b1676249c635af +size 67108291 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds new file mode 100644 index 0000000000000000000000000000000000000000..d5665cb156c25312fc1ccb0196b3aef4065c4fd7 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01760a4e412bf0554a103547124c7a6e51a6fea65b430e59f4d09bce27db9992 +size 67108227 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds new file mode 100644 index 0000000000000000000000000000000000000000..2c72cebaf0ec2d08e452976e6f6187795eb64e33 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fedad6e44e064c63f14bb00c506476372833ec27d53517a61ccfa777baccecd +size 67107320 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds new file mode 100644 index 0000000000000000000000000000000000000000..1a148760cdd710c217876107bcc7d26c97ea01d1 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a7eece28f4988bba6c23282c1a5ce093f876a625c04c26940e5e5b2abd1dee4 +size 67108176 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds new file mode 100644 index 0000000000000000000000000000000000000000..12bedf95c6513f196fdc144ea80ca710f85ea31e --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:087e07cc5f6b96733b5febd3002756798b51830c589a78de25ef17beaf8c7d7c +size 67108384 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds new file mode 100644 index 0000000000000000000000000000000000000000..bb5552dfb5ebf793e558422598c257f314a1085f --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb9fcd8f8abf187d711ef911ac7d8168715f0850151564e7995f050bf9761d8f +size 67108271 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds new file mode 100644 index 0000000000000000000000000000000000000000..d2567f3934a8b5a5a6bcd08e05b5e1f945fdc188 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ac335b61139c3262cf0ca4170b79e56096096c3844286324780295a4017b6b9 +size 67108315 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds new file mode 100644 index 0000000000000000000000000000000000000000..9a25defc21db25bd2f9c1f4df7f5459db2089d46 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d47bf09397ceeae1e052cd8db7cc4c4a35cec0a3de6df81012f4fb5f3168233 +size 67108508 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds new file mode 100644 index 0000000000000000000000000000000000000000..e9e6bf98fea958d6caa0881c33452ca2e33f9f10 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:793d512645e42be670b9899e3cc850ae7b931921ffcaf711d47f722c6788167f +size 67108462 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds new file mode 100644 index 0000000000000000000000000000000000000000..8bf61b977d028d4ee985048621bfd552a73cf29a --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61c05cc519a3ed80c5b15985de1e0ea059050a535b4a4baf4303216874f6920b +size 67108644 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds new file mode 100644 index 0000000000000000000000000000000000000000..1df7b8e5f8af8e4c665e3fb8c5d4a189640a7c8b --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcdc495617137fb152218dd4872b644977fc750ff49dc8d7e6629be7bfeb7a7f +size 67108407 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds new file mode 100644 index 0000000000000000000000000000000000000000..9301e717513b7a4393030a702bbad28809bd522e --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34ef97cdd67cf45cfc91a15b96ddaa4edc46ed3a08a046720167397f0a977923 +size 67107624 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds new file mode 100644 index 0000000000000000000000000000000000000000..37b812593cfcce17e139dc4c40448ede4d9c8963 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b6bfd1a30f052e39ad793f43037097626fabc47af1436a28f6c0a4cd3b21565 +size 67108038 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds new file mode 100644 index 0000000000000000000000000000000000000000..e46d2a4b4a48b0d332fa9938ff0e57a3e05d00c9 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee32d6f5881e4219e6806a1aae00e44e8d02fc166aa380d1df5b4305454ee914 +size 67108565 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds new file mode 100644 index 0000000000000000000000000000000000000000..d77a3eb7a843a033b999c100a254b21cca982af1 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05a8ea0eede2a810401bf8ac2a89dc6f8384b3022968bd0b619b180cf20264d4 +size 67108499 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds new file mode 100644 index 0000000000000000000000000000000000000000..bc7a909371773f1385e0d6d7a1334234e91952e0 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b65af5fa08563d5464db3ff0a52b5750299cff793b39b80cf0e19c384610dda +size 67107744 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds new file mode 100644 index 0000000000000000000000000000000000000000..f957ceb99cea6feace3a3ae4588c2ff2cf7f22ed --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b769b7365a1e1783f8b9821a44ce6366746c2c690cde970c069ed436323154b +size 67108431 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds new file mode 100644 index 0000000000000000000000000000000000000000..6d9f029bd09c0bd890058a3693de04abace3d9ba --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efbcaf63d7674ad916cb676b43d67e16ac527d6ed2da40f807e7a63cc4e2978f +size 67107969 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds new file mode 100644 index 0000000000000000000000000000000000000000..fe8d338fa856bf4d35ecd3eae5c9ffc70928dcd8 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fce38a489c04acc013df17d5f8dda59232be0ae479e1cb0bc486875d99b8bedc +size 67107208 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds new file mode 100644 index 0000000000000000000000000000000000000000..26750ebcd96603e6b15308296525b59f1f646d73 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac13f04123a39159ad7e7ffaffc6036b1583bdb4bb615490765118361638e618 +size 67108787 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds new file mode 100644 index 0000000000000000000000000000000000000000..5d30fd9864f74464b22fcf087d88770a706ab7c4 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25ea2edac217da7df60abf38663ac391c9c34a56588a1a98a4ba9cb2d09c7bb0 +size 67107943 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds new file mode 100644 index 0000000000000000000000000000000000000000..a2c12958eef76f6cd132dc397cff63e92240b71f --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00024.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3aed1a3de7d887ba28f81f69ac27ea870208e1a9f08ebcb9523e2921f6654969 +size 67107016 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds new file mode 100644 index 0000000000000000000000000000000000000000..1395f440f77bc179942476ac8fc17143fd099231 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec90b86a580ef408aed07f59229aa6477aef8f5ae9603de8f4346f98476d7789 +size 67107533 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds new file mode 100644 index 0000000000000000000000000000000000000000..4021b0431a6badf275dde57db43ee931f557e65b --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00026.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cb3bd7803c6479e939ddf6ebffaa7b0af19fdde4174d47b81ed8801fdacea7e +size 67107280 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds new file mode 100644 index 0000000000000000000000000000000000000000..a22b1aeac486a6be193fcf495bac6b10f84bce9b --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00027.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fbcefc1423edc4fff1f967e83bf5fd39436b5e05f7728853bf12c9e433e285e +size 67107592 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds new file mode 100644 index 0000000000000000000000000000000000000000..ca340bbfe3994e71c1a81d3e7bc7c145815bf89d --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a401e9edd3936f67568c280787930ec5b4d45cc105c13b0f480e555364e6cb6 +size 67107501 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds new file mode 100644 index 0000000000000000000000000000000000000000..815313bcd5045978e6f75209ff255c0c7bc318e8 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3650ed4688b5b6cd6e0d2a8c43ce834467a553ecba187044f7bf2ef8197983c1 +size 67108235 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds new file mode 100644 index 0000000000000000000000000000000000000000..b2f2fa8de8c1ae7ab52c26b5dcb0b0d1daae93a3 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00030.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83979802a368efe4e2a56b2a202d8f8f027cfbbda238778616bd94d5e7e24870 +size 67108449 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds new file mode 100644 index 0000000000000000000000000000000000000000..a2b58bb187708bb0966b9d5ac5d7b8dece78751c --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6f8df8e94865e08acf4b23200e652dc46375a27fde2235b03078dfed9ad1295 +size 67108575 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds new file mode 100644 index 0000000000000000000000000000000000000000..9da20853edca4914472c3e9e2a5720550bc70e46 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d507c717ea33536ad243d366146f0827cdd051daac76733be4440aeaf7d48a57 +size 67108703 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds new file mode 100644 index 0000000000000000000000000000000000000000..ff8e706154937944a532cffad67a82421217e7c7 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4eab4838791d1e9c46e12e66b95d0cf3d99d68d3641637c018f81168fc9b17b1 +size 67108000 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds new file mode 100644 index 0000000000000000000000000000000000000000..91135f9e734681a7e113caa03643be017fa678c3 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14874b64b4572e8ca5a49e545c57cf823adebf00d3e813983c7fa252df9c15e1 +size 67107664 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds new file mode 100644 index 0000000000000000000000000000000000000000..3fcf52d42dc0dcee08b2919adc3f0753b28b6b29 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2089d8faa3df537dc0031cf5206cba1c921e37139712915487d1906d4915e53 +size 67107211 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds new file mode 100644 index 0000000000000000000000000000000000000000..5d97cbe009ea218fa221f9ecb3d5a356d82781bd --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec9375dfe6ff286fcde59aac00900b169b871e81ed684ad729b734c0b645c050 +size 67108415 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds new file mode 100644 index 0000000000000000000000000000000000000000..07eee504454ad9f30aef1673d38e83efbd83af2b --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e9ae730d942d6b2d86405ff51dc6c52f11eab249f7fb82601c505fb4466601d +size 67108820 diff --git a/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds new file mode 100644 index 0000000000000000000000000000000000000000..fcfa9c5d45a63edcec132ac40e89742606184022 --- /dev/null +++ b/train/wikipedia_dup/wiki_0001-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6440649473c5d823e7e1e2b75a4ee961393c66a333f98d3be4c46b1a7c09e555 +size 31379957