orionweller commited on
Commit
3fb9e75
1 Parent(s): a401604

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +29 -0
  2. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  3. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  4. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  5. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  6. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds +3 -0
  7. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  8. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  9. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  10. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds +3 -0
  11. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  12. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds +3 -0
  13. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds +3 -0
  14. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  15. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds +3 -0
  16. train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds +3 -0
  17. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  18. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  19. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  20. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  21. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  22. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  23. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds +3 -0
  24. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds +3 -0
  25. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  26. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds +3 -0
  27. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  28. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds +3 -0
  29. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds +3 -0
  30. train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds +3 -0
  31. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_116-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_116-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13312-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13312-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13490-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13490-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14011-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14011-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15343-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15343-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18794-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18794-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20930-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20930-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2633-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2633-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26820-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26820-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2993-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2993-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -11083,3 +11083,32 @@ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/
11083
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text
11084
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11085
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11083
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00020.mds filter=lfs diff=lfs merge=lfs -text
11084
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00019.mds filter=lfs diff=lfs merge=lfs -text
11085
  train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11086
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11087
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
11088
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
11089
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds filter=lfs diff=lfs merge=lfs -text
11090
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11091
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
11092
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
11093
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11094
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds filter=lfs diff=lfs merge=lfs -text
11095
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text
11096
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
11097
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds filter=lfs diff=lfs merge=lfs -text
11098
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
11099
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11100
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
11101
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
11102
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
11103
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text
11104
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds filter=lfs diff=lfs merge=lfs -text
11105
+ train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11106
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
11107
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11108
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds filter=lfs diff=lfs merge=lfs -text
11109
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text
11110
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11111
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11112
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11113
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11114
+ train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:531131c193003ea12b2e56354934d83342679ced66815b132803cc32a4cb866b
3
+ size 67108781
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7fb13f5536c9df6d53b56eee4a960fef813fdf962276cf998baff925ad8d35b
3
+ size 67108226
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8373ebf2c3e70682d02272181bd7484a74a117256f16eb15a6e5526eb0bb3460
3
+ size 67107941
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80ec9729053b662cc9e50ceb7ed8ef3aad661e7dec6f636b63f71b97d16d9180
3
+ size 67108689
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96885533503a8fae0f80c5eda4a2ad1cdb8b9d8df0988e60909e110b2bebe12b
3
+ size 67106851
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bc89896f6d7433438c043290c32ff1fd658384d63fca4cf922f2791f2c330d1
3
+ size 67106906
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:169cea3de0000f4686c93b1396b023db483c24715de42692cc6bb01f128ef0e4
3
+ size 67107663
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eb686abf1531f5e0d1120392a746b0d9fc427e9e3fccd04d7be8c6bb6b1a3b1
3
+ size 67108390
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b565845f85cca3a4139031a35e3f95ac37f01ac1654d209aee3e448623a94ed5
3
+ size 67107581
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9a4e9ddef542f2cfaf86a0df0cf8c3a72c9349f14685939537fca40c37001db
3
+ size 67108307
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3e9773a96d412a320a120f398a454426b221ea054f2e1164db2733aac578022
3
+ size 67108415
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00031.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:525e1cb2aa7eb7956b4b7ad34f18546d3d0ab83ffd17cd71f824df25f2816fb9
3
+ size 67108654
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f36d5d6f54ff363a3b38c669de70c282cb39e6c1d0b747ec98679e047b85164d
3
+ size 67108819
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86497ad5c3a66ee25cb033b1357ccd08597dcb0318e60e91d09e4c90288a4f7f
3
+ size 67108286
train/cc_en_head/cc_en_head_0117-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7debebbc77eac953f8651b9ebe8219238ed2e2361f8c1e31e11a766ba082812
3
+ size 67107309
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60a8d847d1f4a25b9cd95aa34f047266ef0ed74be02afbcf612c9a436f206b11
3
+ size 67106969
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4197d026bff8bbcd0463e30b7a919bc08f91a808434b2ea5294a904046eaf74e
3
+ size 67107776
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99a3b09172f2141bcdfceb09ea1fe068d815f8dac36c258f072058d13656fbe0
3
+ size 67107829
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d01f4329a9d791bfd33e03c7ed99dab8cfba3dbf24d3316ceb7feff136468e5
3
+ size 67108425
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd0bd261e241c47ea8c5dde351b3a8d5ba74ef9a973b70dc7cd437b3b5b24b59
3
+ size 67107824
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9168dd48aaad2f711738623c55c361584294bdf2a3c254359a3f582d6cb85c0d
3
+ size 67108066
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97ca4eb47705c2a8299f45e78cf26a66bcef8235e268d5946b766955c61e962b
3
+ size 67107728
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00025.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59f7b2b01034d677931d26b982c3e194fa66763b708aa0a4d4765c1a4a342d56
3
+ size 67108157
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdb5392eac9aa62406d65a9868a929991b348c67b9f799346701e1ba10597351
3
+ size 67107548
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be556e6d3e09f8a260618841e706f7c851a01797b8d43514d4381885689ce20c
3
+ size 67107336
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae361158dce03fb0fff246b88da80d0f72e9dc7016c910f8a58a0f9f0a41ab01
3
+ size 67107745
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4449c81539d5a30a94bcd0ddb4420d860c5281e3d69bc9e98536bb13d993bb7
3
+ size 67107880
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00038.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:145e5c022e64fdeae23dc2a53af43436fb55450c2ddff736d5916d338773a400
3
+ size 67107698
train/cc_en_head/cc_en_head_0237-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2411955961e08566a130eb43f78b256a9969ef0810e121f0c47bad82c0f170f
3
+ size 58586054
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_116-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108604, "hashes": {}}, "samples": 44608, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47909334, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8594688, "hashes": {}}, "samples": 5770, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6148798, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_116-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36292616,
3
+ "num_truncated_tokens": 36266104
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13312-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108757, "hashes": {}}, "samples": 44800, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47907929, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8163948, "hashes": {}}, "samples": 5495, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5862420, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13312-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36079982,
3
+ "num_truncated_tokens": 36052146
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13490-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108705, "hashes": {}}, "samples": 42508, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47509852, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20914248, "hashes": {}}, "samples": 13533, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14902343, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13490-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42276779,
3
+ "num_truncated_tokens": 42239757
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14011-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108295, "hashes": {}}, "samples": 44429, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47805990, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10085129, "hashes": {}}, "samples": 6700, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7179474, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14011-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37014487,
3
+ "num_truncated_tokens": 36985307
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15343-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108625, "hashes": {}}, "samples": 43428, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47643796, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17368539, "hashes": {}}, "samples": 11038, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12289250, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15343-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40552771,
3
+ "num_truncated_tokens": 40518851
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18794-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107226, "hashes": {}}, "samples": 43379, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47610600, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16761325, "hashes": {}}, "samples": 10906, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11859091, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18794-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40254173,
3
+ "num_truncated_tokens": 40219868
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20930-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106789, "hashes": {}}, "samples": 43625, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47776719, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15277700, "hashes": {}}, "samples": 9758, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10858812, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20930-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39539960,
3
+ "num_truncated_tokens": 39507648
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2633-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108462, "hashes": {}}, "samples": 44669, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47878751, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8442716, "hashes": {}}, "samples": 5472, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5963714, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2633-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36223805,
3
+ "num_truncated_tokens": 36196847
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26820-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108424, "hashes": {}}, "samples": 43699, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47616774, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12653329, "hashes": {}}, "samples": 8315, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8970470, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26820-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38271002,
3
+ "num_truncated_tokens": 38240892
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2993-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107682, "hashes": {}}, "samples": 44141, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47662424, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10733737, "hashes": {}}, "samples": 7099, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7638984, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2993-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37334926,
3
+ "num_truncated_tokens": 37306124
4
+ }