orionweller commited on
Commit
26bf524
1 Parent(s): 3a91610

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +34 -0
  2. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  5. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  6. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  8. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  11. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  12. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  13. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  14. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  15. train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  16. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  17. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  18. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  19. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  20. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  21. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  22. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  23. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  24. train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  25. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  26. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  27. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  28. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  29. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  30. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  31. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  32. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  33. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  34. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  35. train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds +3 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14286-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14286-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16397-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16397-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16506-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16506-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16528-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16528-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19203-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19203-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21289-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21289-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23750-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23750-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25601-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
.gitattributes CHANGED
@@ -12806,3 +12806,37 @@ train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
12806
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12807
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12808
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12806
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12807
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12808
  train/arxiv/arxiv_0070-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12809
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12810
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12811
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12812
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12813
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12814
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12815
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12816
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12817
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12818
+ train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12819
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12820
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12821
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12822
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12823
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12824
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12825
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12826
+ train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12827
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12828
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12829
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12830
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12831
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12832
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text
12833
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12834
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12835
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12836
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
12837
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12838
+ train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12839
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12840
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12841
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12842
+ train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b661e9b50b79eb35288fffea1b7f3c23bda79fb18e7c0ccbb5bdc2d5e1012a28
3
+ size 67107053
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ece9fd780741e014c6e6c8e3c57a3c181a8bfac7affd316e58cdde5267676896
3
+ size 67107517
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bbdf5188aaa2b26060bf20f399ebaa824441144edcfa8e4d3eae72bff7638b5
3
+ size 67107529
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f32fd0ad97d972ea9337ff8580a89beda57a8d45ab09ae8a38355e9fef0e33ce
3
+ size 67107477
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65faa7105d4b79411d131b29acccb7d736181d67d4d573c9910c4823a285ad83
3
+ size 67107389
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b772374ae6caff097311a43396c0a2fcd57e3ba3af126df3cc25d498745ed3f7
3
+ size 67108364
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b91ad8192d4a9a35760a2612882af450e0ceb77f2c2699d97cf890dc82d29e5e
3
+ size 67108615
train/arxiv/arxiv_0013-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c69592406b8b577c572f28fd5f2f55b91029bfb54354ae0e87d8318a62442448
3
+ size 65771457
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55c4c95d00a08cbe938ac34d6ad1f37f55bc8ff97ad26fa55417954a6a30b0ae
3
+ size 67107187
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:427b9f7d636119ac329c328d50622b205d6eaea3bd1f1e888e15da149f01f67a
3
+ size 67108769
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92cabfadc2d9a06517ea9d0b42685db05bd5649620717896ec929afca8f83f76
3
+ size 67107856
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba023a25d3a344debc312d5ec4f558cc856dec62b39182efeb236de62dea999a
3
+ size 67108606
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb76e05323fb6040986c84311ecc3f0e143cc0239b068bdf2128c72f2c3ec646
3
+ size 67106830
train/arxiv/arxiv_0040-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3be86212b249e2778061767bca8b6562200f4a091d1819abf85fd2d437204f8b
3
+ size 9285209
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7def4e2144b2863b1c5f2f0b158e55dcf0bac9219e2f238e67519d6984b041a
3
+ size 67107285
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77a2b358e38c11f94126e5c3d0b58c8b6661af7361658a7d239403b670dbcdfe
3
+ size 67108783
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:257e9e3cf93ac20a6bef6ab1093ef00aafa9f1465341a23e35b53fde5a662feb
3
+ size 67108476
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b9111d7841cf92eec6c44e0c306ea74a60f615b1627e5d64f3dca5133dd501b
3
+ size 67108599
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65dc2ff50a725e41aaee48af0bcbe289c126cac728f6db7232f5450e8b97c1bf
3
+ size 67108249
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79ebd3ed347cdc442e5b99bcdab8624ec52873e7db32f99fc4b36969159609cf
3
+ size 67108795
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94728f0a8e1f0778a0649260062a7c45d5505e2440dc03677968ea6b703e8068
3
+ size 67108032
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bb42351006160f1be2ee4986fb57fe7df30ea7588c3c48cda2a5ab1bb89b675
3
+ size 67107227
train/arxiv/arxiv_0057-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fbe477b5541a21ee51e3c46fd1a2863e03aff48c5924313bf076a916d2c849f
3
+ size 41140404
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:357a2d7aef0591b36c44c14406162e62033aea507bff8fba3ad17e7d456c852a
3
+ size 67107154
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bf5c5b9252821ac0a2c9c34c45d7b8388f183c9f52d096c50d4f58b2b48c05a
3
+ size 67108484
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e29db58cf8396a9bd84beee68dc7390b75cee3f1f648998ccdfeab3ac3b2a3bc
3
+ size 67106976
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b025afad640a734a413c757220f43f92a751b2713af3a00622ace16b18279ab
3
+ size 67108687
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8346028b88c7ab087f91ebdb66d277da168c9d3b0c19c3e4f8b2322c161422f1
3
+ size 67107488
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de03e6254ddc1856a8e441b4a9653c0c78897f6fbe58d66c7ebaf4ee1fb508b6
3
+ size 67108235
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6140a1f414ba75b6acb3adf43f3cd53a71ccca28929b7895c23dc94af70f76cf
3
+ size 67107020
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:260bd26fab95294c424f1570b639e72e5071b0305a3ab400dea209721d56314f
3
+ size 67107731
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6af374ed02c0d4c77b341b6c78404b47d1442ab57f6c2af7b3d91691045aee6
3
+ size 67107783
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:268bac27afbbcae4de79caf8a8b574a5fc365fe0861f3058725a107b4f5f7f1a
3
+ size 67107846
train/arxiv/arxiv_0072-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a88d2eec94349bc9b5589a0e46a6ff073c9a6d39243a7868ccb12e2a20ec22d4
3
+ size 7677926
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14286-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106997, "hashes": {}}, "samples": 42959, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47925612, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18301406, "hashes": {}}, "samples": 11719, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13113380, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14286-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41011577,
3
+ "num_truncated_tokens": 40977023
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16397-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108100, "hashes": {}}, "samples": 42426, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47435039, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21822502, "hashes": {}}, "samples": 13988, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15587325, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16397-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42719014,
3
+ "num_truncated_tokens": 42681505
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16506-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107344, "hashes": {}}, "samples": 45156, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47992807, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 6652949, "hashes": {}}, "samples": 4323, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 4763854, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16506-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 35349032,
3
+ "num_truncated_tokens": 35322312
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16528-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108504, "hashes": {}}, "samples": 44226, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47536204, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9857900, "hashes": {}}, "samples": 6557, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7020610, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_16528-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36911527,
3
+ "num_truncated_tokens": 36883894
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19203-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107739, "hashes": {}}, "samples": 43011, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47562088, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18439560, "hashes": {}}, "samples": 11853, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13068965, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19203-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41075367,
3
+ "num_truncated_tokens": 41040582
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21289-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108262, "hashes": {}}, "samples": 42547, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47338630, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21010171, "hashes": {}}, "samples": 13422, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14870533, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21289-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42326679,
3
+ "num_truncated_tokens": 42290281
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23750-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108241, "hashes": {}}, "samples": 44066, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47612723, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13703404, "hashes": {}}, "samples": 8756, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9735105, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23750-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38771031,
3
+ "num_truncated_tokens": 38739392
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25601-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107763, "hashes": {}}, "samples": 43021, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47641477, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19643700, "hashes": {}}, "samples": 12509, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13918706, "hashes": {}}}], "version": 2}