orionweller commited on
Commit
dc49379
1 Parent(s): 8c5e668

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +31 -0
  2. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  5. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  6. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  8. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  11. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  12. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  13. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  14. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  15. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  16. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  17. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  18. train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  19. train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  20. train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  21. train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  22. train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  23. train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  24. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  25. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  26. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  27. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  28. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  29. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  30. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  31. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  32. train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11229-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11229-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12808-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12808-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15408-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15408-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19543-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19543-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20076-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20076-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20648-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20648-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24478-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24478-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26002-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26002-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26671-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26671-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -12563,3 +12563,34 @@ train/arxiv/arxiv_0067-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
12563
  train/arxiv/arxiv_0067-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12564
  train/arxiv/arxiv_0067-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12565
  train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12563
  train/arxiv/arxiv_0067-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12564
  train/arxiv/arxiv_0067-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12565
  train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12566
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12567
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12568
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12569
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12570
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12571
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12572
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12573
+ train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12574
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12575
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12576
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12577
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12578
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12579
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12580
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12581
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12582
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12583
+ train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12584
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12585
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12586
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12587
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12588
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12589
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12590
+ train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12591
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12592
+ train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12593
+ train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12594
+ train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12595
+ train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12596
+ train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b76873a76a44c976dbbe09e1aa4e812f7ccac21c3ffd1fe1ed54036becafe8cd
3
+ size 67108011
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2201e03c16891971b30196c4739eeb99c5a8b967e10aac8656a54123b3fc3900
3
+ size 67108027
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e54bbf636b10b48db5a8c1bebae9f86ba806b2aded43c11106769f3f7df4bab
3
+ size 67108094
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dfc5fd5a66567a11cf8d68dc7bdd54a6d9951b3a62ca3205467aa1bf57a7ce2
3
+ size 67107333
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9214346c30de461458b78bebb1e657882d63fa49917afa3728e01e9f25b7bb91
3
+ size 67108681
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd4858423b20763f54879ab0235f5bfd0e231a9ad2d29cdb6b0031cca3fde30f
3
+ size 67107516
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bb5bedf1c9dd43d9cef8dcd92a454bc3215b2a80e3207f7ae88a05b22852eb4
3
+ size 67108856
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f59828b4cccfc012dae91a21aa5bcc19339fd8c62998c187a5dc1df2f37e37f3
3
+ size 67107253
train/arxiv/arxiv_0003-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d352d52240a36c56232874287fb5d7b09a996ad7ddf3d3c829ccfceda2b0f5
3
+ size 22220196
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e199198c9b184f6f00fc0b135277512208b1d5e8fad14bd91f86ac30b383779
3
+ size 67107745
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a78abd48b80a26f1b08523bd33d2f96f9a538ded568de64407c7ac2e322aea
3
+ size 67108653
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:852789aa03b55b470a9d1a4201ee548ec4cc0e5352028af90aeea9093b4ceb1f
3
+ size 67107872
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f3a21ec26c529537aeb35eb6fba000964aef54f5bf889661ae599eb733b8a81
3
+ size 67108562
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4931327aa30c9af31c1e473507e8bedebf6ade3570f89ed0438c1c3051a60886
3
+ size 67107455
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bd0a11c1b33668b7396c16c46643e18159ca5e27114df32f670a5e040572e68
3
+ size 67108168
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8163374b88041e1ce19f79d737515cb6cda25606b698a38946b8d9c14748aa6b
3
+ size 67107599
train/arxiv/arxiv_0021-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a6f71134e805782eeca311cf797ea370a30782fc6da057522c5d4fc5aac1455
3
+ size 3803332
train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c38c2fa1f97078ed0c69a9abbfeaceb9d063b0d2482bf41df3724c8308c3d94
3
+ size 67108600
train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8ddf11670580d0e3660f4c285a6572667d6557e14477a4f7db364c70f21f660
3
+ size 67106768
train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f601600b89fdded8eaa31452a5db5580b08af3ad4e4fb42490cb8e50d1a5551
3
+ size 67107549
train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f4772a90ce8cf1d525642b7a28c04a50c5c36b42431923f332a23c684ad75e5
3
+ size 67107032
train/arxiv/arxiv_0050-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8e1d198b7c8f010d697a1ed5133db9f4220b6b709d667978159b7a4bb847ef8
3
+ size 7801420
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb652b8c89289fff8cfc1c60bc85568bc361f5f1fb0460f90a623cd2eac461e6
3
+ size 67108665
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ff2c68f4a33f8cdac8344324345527bec319f16940487bb2a4525e0b2c9bf60
3
+ size 67107277
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d83032f674f526e1b25e432ebf8ef5b1d1cc3002ca8b88b403af61748848eda
3
+ size 67107270
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f07cbae421b4f296a6b21213185d8aa51a81e4bdbe34ac20f993e44f3393e19
3
+ size 67108464
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f2394c036d5b1170691e50afcd6b65fe8b8dbbccbd414e296b07822fa8abcc6
3
+ size 67107804
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91ce8c33ea86ce3749c6b1fdb368dd808772f290602eac7084e6033a44524d24
3
+ size 67108027
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fee4f21914208472ca5a9ce5336055eb04a4436ca515692318767c026d7a137c
3
+ size 67108787
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:354db5814c2f88161cdca5ebae15102c0ba95805690b602d393dd0ec81592be6
3
+ size 67108069
train/arxiv/arxiv_0087-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:221b0e32cc8b311a3c1090707597b5098b4dbf792d93d841070fb16c0f418e1a
3
+ size 43731676
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11229-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107191, "hashes": {}}, "samples": 44260, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48021499, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11804884, "hashes": {}}, "samples": 7570, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8474952, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_11229-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37851956,
3
+ "num_truncated_tokens": 37821396
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12808-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107428, "hashes": {}}, "samples": 44032, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47668307, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10572842, "hashes": {}}, "samples": 7103, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7522170, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12808-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37257559,
3
+ "num_truncated_tokens": 37229237
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15408-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108549, "hashes": {}}, "samples": 42846, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47608096, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19599186, "hashes": {}}, "samples": 12562, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13952856, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15408-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41638786,
3
+ "num_truncated_tokens": 41603502
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19543-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108736, "hashes": {}}, "samples": 42624, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47860403, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 23839312, "hashes": {}}, "samples": 14614, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16867854, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19543-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 43702020,
3
+ "num_truncated_tokens": 43663238
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20076-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67106863, "hashes": {}}, "samples": 43432, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48081219, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 15559583, "hashes": {}}, "samples": 9962, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11147979, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20076-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39680480,
3
+ "num_truncated_tokens": 39647938
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20648-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108397, "hashes": {}}, "samples": 43673, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47681137, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13893911, "hashes": {}}, "samples": 9158, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9935611, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20648-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38866041,
3
+ "num_truncated_tokens": 38834564
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24478-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107987, "hashes": {}}, "samples": 42683, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47606144, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21923851, "hashes": {}}, "samples": 13737, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15446158, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24478-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42769415,
3
+ "num_truncated_tokens": 42732260
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26002-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108224, "hashes": {}}, "samples": 44475, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47754045, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8873036, "hashes": {}}, "samples": 5821, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6306316, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26002-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36434077,
3
+ "num_truncated_tokens": 36406467
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26671-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107862, "hashes": {}}, "samples": 43255, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47673473, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17969229, "hashes": {}}, "samples": 11507, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12742759, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26671-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40843643,
3
+ "num_truncated_tokens": 40808845
4
+ }