orionweller commited on
Commit
0070df0
1 Parent(s): a5f39b4

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +32 -0
  2. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  5. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  6. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  8. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  11. train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  12. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  13. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  14. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  15. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  16. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  17. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  18. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  19. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  20. train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  21. train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  22. train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  23. train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  24. train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  25. train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  26. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  27. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  28. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  29. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  30. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  31. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  32. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  33. train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18191-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18191-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20666-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20666-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2124-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2124-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25630-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25630-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26575-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27417-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27417-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29671-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29671-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_35198-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_35198-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_37056-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_37056-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -12371,3 +12371,35 @@ train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
12371
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12372
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12373
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12371
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12372
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12373
  train/arxiv/arxiv_0052-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12374
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12375
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12376
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12377
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12378
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12379
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12380
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12381
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12382
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12383
+ train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12384
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12385
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12386
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12387
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12388
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12389
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12390
+ train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12391
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
12392
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12393
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12394
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
12395
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12396
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
12397
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12398
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
12399
+ train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
12400
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
12401
+ train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
12402
+ train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
12403
+ train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
12404
+ train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
12405
+ train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15bd437ba27431ae4adf918bf9314fe8b242c4c59f960347ec787202b2ce7026
3
+ size 67107668
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69898bee4a771fbb1788601fa1d8b841643399f61edce29128743122eee4349e
3
+ size 67108460
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1653bb78f415e87a5f1e0eb2752ce80b5edb2a37446259c0727953f94f39d38
3
+ size 67107557
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7f27faf34f4d3900669ce59434ed3427a59852575607cc2e55486313b0ff601
3
+ size 67107930
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8a08890d1ee703a5e49936e4b7336a8f5d260527814ba3d38289ec14524dab0
3
+ size 67108024
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a683604b0eb3a9cf0743967aff9432723b9954ca3122ed37a113427d8d71f7b
3
+ size 67108846
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:591c1b2faab6fb0ad23429fce815448137786e26a71f2b552da509fcb0d7f6e9
3
+ size 67108643
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5b6adfd9b6e8c8e98893acaa0fed92c0967b5a1dc8f75a839e7d9c1be1000e7
3
+ size 67106821
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ac10259d058525d6a4213307b2a3b1441b0dc893ac5176ec70fe09408a03869
3
+ size 67107184
train/arxiv/arxiv_0042-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d345b022e08369e53c7a210fbdcf2822097630050f8eb79f29022eac41ea2516
3
+ size 6526565
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fcf82bb818aa4daaca3b43f5b1b6c4612bd4940b0049f1cefb01a3eccea179b
3
+ size 67108164
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08077e5415cf0bf90c4cd232e8d0221392ca69183f3496dcd20892e8105e3fcf
3
+ size 67107235
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e9403a20994f1b0957b9cd647a789ddd38da0db37eee313895c1a27ac4dd45d
3
+ size 67107013
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5049819e4057e0c817f5b0ccdd3542a0a0613ae4def693e251fd38fc54ba7bd
3
+ size 67108677
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0659dd3c36f8703b9c4de57bf5f0bbddfb8ad26e3fa7202e301476ce2c0f75d
3
+ size 67108464
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9796b496213f670bebc2e91241bc1562aef1e9f3d006b489a860d631a9fa6ab7
3
+ size 67108824
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07b3125246a1c2df41b075f558b4f7d5a6a03bad27a73fb1c5856878cb252626
3
+ size 67108248
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbcb3b65b48f42ed933add0e2c4b4141c039f037c0ed0e528a2ab29f95409736
3
+ size 67108597
train/arxiv/arxiv_0044-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85bb817f74673505b3dda3449a2d350f57fd4cd97b3c2ef3ce30725d0f3a3a83
3
+ size 13813607
train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeed3a3822ae3ceee8bb0de110ed6f190eecf8d70f885a44d4bfe5139413fbd6
3
+ size 67108528
train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae8c9833ff033ebeaffbe26c960315fe91598ca683953bef743fcff84790b060
3
+ size 67107855
train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:013e7530639a5edcaf05dc6d7711e11414d562abd76ddbbad50dcdbe08cac885
3
+ size 67108790
train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3bd0288ce81019a8f2c879c25acf788994ff055a3942f937927dc902e19e1f5
3
+ size 67107114
train/arxiv/arxiv_0078-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56714cbde2e00c5fdfd5b94eccc5074207d22942283d271ce4b5a5a3b52b1108
3
+ size 12393412
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b77404d7659eccc463ffddb9f11147b82f2f606fe364c425cd536d39f9e8454e
3
+ size 67107917
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921c90fd79d6e0e5afa917107dd6b69c980ff8b6a6bd7c98dbc7dc0ebf3be33a
3
+ size 67108714
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b98f09228827381efd9869768acfe88eab1d5c5ecad54bc0d0cb362c53242bb3
3
+ size 67108439
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:307b44c06652d1aacc3486a24a7616693febd50f6a5734e22868871e5982a082
3
+ size 67108248
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e1fe2c70ff5da3ccd6409fb9b6f9aacba9f1eab16eb38977adae3f8eb87f34
3
+ size 67107577
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33c2ff47538a30cade4f486a04da833d66a1308873a35d16c5b1646f855d823e
3
+ size 67108805
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d84a69053bad2b9f1b0ff64f070e4fc85292b86d46bc5cba81a48d06e9200f52
3
+ size 67107624
train/arxiv/arxiv_0098-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e69b019379f27ac2b4e0dd5525d71d1b4aef08b6060bd9cd378304bd5fe16828
3
+ size 30793604
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18191-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108059, "hashes": {}}, "samples": 42393, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47625563, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22824158, "hashes": {}}, "samples": 14429, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16167406, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18191-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 43207169,
3
+ "num_truncated_tokens": 43168954
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20666-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107618, "hashes": {}}, "samples": 44300, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47800734, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9991370, "hashes": {}}, "samples": 6654, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7109948, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20666-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36972636,
3
+ "num_truncated_tokens": 36944410
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2124-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108305, "hashes": {}}, "samples": 44388, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47804008, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8740489, "hashes": {}}, "samples": 5803, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6236703, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2124-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36371032,
3
+ "num_truncated_tokens": 36344146
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25630-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107268, "hashes": {}}, "samples": 44148, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47606575, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9402575, "hashes": {}}, "samples": 6347, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6707173, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25630-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36692191,
3
+ "num_truncated_tokens": 36664762
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_26575-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108609, "hashes": {}}, "samples": 44304, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47884386, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10458431, "hashes": {}}, "samples": 6978, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7470991, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27417-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108084, "hashes": {}}, "samples": 44642, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47976312, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8780821, "hashes": {}}, "samples": 5721, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6273771, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_27417-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36385831,
3
+ "num_truncated_tokens": 36358030
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29671-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108165, "hashes": {}}, "samples": 43011, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47622285, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18617700, "hashes": {}}, "samples": 11934, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13179222, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_29671-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41162199,
3
+ "num_truncated_tokens": 41127196
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_35198-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107872, "hashes": {}}, "samples": 42478, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47747771, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21796078, "hashes": {}}, "samples": 13889, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15414885, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_35198-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42707114,
3
+ "num_truncated_tokens": 42669544
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_37056-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107645, "hashes": {}}, "samples": 43137, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47949948, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16995064, "hashes": {}}, "samples": 10911, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12113816, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_37056-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40378292,
3
+ "num_truncated_tokens": 40344514
4
+ }