orionweller commited on
Commit
0eb695f
1 Parent(s): 2b11b34

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +29 -0
  2. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  3. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  4. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  5. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds +3 -0
  6. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  7. train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds +3 -0
  8. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  9. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  10. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  11. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  12. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  13. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  14. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  15. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds +3 -0
  16. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  17. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  18. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds +3 -0
  19. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds +3 -0
  20. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds +3 -0
  21. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds +3 -0
  22. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  23. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds +3 -0
  24. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  25. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds +3 -0
  26. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds +3 -0
  27. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds +3 -0
  28. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds +3 -0
  29. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00050.mds +3 -0
  30. train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00053.mds +3 -0
  31. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1089-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1089-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13879-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13879-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15746-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15746-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17221-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17221-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18333-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18333-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21712-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21712-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2216-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2216-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23461-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23461-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28172-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28172-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28461-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28461-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -11640,3 +11640,32 @@ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/
11640
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11641
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
11642
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11640
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11641
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00033.mds filter=lfs diff=lfs merge=lfs -text
11642
  train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11643
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds filter=lfs diff=lfs merge=lfs -text
11644
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text
11645
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
11646
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11647
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11648
+ train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
11649
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds filter=lfs diff=lfs merge=lfs -text
11650
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
11651
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
11652
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
11653
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds filter=lfs diff=lfs merge=lfs -text
11654
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
11655
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
11656
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds filter=lfs diff=lfs merge=lfs -text
11657
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
11658
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00050.mds filter=lfs diff=lfs merge=lfs -text
11659
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds filter=lfs diff=lfs merge=lfs -text
11660
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
11661
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00053.mds filter=lfs diff=lfs merge=lfs -text
11662
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds filter=lfs diff=lfs merge=lfs -text
11663
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
11664
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
11665
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds filter=lfs diff=lfs merge=lfs -text
11666
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
11667
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds filter=lfs diff=lfs merge=lfs -text
11668
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds filter=lfs diff=lfs merge=lfs -text
11669
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text
11670
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds filter=lfs diff=lfs merge=lfs -text
11671
+ train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:114448aef6058c1348e69495d7a99a27760d5a1fc8a65adbad1fbc4aa099a3d1
3
+ size 67107548
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:646e07de6527dcc3694f44c33729089986dadc89f4338b23c50d7fc29219c3c8
3
+ size 67107514
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0fa2714a992e49c54f14d185e331f1bd8198ab196c9b3cd92acc80f2f0fc3a6
3
+ size 67108392
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57e95826d18262816fc7beafc0a989b63d59df6ae036e1e26ddc350716688502
3
+ size 67107543
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b48a52f4fb7db479a4ca4b2ad9d2bf08ffe93f108610efc782e6beaa2a20b86
3
+ size 67108774
train/cc_en_head/cc_en_head_0037-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:837d31e81befabeca65be9c4bf7b1d9420fc96ad03e9b6285d2b0d45b17ab4d5
3
+ size 67107773
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d35103e672ab1ed7ca1097e3fbfcb7fe3d6ede7d7febc186f45f3c498c28470
3
+ size 67108213
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c88d9540ec877d0230c8c135a89480e34ecb7aaa10b409924c8ca4891de747fb
3
+ size 67107840
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c868732f38584fa680fc0ffe5ab921b0e05c89a26791930778e7028efd452d
3
+ size 67108162
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d1a3f98d5bd5f8be584f13dfd3212a24bc4f7b5bf6bae0bb7f75507d1d72c23
3
+ size 67108756
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a559bdf699e11860e643b00efdce5e49af21381659b738a64830b4895d6da8a
3
+ size 67107861
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a317068243cccc7fdf9331b454e87f5f7bc4c30ee248bdead1c95d147e64e41f
3
+ size 67106950
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaad74b29979a7958966eeba15b9e991560b66121ea41e82149161cdf6541b05
3
+ size 67108069
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00015.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076fa9c9c1246aa3ff301ad78999f3d44bc9465ba339c02a57e2b1a4bd9218f6
3
+ size 67108416
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87f2babc5942bb0718338ffd6a50e206b1b9feeb07399ef8698aa1d39524b595
3
+ size 67108636
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2819d79c8c2182bb7eb4cbaa8f0eade0567c9b80115fd92845916555ab269abc
3
+ size 67106866
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00022.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b23e6d6fe006c1bbf8d070a16d331564d192e2e6fc96e91bcf635080f3b779ed
3
+ size 67107481
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00023.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b97273392157cc05b576f0d93c8dc313a454619a97c662bb36e949f74d7f440
3
+ size 67107721
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00028.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83618ef1e6b9eae31b5783483d0da31309c23b4354dfaa083f6af2788bc7feeb
3
+ size 67108422
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:051be0de8264ce5f190bce05755a630b4e60fcaac2ec7a0507a681463e1bef27
3
+ size 67108553
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6c827202a80ebf112c8cde5814107c79dfe7051b88fa9c84d5da2b70801b6c3
3
+ size 67108504
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00035.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d60e18398ed63e6c8e1278a8a8998a65e17f6ebe4bec4919b2addab838017aa
3
+ size 67107011
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40084de3482c2482bf98c6d5c99f87464a5b28d1e36e9097c3ea88c9ee62c9aa
3
+ size 67107091
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00039.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a4dc2a884853c29c268cdb22525330e538d0003959e11b86cd9786a4d99c28
3
+ size 67107318
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afddde14eb077755e8277085ed8de3d905de3f2a6252ae86625f55e97f290dfa
3
+ size 67107259
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00041.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4b9e98ab18a4e504f07bfe10ed5d7959f3dbbb2ac9544d9e919e98d64a7ee3d
3
+ size 67107974
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00046.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b8adc33af740b0adc20487613016dffa5c15ed8668b6ab353158fc4917f351f
3
+ size 67108204
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00050.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32c93508d5f4fb59dcd9fc40bde3b35b9674e89941a4348718a7feecd04f5370
3
+ size 67108707
train/cc_en_head/cc_en_head_0139-tokenized-chunked-1024-512-128-backfill-nodups/shard.00053.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:022706cb9ae6cf05a396eb3540e6b1afd71e4ac266eff06a04c3196f826118ae
3
+ size 46002251
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1089-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108645, "hashes": {}}, "samples": 43125, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48001802, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 19151687, "hashes": {}}, "samples": 11997, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13593972, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_1089-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41423826,
3
+ "num_truncated_tokens": 41388182
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13879-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108364, "hashes": {}}, "samples": 44081, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47762000, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10187100, "hashes": {}}, "samples": 6858, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7262772, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_13879-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37071213,
3
+ "num_truncated_tokens": 37042654
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15746-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107821, "hashes": {}}, "samples": 44084, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47813714, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9956415, "hashes": {}}, "samples": 6647, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7118167, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_15746-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36962078,
3
+ "num_truncated_tokens": 36933949
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17221-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107197, "hashes": {}}, "samples": 43324, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47435294, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17160805, "hashes": {}}, "samples": 11041, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12191910, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_17221-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40451230,
3
+ "num_truncated_tokens": 40417694
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18333-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107560, "hashes": {}}, "samples": 42966, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47575089, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20629534, "hashes": {}}, "samples": 12815, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14713669, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_18333-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42141881,
3
+ "num_truncated_tokens": 42105675
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21712-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107826, "hashes": {}}, "samples": 42734, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47434026, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20935543, "hashes": {}}, "samples": 13243, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14856769, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21712-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42288897,
3
+ "num_truncated_tokens": 42252454
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2216-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108424, "hashes": {}}, "samples": 44430, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47835674, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9995084, "hashes": {}}, "samples": 6640, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7137107, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2216-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36971356,
3
+ "num_truncated_tokens": 36942557
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23461-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107558, "hashes": {}}, "samples": 43867, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47662914, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 14382288, "hashes": {}}, "samples": 9162, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 10149395, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23461-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 39103693,
3
+ "num_truncated_tokens": 39071900
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28172-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108541, "hashes": {}}, "samples": 42534, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47513025, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22368472, "hashes": {}}, "samples": 14347, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15902948, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28172-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42977890,
3
+ "num_truncated_tokens": 42940283
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28461-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108718, "hashes": {}}, "samples": 43993, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47659220, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13300952, "hashes": {}}, "samples": 8767, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9493924, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28461-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38572103,
3
+ "num_truncated_tokens": 38540414
4
+ }