orionweller commited on
Commit
0272ec3
1 Parent(s): 8be0470

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +38 -0
  2. train/arxiv/arxiv_0012-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  3. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  4. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  5. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  6. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  7. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  8. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  11. train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  12. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  13. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  14. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  15. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  16. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  17. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  18. train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  19. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  20. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  21. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  22. train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds +3 -0
  23. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12651-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  24. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14348-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  25. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14348-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  26. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19157-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  27. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19157-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  28. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19902-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  29. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19902-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  30. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20738-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  31. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20738-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20794-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20794-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21094-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21094-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22840-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22840-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23730-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23730-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25644-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28191-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28419-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2884-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2884-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30471-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30471-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
.gitattributes CHANGED
@@ -13064,3 +13064,41 @@ train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.0000
13064
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13065
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13066
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13064
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13065
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13066
  train/arxiv/arxiv_0081-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13067
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
13068
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13069
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13070
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13071
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
13072
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13073
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_88320-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13074
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31799-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13075
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28419-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13076
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12651-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13077
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48402-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13078
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_31799-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13079
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_48402-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13080
+ train/arxiv/arxiv_0012-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13081
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_8713-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13082
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13083
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28191-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13084
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_49686-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13085
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_8713-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13086
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_49686-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13087
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_33793-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13088
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13089
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13090
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13091
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13092
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13093
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13094
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
13095
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
13096
+ train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
13097
+ train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
13098
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
13099
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds filter=lfs diff=lfs merge=lfs -text
13100
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
13101
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
13102
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
13103
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
13104
+ train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
train/arxiv/arxiv_0012-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0b25698bc70f37e97dc446267de23bd5f960478df33609a3272b2bc2f6d491c
3
+ size 67106775
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b554c4f030cd4da4212d0aa27a9771d04ef08aa7ddf2943cfabd595e6db58d0e
3
+ size 67106896
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb78ec4772cfbe4b7701739003ebb8a9e80fda13ee87716aa93072540a2b43e3
3
+ size 67106797
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41fe942d34e07967a80f1a855516e551544835153a01df92805ddc9b793adf88
3
+ size 67108055
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cf89fcd829b25ee7b6d05f67b60ce15626cd460e6796fb86998622ed8ae26d5
3
+ size 67106891
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42d09afb2b8424de6fe577813af0c7a489c4119bd050aa2fbe0cb23b66b860d9
3
+ size 67108616
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9a98fc0d347bef83b3de7943e34b631efe024b6c0d63a2d5e26a984e5095a2a
3
+ size 67106982
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65a5dc22cda12ba51fe7469af14b86b31c21eeeba4fdb382a45685b7630cc01f
3
+ size 67107465
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f641a8bf65975b528d975ab1239d83e0415969b0533017eb90f11f636210398c
3
+ size 67106878
train/arxiv/arxiv_0083-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30c615ac4a1053c3ae0f82d2df94f8936f12cbda821a685a4305c326f12cc96f
3
+ size 1849221
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2d255e143f666f9b5fe3da4c3388bfa7b68e9ea9d221057576509da9686e0b5
3
+ size 67108074
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef8e24f9e221f29e0cacb7c048b379c49f588f46ec1d020811e6a2766f50ea6c
3
+ size 67107812
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfd67c0019daad336f29f6cba8a7c6aebb0899355eeea408e42778f10d4c3a48
3
+ size 67107772
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9124f1b506e9079892ace0cc86355522d395024db0193bd3eb9f73d1b8df510
3
+ size 67108495
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0802b002887f238256edecf88e79c8a56b5bdacbd487756b41b882547c1a6ab
3
+ size 67107017
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e994689fdc4166875307fd1358097f861929aa8fa450c5f8187fdfbee97e0b15
3
+ size 67107601
train/arxiv/arxiv_0097-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5aa440580087f48b3ca3369e3a770af523a4b66af5c40f306f479d5866975267
3
+ size 56691514
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e88250f3bed18df313b0f01d3a23a3f0399388798d8bcc5789bb90e116535d0
3
+ size 67107656
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46f93f7c661e963ddc4e17be974d21a98a82d966ebd3c400cf76b29ff6024e77
3
+ size 67108418
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0271c958b3710d4a9ac06372289dbddba931e037d83243754053767acc428ac
3
+ size 67107109
train/arxiv/arxiv_0099-tokenized-chunked-1024-512-128-backfill-nodups/shard.00008.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6ae0cc3068e2920720578d26b9f44841f66374700bb9bea283b869ba8bd4b76
3
+ size 15704000
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_12651-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eb27c140e71614047c427abebd1b89c865746db253b4de2e1edb245d3e01297
3
+ size 67107444
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14348-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107779, "hashes": {}}, "samples": 44314, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 48092933, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10342462, "hashes": {}}, "samples": 6802, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7407294, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_14348-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37143104,
3
+ "num_truncated_tokens": 37113764
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19157-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108344, "hashes": {}}, "samples": 42576, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47710683, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 21180806, "hashes": {}}, "samples": 13568, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 15018512, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19157-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42406673,
3
+ "num_truncated_tokens": 42370080
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19902-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108090, "hashes": {}}, "samples": 43643, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47491682, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13330222, "hashes": {}}, "samples": 8622, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9391379, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_19902-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38601381,
3
+ "num_truncated_tokens": 38571640
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20738-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107735, "hashes": {}}, "samples": 44499, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47804359, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9523666, "hashes": {}}, "samples": 6402, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6798809, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20738-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36740585,
3
+ "num_truncated_tokens": 36711777
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20794-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108387, "hashes": {}}, "samples": 43528, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47579347, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 17130646, "hashes": {}}, "samples": 10822, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12127755, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_20794-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40437282,
3
+ "num_truncated_tokens": 40403600
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21094-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107466, "hashes": {}}, "samples": 43607, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47651872, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16689270, "hashes": {}}, "samples": 10469, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11749903, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_21094-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40224614,
3
+ "num_truncated_tokens": 40191329
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22840-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107214, "hashes": {}}, "samples": 44162, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47801616, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11913584, "hashes": {}}, "samples": 7754, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8504482, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_22840-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37903710,
3
+ "num_truncated_tokens": 37874004
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23730-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107895, "hashes": {}}, "samples": 43244, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47912628, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 16247817, "hashes": {}}, "samples": 10473, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 11622130, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_23730-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 40015079,
3
+ "num_truncated_tokens": 39981643
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6196c6eb5c2f720e5611467fbd310e1126b6a243b0714c0355f3f05474ff2f8c
3
+ size 67107176
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_24872-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76b29bbef3324be43f02c07d1b6c55548ff027a4e153bd60e598f518986a0db4
3
+ size 15924061
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_25644-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108559, "hashes": {}}, "samples": 44725, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47109741, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9482831, "hashes": {}}, "samples": 6178, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6704827, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28191-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b92d3fee0a00cae2e1f5cb1a248e06af51e780af978be61eb8b260e4bfd54070
3
+ size 17732071
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28419-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:184aef6d01f6fb9e3850ceb749cb1634600682ccc9fd42136aa8c3210c137bd2
3
+ size 67107563
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2884-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108682, "hashes": {}}, "samples": 43987, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47745972, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12099718, "hashes": {}}, "samples": 8080, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8655581, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_2884-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37992854,
3
+ "num_truncated_tokens": 37962190
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4909582b6788f72b8d3e0ebbee1e09d56a2306d672e8401738f8ff69d3bae498
3
+ size 67108629
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_28919-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1448ac62c740c6353fcf5d14731d7573691bf22b352c74e67a321dbb6072ed6f
3
+ size 17886552
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30471-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108348, "hashes": {}}, "samples": 43331, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47584212, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18296211, "hashes": {}}, "samples": 11549, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 12965934, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v2/split_30471-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41003611,
3
+ "num_truncated_tokens": 40969021
4
+ }