orionweller commited on
Commit
9f538c2
1 Parent(s): c39a967

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +32 -0
  2. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  3. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds +3 -0
  4. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds +3 -0
  5. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds +3 -0
  6. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds +3 -0
  7. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  8. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds +3 -0
  9. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds +3 -0
  10. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  11. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds +3 -0
  12. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds +3 -0
  13. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  14. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds +3 -0
  15. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds +3 -0
  16. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  17. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  18. train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds +3 -0
  19. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds +3 -0
  20. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds +3 -0
  21. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds +3 -0
  22. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds +3 -0
  23. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds +3 -0
  24. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds +3 -0
  25. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds +3 -0
  26. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds +3 -0
  27. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds +3 -0
  28. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds +3 -0
  29. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds +3 -0
  30. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds +3 -0
  31. train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds +3 -0
  32. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_10492-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  33. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_10492-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  34. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_11901-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  35. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_11901-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  36. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_1689-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  37. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_1689-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  38. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20786-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  39. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20786-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  40. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_23918-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  41. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_23918-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  42. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_29921-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  43. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_29921-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  44. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30174-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  45. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30174-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  46. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32454-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  47. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32454-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  48. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34912-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
  49. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34912-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json +4 -0
  50. train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_3682-tokenized-chunked-1024-512-128-backfill-nodups/index.json +1 -0
.gitattributes CHANGED
@@ -27140,3 +27140,35 @@ train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_46997-tokenized-chun
27140
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_59689-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
27141
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_59689-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
27142
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_46997-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
27140
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_59689-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
27141
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_59689-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
27142
  train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_46997-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
27143
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_46566-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
27144
+ train/mlfoundations-dclm-baseline-1.0-parquet-sampled/split_81884-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
27145
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds filter=lfs diff=lfs merge=lfs -text
27146
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
27147
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds filter=lfs diff=lfs merge=lfs -text
27148
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds filter=lfs diff=lfs merge=lfs -text
27149
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds filter=lfs diff=lfs merge=lfs -text
27150
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds filter=lfs diff=lfs merge=lfs -text
27151
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
27152
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
27153
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
27154
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds filter=lfs diff=lfs merge=lfs -text
27155
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
27156
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds filter=lfs diff=lfs merge=lfs -text
27157
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds filter=lfs diff=lfs merge=lfs -text
27158
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
27159
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds filter=lfs diff=lfs merge=lfs -text
27160
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds filter=lfs diff=lfs merge=lfs -text
27161
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds filter=lfs diff=lfs merge=lfs -text
27162
+ train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds filter=lfs diff=lfs merge=lfs -text
27163
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds filter=lfs diff=lfs merge=lfs -text
27164
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds filter=lfs diff=lfs merge=lfs -text
27165
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds filter=lfs diff=lfs merge=lfs -text
27166
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds filter=lfs diff=lfs merge=lfs -text
27167
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds filter=lfs diff=lfs merge=lfs -text
27168
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds filter=lfs diff=lfs merge=lfs -text
27169
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds filter=lfs diff=lfs merge=lfs -text
27170
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds filter=lfs diff=lfs merge=lfs -text
27171
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds filter=lfs diff=lfs merge=lfs -text
27172
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds filter=lfs diff=lfs merge=lfs -text
27173
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds filter=lfs diff=lfs merge=lfs -text
27174
+ train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds filter=lfs diff=lfs merge=lfs -text
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d01577922e79d38c249d13c32bb7c681fd3e455e587cd202308275a36c9bf71
3
+ size 67108841
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00001.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca10d40ff618ea8b688d30fd7e70c6bf1ea0707f825ebd0082dfc31481f9710d
3
+ size 67108099
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00002.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8ce20afa86800536393d94d56cc5d71b8b9db1c44acf91df6c694a3d492ce1
3
+ size 67107354
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00003.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52d959a0bfbd17b6f4bb255cc8b813d692dfdd9a0605a20472cfe8004d761338
3
+ size 67107399
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00004.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3064dc93c5e84a3900e6e10411e35b0033f5ea92fb2a8b94e208f746e963c3f1
3
+ size 67107102
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b18935c8df10416bb77aab40add80d2e4535c4812a5195994a90fe4c51793532
3
+ size 67108545
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00006.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:247f708c04142848b7cf3b81c4f633e114bd43913895b96f2ad711fdc412f83e
3
+ size 67107545
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00007.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fa7a9fe5101e2d41680259ddcf9ed442e51007d5a03aff1951721d64496005a
3
+ size 67107404
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82115528445ea6098261d1eefd6e9ef02d4b467ca5d8a8465e4266d2ecfb7e2b
3
+ size 67106777
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00010.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e40f3cf87dfcfda936a4b1d47007f8cb5e8064ea2809063774ca7328cc1d1ea8
3
+ size 67108228
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00011.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4ef9bbd1af25031063ca58dbe32809e7dc55515a99b8da5b6f7eee33e066ee7
3
+ size 67107848
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:706c70a0d504d9e637c8d525cc0cd34d4f6b696b67b20b9ff9c5ad4f4aa60bc4
3
+ size 67107415
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00013.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82fe3b644fe862ae24924395d08c77fc5b503faf7a3a3544bbd4a324b1ed2625
3
+ size 67108306
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00014.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2da78dd5c6d81cdb7067ffb69dcb8113679aed86b46d6fd3fc597ad603f7e84f
3
+ size 67107472
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:103a18ffb1dd64a1f63d8ea7d5689e4551138cf93f343691367ff2cb7c79c9ab
3
+ size 67108816
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9852019c6c5c2d24828e1f31c24f91d8216d9b4031bfb0d7f8a57dc1956c937
3
+ size 67107135
train/algebraic-stack/algebraic_stack_train_0011-tokenized-chunked-1024-512-128-backfill-nodups/shard.00018.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:caf1ca9c11e278866fc7a0ad6881fa2b6abe368af5c58a5635ad95ac372b6e2b
3
+ size 62710419
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00000.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af454ea3b5c8e94cf862910a16b52951ca73eb76e7e05a01a61de592d420a08e
3
+ size 67108363
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00005.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5095eee7bdd57e21bcae2b25bcfcaad065dee62b197ec4a42b2f11f741a80478
3
+ size 67107528
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00009.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90a9ab07bf97095a315c74add384005957e0bba15f96d38eb71d5a37d7d54691
3
+ size 67107346
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00012.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9778dd79933ee6456ad9e164ecebf51d297f90fc922bb904f859d1fda13bee08
3
+ size 67107026
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00016.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:026e4ce8701f807beed9ad6e62aaea3a467d80d655e303de203acfe32523aa15
3
+ size 67106943
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00017.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f232c26d51b5d8edefe28e16f5040c2dc63e7f6d7a82417128caa51562b0d495
3
+ size 67108862
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00021.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:991b03e512593bcaba3122e45dda4b993b8b451655a0de067f430a89b0cd836a
3
+ size 67107129
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00029.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:380b7665c86ff960fecd59fb76e153eddb5acbb00a7516b4af61fcbc496df857
3
+ size 67107278
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00032.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b634c68d0c6d697da62d5157ba12ef5e3f8d8eece5cb556a37122edadb743d78
3
+ size 67108283
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00034.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9561be5cd691371d73a3f001feb145a1cf32db2504e77f682fd51a17df0f77f1
3
+ size 67107177
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00036.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a9195633d631bcc25ee07068274ae946363a84cb6b0aa0f376fe335e107ec8b
3
+ size 67106946
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00037.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46119686fc319440b820e2c38d92ac81f331cb50051ec24ab4cf06f8a6d38324
3
+ size 67106781
train/algebraic-stack/algebraic_stack_train_0014-tokenized-chunked-1024-512-128-backfill-nodups/shard.00040.mds ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:187a1e1bae449f375fe887be8ea4d38281b361974d6a688246169d22d4f139f0
3
+ size 67107577
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_10492-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107129, "hashes": {}}, "samples": 44529, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47935603, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 9472713, "hashes": {}}, "samples": 6258, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6780669, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_10492-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36718244,
3
+ "num_truncated_tokens": 36690058
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_11901-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107558, "hashes": {}}, "samples": 44510, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47831963, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8907074, "hashes": {}}, "samples": 5851, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 6340361, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_11901-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36448792,
3
+ "num_truncated_tokens": 36421600
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_1689-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108856, "hashes": {}}, "samples": 43710, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47415050, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 12393884, "hashes": {}}, "samples": 8099, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 8768780, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_1689-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38147762,
3
+ "num_truncated_tokens": 38118878
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20786-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108765, "hashes": {}}, "samples": 44043, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47626526, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 10785413, "hashes": {}}, "samples": 7066, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7635202, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_20786-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37365222,
3
+ "num_truncated_tokens": 37336832
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_23918-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108258, "hashes": {}}, "samples": 44715, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47781126, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 8364449, "hashes": {}}, "samples": 5397, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 5936579, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_23918-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 36185545,
3
+ "num_truncated_tokens": 36158460
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_29921-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108382, "hashes": {}}, "samples": 43367, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47792336, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 18445658, "hashes": {}}, "samples": 11448, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 13083616, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_29921-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 41080327,
3
+ "num_truncated_tokens": 41045860
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30174-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107329, "hashes": {}}, "samples": 44206, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47779942, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 11001207, "hashes": {}}, "samples": 7300, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 7847876, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_30174-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 37460335,
3
+ "num_truncated_tokens": 37431293
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32454-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108618, "hashes": {}}, "samples": 42827, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47562897, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 20543821, "hashes": {}}, "samples": 13121, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 14631297, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_32454-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 42094514,
3
+ "num_truncated_tokens": 42058317
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34912-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67107806, "hashes": {}}, "samples": 43926, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47785998, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 13732621, "hashes": {}}, "samples": 8874, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 9737621, "hashes": {}}}], "version": 2}
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_34912-tokenized-chunked-1024-512-128-backfill-nodups/num_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "num_tokens": 38786116,
3
+ "num_truncated_tokens": 38754436
4
+ }
train/mlfoundations-dclm-baseline-1.0-parquet-sampled-v3/split_3682-tokenized-chunked-1024-512-128-backfill-nodups/index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"shards": [{"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00000.mds", "bytes": 67108402, "hashes": {}}, "samples": 42480, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00000.mds.zstd", "bytes": 47716813, "hashes": {}}}, {"column_encodings": ["str", "ndarray:uint16"], "column_names": ["id", "input_ids"], "column_sizes": [null, null], "compression": "zstd", "format": "mds", "hashes": [], "raw_data": {"basename": "shard.00001.mds", "bytes": 22653306, "hashes": {}}, "samples": 14389, "size_limit": 67108864, "version": 2, "zip_data": {"basename": "shard.00001.mds.zstd", "bytes": 16005484, "hashes": {}}}], "version": 2}