Datasets:
Tasks:
Token Classification
Modalities:
Text
Formats:
parquet
Languages:
Thai
Size:
100K - 1M
Tags:
word-tokenization
License:
Delete old duckdb index files
Browse files
best2009/test/index.duckdb
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:f2508845d9f3168f908bc32933101ede4c2d710abd5b6c6396df2474966018b8
|
3 |
-
size 1847296
|
|
|
|
|
|
|
|
best2009/train/index.duckdb
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:1e7bf303347e24fd84202f28201d12384462086fb9596d03c5bd564f08f72c31
|
3 |
-
size 37498880
|
|
|
|
|
|
|
|