Datasets:
Tasks:
Token Classification
Modalities:
Text
Formats:
parquet
Languages:
Yue Chinese
Size:
10K - 100K
License:
dataset_info: | |
features: | |
- name: chars | |
sequence: string | |
- name: labels | |
sequence: | |
class_label: | |
names: | |
'0': D | |
'1': I | |
'2': P | |
'3': S | |
- name: logits | |
sequence: | |
sequence: float32 | |
length: 4 | |
splits: | |
- name: train | |
num_bytes: 175352249 | |
num_examples: 46033 | |
- name: validation | |
num_bytes: 9697876 | |
num_examples: 2557 | |
- name: test | |
num_bytes: 9693002 | |
num_examples: 2558 | |
download_size: 144555783 | |
dataset_size: 194743127 | |
configs: | |
- config_name: default | |
data_files: | |
- split: train | |
path: data/train-* | |
- split: validation | |
path: data/validation-* | |
- split: test | |
path: data/test-* | |
language: | |
- yue | |
license: mit | |
task_categories: | |
- token-classification | |
This segmentation dataset was generated with [AlienKevin/electra-hongkongese-base-hkcancor-multi](https://huggingface.co/AlienKevin/electra-hongkongese-base-hkcancor-multi) on [R5dwMg/zh-wiki-yue-long](https://huggingface.co/datasets/R5dwMg/zh-wiki-yue-long). | |
**See https://github.com/AlienKevin/dips for details.** |