Datasets:

Languages:
Indonesian
ArXiv:
License:
holylovenia commited on
Commit
bbd3a53
1 Parent(s): b64e724

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +17 -17
README.md CHANGED
@@ -11,14 +11,14 @@ tags:
11
  ---
12
 
13
  A dataset of Indonesian News for Named-Entity Recognition task.
14
- This dataset re-annotated the dataset previously provided by Syaifudin & Nurwidyantoro (2016)
15
- (https://github.com/yusufsyaifudin/Indonesia-ner) with a more standardized NER tags.
16
- There are three subsets, namely train.txt, dev.txt, and test.txt.
17
- Each file consists of three columns which are Tokens, PoS Tag, and NER Tag respectively.
18
- The format is following CoNLL dataset. The NER tag use the IOB format.
19
- The PoS tag using UDPipe (http://ufal.mff.cuni.cz/udpipe),
20
- a pipeline for tokenization, tagging, lemmatization and dependency parsing
21
- whose model is trained on UD Treebanks.
22
 
23
 
24
  ## Languages
@@ -28,25 +28,25 @@ ind
28
  ## Supported Tasks
29
 
30
  Named Entity Recognition
31
-
32
  ## Dataset Usage
33
  ### Using `datasets` library
34
  ```
35
- from datasets import load_dataset
36
- dset = datasets.load_dataset("SEACrowd/idner_news_2k", trust_remote_code=True)
37
  ```
38
  ### Using `seacrowd` library
39
  ```import seacrowd as sc
40
  # Load the dataset using the default config
41
- dset = sc.load_dataset("idner_news_2k", schema="seacrowd")
42
  # Check all available subsets (config names) of the dataset
43
- print(sc.available_config_names("idner_news_2k"))
44
  # Load the dataset using a specific config
45
- dset = sc.load_dataset_by_config_name(config_name="<config_name>")
46
  ```
47
-
48
- More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
49
-
50
 
51
  ## Dataset Homepage
52
 
 
11
  ---
12
 
13
  A dataset of Indonesian News for Named-Entity Recognition task.
14
+ This dataset re-annotated the dataset previously provided by Syaifudin & Nurwidyantoro (2016)
15
+ (https://github.com/yusufsyaifudin/Indonesia-ner) with a more standardized NER tags.
16
+ There are three subsets, namely train.txt, dev.txt, and test.txt.
17
+ Each file consists of three columns which are Tokens, PoS Tag, and NER Tag respectively.
18
+ The format is following CoNLL dataset. The NER tag use the IOB format.
19
+ The PoS tag using UDPipe (http://ufal.mff.cuni.cz/udpipe),
20
+ a pipeline for tokenization, tagging, lemmatization and dependency parsing
21
+ whose model is trained on UD Treebanks.
22
 
23
 
24
  ## Languages
 
28
  ## Supported Tasks
29
 
30
  Named Entity Recognition
31
+
32
  ## Dataset Usage
33
  ### Using `datasets` library
34
  ```
35
+ from datasets import load_dataset
36
+ dset = datasets.load_dataset("SEACrowd/idner_news_2k", trust_remote_code=True)
37
  ```
38
  ### Using `seacrowd` library
39
  ```import seacrowd as sc
40
  # Load the dataset using the default config
41
+ dset = sc.load_dataset("idner_news_2k", schema="seacrowd")
42
  # Check all available subsets (config names) of the dataset
43
+ print(sc.available_config_names("idner_news_2k"))
44
  # Load the dataset using a specific config
45
+ dset = sc.load_dataset_by_config_name(config_name="<config_name>")
46
  ```
47
+
48
+ More details on how to load the `seacrowd` library can be found [here](https://github.com/SEACrowd/seacrowd-datahub?tab=readme-ov-file#how-to-use).
49
+
50
 
51
  ## Dataset Homepage
52