File size: 36,696 Bytes
6fa4bc9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
{
    "paper_id": "O12-1016",
    "header": {
        "generated_with": "S2ORC 1.0.0",
        "date_generated": "2023-01-19T08:02:46.464674Z"
    },
    "title": "English-to-Traditional Chinese Cross-lingual Link Discovery in Articles with Wikipedia Corpus",
    "authors": [
        {
            "first": "Liang-Pu",
            "middle": [],
            "last": "Chen",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Yu-Lun",
            "middle": [],
            "last": "Shih",
            "suffix": "",
            "affiliation": {
                "laboratory": "",
                "institution": "National Taipei Univeristy of Technology",
                "location": {
                    "country": "Taiwan"
                }
            },
            "email": ""
        },
        {
            "first": "Chien-Ting",
            "middle": [],
            "last": "Chen",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Tsun",
            "middle": [],
            "last": "Ku",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Wen-Tai",
            "middle": [],
            "last": "Hsieh",
            "suffix": "",
            "affiliation": {},
            "email": ""
        },
        {
            "first": "Hung-Sheng",
            "middle": [],
            "last": "Chiu",
            "suffix": "",
            "affiliation": {},
            "email": ""
        }
    ],
    "year": "",
    "venue": null,
    "identifiers": {},
    "abstract": "In this paper, we design a processing flow to produce linked data in articles, providing anchorbased term's additional information and related terms in different languages (English to Chinese). Wikipedia has been a very important corpus and knowledge bank. Although Wikipedia describes itself not a dictionary or encyclopedia, it is if high potential values in applications and data mining researches. Link discovery is a useful IR application, based on Data Mining and NLP algorithms and has been used in several fields. According to the results of our experiment, this method does make the result has improved.",
    "pdf_parse": {
        "paper_id": "O12-1016",
        "_pdf_hash": "",
        "abstract": [
            {
                "text": "In this paper, we design a processing flow to produce linked data in articles, providing anchorbased term's additional information and related terms in different languages (English to Chinese). Wikipedia has been a very important corpus and knowledge bank. Although Wikipedia describes itself not a dictionary or encyclopedia, it is if high potential values in applications and data mining researches. Link discovery is a useful IR application, based on Data Mining and NLP algorithms and has been used in several fields. According to the results of our experiment, this method does make the result has improved.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Abstract",
                "sec_num": null
            }
        ],
        "body_text": [
            {
                "text": "For our goal, we have to conquer some issues to find every potential linked data on articles. This paper focuses on Cross-lingual link discovery. Cross-lingual link discovery contains a lot of important tasks of NLP(Natural Language Processing) such as WSD(Word Sense Disambiguation) [1] , NED(Named Entities Disambiguation) [2] or Machine Translation. The cross-lingual links in the Wikipedia 1 are established by the human contributors, and not all Wikipedia Pages have cross lingual links because no human editors established these links yet. Thus, when one visits English Wikipedia page which describes some special information, users cannot find any cross lingual link to visit the Wikipedia page whose language is the same as the user's mother tongue. This problem has been raised by many recent studies [3, 4] , and recovering these missing links between two languages is the main goal of the CLLD (Cross-Lingual Link Discovery). In this paper, we propose a system which can automatically help users to tag potential links in their articles, and automatically find out the cross language link of the tag based on Wikipedia cross language links. As for cross lingual link discovery, our system is able to find the missing links between two related Wikipedia pages in two different language systems by exploiting and extracting data from Wikipedia dump files in two languages. In addition, we use two additional translation mechanisms to help find out the corresponding cross lingual translation , one is the Pattern Translate , the other one is Google Translate 2 . We further integrate the Lucene 3 software package to deal with the ambiguous phases in the articles. In order to find out the missing links between two pages, and automatically tagged this cross language link in users' articles.",
                "cite_spans": [
                    {
                        "start": 284,
                        "end": 287,
                        "text": "[1]",
                        "ref_id": "BIBREF0"
                    },
                    {
                        "start": 325,
                        "end": 328,
                        "text": "[2]",
                        "ref_id": "BIBREF1"
                    },
                    {
                        "start": 810,
                        "end": 813,
                        "text": "[3,",
                        "ref_id": "BIBREF2"
                    },
                    {
                        "start": 814,
                        "end": 816,
                        "text": "4]",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "The remainder of this paper is organized as follows: First, we described corresponding background of Wikipedia and cross-lingual link discovery in Section 2. In Section 3,The proposed WSD method and translation mechanism will be described in detail. Finally, the experiment and conclusion will be discussed in Section 4.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Introduction",
                "sec_num": "1"
            },
            {
                "text": "Wikipedia is a free, collaboratively edited, and multilingual Internet encyclopedia supported by the non-profit Wikimedia Foundation 4 . Recently, many researchers focus on developing data mining applications with Wikipedia's large-scale collaborative user data. Although Wikipedia describes itself not a dictionary, textbook or encyclopedia, exploiting its characteristics to develop new services is regarded as a promising method on auto text explanation.",
                "cite_spans": [
                    {
                        "start": 133,
                        "end": 134,
                        "text": "4",
                        "ref_id": "BIBREF3"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Background 2.1 Wikipedia",
                "sec_num": "2"
            },
            {
                "text": "One of the special feature of Wikipedia is that it contains many hypertext links to help users easily retrieve the information they need. These hypertext links might be embedded within the text content under the corresponding pages, and each of these links is linking to other pages related with different terms. Obviously, information flow is thus being traversed very easy and smoothing when the hypertext links are extensively tagged. Unfortunately, the hypertext links between different languages are mostly not being tagged because of the hypertext link is generated by human contributor, mostly monolingual ones. To solve this problem, we design a process flow trying to make it more completely.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Background 2.1 Wikipedia",
                "sec_num": "2"
            },
            {
                "text": "The goal of cross-lingual link discovery(CLLD) is trying to find the potential links that are missing between the two different languages. There are three main challenges for the system to overcome. First, the system providing solution on CLLD can proactively recommends a set of words which called anchors. The set of words have higher chances to have their corresponding cross lingual links than other words in the same article. For example, considering different cases as following:",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "1. Let's go dutch. 2.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "A Dutch auction is a type of auction that starts with a high bid.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "The system must determine the boundaries between anchor and rest of words, considering the first case above, the word \"dutch\" is meaning to share the money on something instead of meaning some behavior or something related to the country \"Holland\". In other words, the word \"dutch\" should not be chosen as an anchor here and choosing the phase of \"go dutch\" is more significant. Considering the second case above, the word \"Dutch auction\" is an appropriate anchor rather than \"Dutch\".",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "After the system identifies these anchors, there must exist many highly ambiguous cases in these anchors and this is the second challenge of CLLD, for example, the anchor Apple can be refer to the link which is related with Apple(Computer Manufacturer), or the link which is related to Apple(Fruit). The system must be able to choosing the most related corresponding links and also ensure the correctness of link discovery.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "Once the system can return the most related links of each anchor, there is only one more problem need to solve. In the end of the CLLD flow, the system have to automatically discover the cross-lingual link based on the anchors which generated from previous steps. The system can just use simple parser or crawler to check the content of corresponding wikipedia page or combines some different mechanism to increase the accuracy of link discovery. In this paper, we implement these CLLD steps to help us find the corresponding cross-lingual links and we focus on anchor disambiguation and cross-lingual link discovery, which are both described in Section 3.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-lingual link discovery",
                "sec_num": "2.2"
            },
            {
                "text": "In English-to-Chinese cross-lingual link discovery, the goal is to find every potential links in documents. At first, the system searches out potential terms as candidate terms. Overlapping problem happens in this stage, and adequate candidate term selection is required. We propose an similarity-scoring formula to calculate score of relevance. When candidate terms are selected, relevant pages in Chinese Wikipedia need to be linked with these terms. There are some cross-lingual articles in Wikipedia; however, many more links are still missed. (eg. \"Hundred Schools of Thought\" with \"\u8af8\u5b50\u767e\u5bb6\").",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Method and System Description",
                "sec_num": "3"
            },
            {
                "text": "To find cross-lingual links in a language, every potential term or phrase is to be listed in the beginning. Here we adopt n-gram tokenizer [5] and Maximum Matching algorithm [6] to segment. For example, assume a sentence \"Magic Johnson is one of the best basketball player in NBA\", in our method , our approach will take \"Magic Johnson\" as an anchor rather than \"Magic\" or \"John\". The system will first examine the longer term in the sentence and exploit the Wikipedia as a anchor look-up table to check whether this long term is meaningful or not.",
                "cite_spans": [
                    {
                        "start": 139,
                        "end": 142,
                        "text": "[5]",
                        "ref_id": "BIBREF4"
                    },
                    {
                        "start": 174,
                        "end": 177,
                        "text": "[6]",
                        "ref_id": "BIBREF5"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Candidates finding",
                "sec_num": "3.1"
            },
            {
                "text": "Many terms in Wikipedia have the same title but different meanings depending on their occurrences in contexts. To address this problem, Wikipedia has already define it as \"Disambiguation\". In our system, we use redirect page, providing disambiguation information and candidate terms, to analysis and select one from terms for users by this system. For instance, a term \"Virus\" is shown in \"A virus is a parasitic agent that is smaller than a bacterium and that can only reproduce after infecting a host cell.\" and \"Virus (clothing), an Israeli clothing brand\"...etc. It indicates users may look out the clothing brand but Wikipedia gives him a virus' definition in biology domain. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Anchor decision",
                "sec_num": "3.2"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "SimilarityScore(D i , D j ) = T ermRecog(D i ) T ermRecog(D j ) T ermRecog(D i ) T ermRecog(D j )",
                        "eq_num": "(1)"
                    }
                ],
                "section": "Anchor decision",
                "sec_num": "3.2"
            },
            {
                "text": "EQUATION",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [
                    {
                        "start": 0,
                        "end": 8,
                        "text": "EQUATION",
                        "ref_id": "EQREF",
                        "raw_str": "Anchor = max(SimilarityScore(D current , D i )), \u2200i \u2208 candidates",
                        "eq_num": "(2)"
                    }
                ],
                "section": "Anchor decision",
                "sec_num": "3.2"
            },
            {
                "text": "In our work, we design a content-aware approach to perform auto selection among disambiguation terms. Our design principle is to analyze the input article, especially the source of terms, and use full-featured text search engine with a prepared index file. If a term has the disambiguation property, the system will extract the features from article and search the existed index to decide which term is more likely to the source article.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Anchor decision",
                "sec_num": "3.2"
            },
            {
                "text": "In this section, we describe how we translate the anchor first and than how we find the cross lingual Wikipedia link after the translation. There are two main approaches of the translation mechanism, namely Cross-Lingual Link Dictionary and Google Translate. We first use a Cross-Lingual Link Dictionary as the translation scheme, once if Cross-Lingual Link Dictionary can not provide any corresponding translation, Google Translate is then used by the system to discover the corresponding translation from the online Machine Translation mechanism. Google Translate is a state-of-the-art online commercial machine translation scheme, and it is exploited by our system to trying find out some possible translation when there doesn't have any corresponding translation which can be provided by the Cross-Lingual Link Dictionary. With the support by the Google Translate, the system can provide higher translation coverage compared to using Cross-Lingual Link Dictionary only. We will describe the detail about the two translation mechanisms below and will also discuss the missing link recovery approach in the end of this section. ",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "English-Chinese Link Discovery",
                "sec_num": "3.3"
            },
            {
                "text": "We first describe the Google Translate here because we are going to introduce the translation and missing link recovery within Cross-Lingual Dictionary in the end of this section together.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Google Translate",
                "sec_num": "3.4.1"
            },
            {
                "text": "Google Translate has been a famous automatic translation mechanism, one distinguishing feature of this online translator is that it enables users to choose different languages that users want to translate. As for whole sentence translations, the users have a chance to modify the translation sentence once they find the output translation inadequate. As Google collects enough user data of modifying the translation sentence, Google Translator gets higher translation accuracy.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Google Translate",
                "sec_num": "3.4.1"
            },
            {
                "text": "Although Google Translate has such special characteristic, it can not providing good accuracy at Anchor translation [7] . However, there is a special characteristic of Google Translate; that is, it can provide more possible translation candidates than previous methods such like Cross-Lingual Link Dictionary. The reason is that Google Translate is tends to adopt a best-effort approach, it aims to provide many translation candidates which enable users to understand what the untranslated sentence might be supposed to mean.",
                "cite_spans": [
                    {
                        "start": 116,
                        "end": 119,
                        "text": "[7]",
                        "ref_id": "BIBREF6"
                    }
                ],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Google Translate",
                "sec_num": "3.4.1"
            },
            {
                "text": "As a result, we put the lowest translation priority in Google Translate, namely, once the previous method(Cross-Lingual Dictionary) can not find out any possible translation candidates, we will try to get some translation suggested from Google Translate. The main reason is just what we describe above, we want to take a chance to find out the corresponding translation when we do not have any other translation candidate, only to use some anchor translation from Google Translate to find out the corresponding cross-language links.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Google Translate",
                "sec_num": "3.4.1"
            },
            {
                "text": "For example, in our Cross-Lingual Link Dictionary, it does not contain the Chinese Translation of \"Freeway\". However, Google Translate can provide some useful Chinese translation like \"\u9ad8 \u9ad8 \u9ad8\u901f \u901f \u901f\u516c \u516c \u516c\u8def \u8def \u8def\", thus we can find the corresponding link of Chinese article page of Wikipedia page at \"http://zh.wikipedia.org/wiki/\".",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Google Translate",
                "sec_num": "3.4.1"
            },
            {
                "text": "Wikipedia provides a well formatted dump file for all languages. As a result, we can get the chinese translation from the english dump files and vise-versa. We exploit this property to construct both Chinese-English bilingual link dictionary and an English-Chinese bilingual link dictionary. Furthermore, once the translation in the dictionary has be found, there is a high probability that we can directly discover the link by adding the translated anchor after the specific wikipedia URL(e.g. http://en.wikipedia.org/wiki/Computer_ accessibility), both in English and Chinese. We refer these two dictionaries as the translation dictionaries, one is the English to Chinese (E-C) translation dictionary and the other one is Chinese to English (C-E) translation dictionary. Once we use these two bilingual dictionaries as translation dictionaries, in our case, English-to-Chinese vise versa,we can have a chance to retrieve the link informations bidirectional. The reason is that we have noticed that links for Chinese-to-English are more than English-to-Chinese, because many Chinese editors will add English link for annotation or reference.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-Lingual Link Dictionary",
                "sec_num": "3.4.2"
            },
            {
                "text": "On link discovery part, we find out that some links may be missing in one translation dictionary, such as the term \"Flag of Republic of China\" is not able to found any corresponding Chinese translation in E-C translation dictionary. However, we can find the corresponding english translation of chinese term \"\u8af8\u5b50\u767e\u5bb6\" in the C-E translation dictionary, which is the \"Hundred Schools of Thought\".",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-Lingual Link Dictionary",
                "sec_num": "3.4.2"
            },
            {
                "text": "There is an additional problem about the English-Chinese dictionary with the Wikipedia disambiguation page. If the anchor which exist in the English-Chinese dictionary is a title of the Wikipedia disambiguation page, then we can not directly get the Chinese translation from the page content of disambiguation page. The reason is that a Wikipedia disambiguation page only contains the possible candidates that are referring to this title.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-Lingual Link Dictionary",
                "sec_num": "3.4.2"
            },
            {
                "text": "Fortunately, Wikipedia have a complete dump file format and it provide the redirect information of the disambiguation page. Therefore, we can using the redirect link information to find out the corresponding Chinese translation. The problem may also occur at Chinese Wikipedia disambiguation page, and it can be also solved by redirection information.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Cross-Lingual Link Dictionary",
                "sec_num": "3.4.2"
            },
            {
                "text": "We use four articles as evaluation to see the performance of cross-lingual discovery, we first randomly choose four Bilingual news article from Yahoo! News, all terms in the Chinese articles are tagged by two human experts to generate correct answers. We apply two methods, the first method is tagging the English articles with English Wikipedia entries by means of longterm-first algorithm. Those tagged terms are then directly transformed into Chinese Wikipedia entries by original anchored links; the second method is to implement our proposed method, we then compare these two methods to see the coverage rates. As Figure 4 shows, the experiment result shows that our proposed method has 8% coverage rates higher than the that of direct anchor transformation method. ",
                "cite_spans": [],
                "ref_spans": [
                    {
                        "start": 619,
                        "end": 627,
                        "text": "Figure 4",
                        "ref_id": null
                    }
                ],
                "eq_spans": [],
                "section": "Results and Discussion",
                "sec_num": "4"
            },
            {
                "text": "In conclusion, we present a system to find potential cross-lingual linked data on articles, trying to discover miss cross-lingual links. The main contribution of our proposed method includes finding anchor and discovering missing cross-lingual links. We have successfully designed a practical system to perform tagging task on real-world articles. and proved that maximum match algorithm has a better performance than the original Wikipedia anchor links transformation. However, there are still issued to be improved for future work. First, the precision of WSD is still low, and second, we can apply machine learning approaches in our method, in which we are convinced that our proposed method might have higher performance in the future.",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "Conclusion",
                "sec_num": "5"
            },
            {
                "text": "http://wikipedia.org Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "http://translate.google.com 3 http://lucene.apache.org 4 http://en.wikipedia.org/wiki/Wikipedia Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing (ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            },
            {
                "text": "Proceedings of the Twenty-Fourth Conference on Computational Linguistics and Speech Processing(ROCLING 2012)",
                "cite_spans": [],
                "ref_spans": [],
                "eq_spans": [],
                "section": "",
                "sec_num": null
            }
        ],
        "back_matter": [],
        "bib_entries": {
            "BIBREF0": {
                "ref_id": "b0",
                "title": "Wikify! linking documents to encyclopedic knowledge",
                "authors": [
                    {
                        "first": "Csomai",
                        "middle": [],
                        "last": "Mihalcea",
                        "suffix": ""
                    }
                ],
                "year": 2007,
                "venue": "sixteenth ACM conference",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Mihalcea and Csomai, \"Wikify! linking documents to encyclopedic knowledge\", in six- teenth ACM conference, 2007.",
                "links": null
            },
            "BIBREF1": {
                "ref_id": "b1",
                "title": "Using encyclopedic knowledge for named entity disambiguation",
                "authors": [
                    {
                        "first": "Pasca",
                        "middle": [],
                        "last": "Bunescu",
                        "suffix": ""
                    }
                ],
                "year": 2006,
                "venue": "EACL",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Bunescu and Pasca, \"Using encyclopedic knowledge for named entity disambiguation\", in EACL, 2006.",
                "links": null
            },
            "BIBREF2": {
                "ref_id": "b2",
                "title": "Ukp at crosslink: Anchor text translation for cross-lingual link discovery",
                "authors": [
                    {
                        "first": "J",
                        "middle": [],
                        "last": "Kim",
                        "suffix": ""
                    },
                    {
                        "first": "I",
                        "middle": [],
                        "last": "Gurevych",
                        "suffix": ""
                    }
                ],
                "year": 2011,
                "venue": "NTCIR-9",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "J. Kim and I. Gurevych, \"Ukp at crosslink: Anchor text translation for cross-lingual link discovery\", in NTCIR-9, 2011.",
                "links": null
            },
            "BIBREF3": {
                "ref_id": "b3",
                "title": "Hits'graph-based system at the ntcir-9 cross-lingual link discovery task",
                "authors": [
                    {
                        "first": "A",
                        "middle": [
                            "F V"
                        ],
                        "last": "Nastase",
                        "suffix": ""
                    },
                    {
                        "first": "M",
                        "middle": [],
                        "last": "Strube",
                        "suffix": ""
                    }
                ],
                "year": 2011,
                "venue": "NTCIR-9",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "A.F.V. Nastase and M. Strube, \"Hits'graph-based system at the ntcir-9 cross-lingual link discovery task\", in NTCIR-9, 2011.",
                "links": null
            },
            "BIBREF4": {
                "ref_id": "b4",
                "title": "N-gram based text categorization",
                "authors": [
                    {
                        "first": "W",
                        "middle": [
                            "B"
                        ],
                        "last": "Cavnar",
                        "suffix": ""
                    },
                    {
                        "first": "J",
                        "middle": [
                            "M"
                        ],
                        "last": "Trenkle",
                        "suffix": ""
                    }
                ],
                "year": 1994,
                "venue": "Proceeding of the Symposium on Document Analysis and Information Retrieval",
                "volume": "",
                "issue": "",
                "pages": "161--175",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "W. B. Cavnar and J. M. Trenkle., \"N-gram based text categorization\", in Proceeding of the Symposium on Document Analysis and Information Retrieval. University of Nevada, Las Vegas, 1994, pp. 161-175.",
                "links": null
            },
            "BIBREF5": {
                "ref_id": "b5",
                "title": "An improved parallel algorithm for maximal matching",
                "authors": [
                    {
                        "first": "Y. Shiloach Amos",
                        "middle": [],
                        "last": "Israeli",
                        "suffix": ""
                    }
                ],
                "year": 1986,
                "venue": "Information Processing Letters",
                "volume": "22",
                "issue": "",
                "pages": "57--60",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Y. Shiloach Amos Israeli, \"An improved parallel algorithm for maximal matching\", in Information Processing Letters, 1986, vol. 22, pp. 57-60.",
                "links": null
            },
            "BIBREF6": {
                "ref_id": "b6",
                "title": "Exploit wikipedia and name-entity pattern as translation method on chinene-korean cross language multimedia information retrival",
                "authors": [
                    {
                        "first": "Yu-Chun Wang Richard Tzong-Han Tsai Liang-Pu",
                        "middle": [],
                        "last": "Chen",
                        "suffix": ""
                    },
                    {
                        "first": "Chen-Ting",
                        "middle": [],
                        "last": "Chen",
                        "suffix": ""
                    }
                ],
                "year": 2009,
                "venue": "International Conference on Digital Contents",
                "volume": "",
                "issue": "",
                "pages": "",
                "other_ids": {},
                "num": null,
                "urls": [],
                "raw_text": "Yu-Chun Wang Richard Tzong-Han Tsai Liang-Pu Chen, Chen-Ting Chen, \"Exploit wikipedia and name-entity pattern as translation method on chinene-korean cross language multimedia information retrival\", in International Conference on Digital Contents, 2009.",
                "links": null
            }
        },
        "ref_entries": {
            "FIGREF0": {
                "text": "Processing flow of our system.",
                "num": null,
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF1": {
                "text": "Flow of our English-Chinese Link Discovery system.",
                "num": null,
                "uris": null,
                "type_str": "figure"
            },
            "FIGREF2": {
                "text": "Results of English to Chinese link discovery.",
                "num": null,
                "uris": null,
                "type_str": "figure"
            }
        }
    }
}