Datasets:
Tasks:
Token Classification
Sub-tasks:
named-entity-recognition
Languages:
English
Size:
1K<n<10K
Tags:
named-entity-linking
License:
annotations_creators: | |
- crowdsourced | |
language_creators: | |
- found | |
language: | |
- en | |
license: | |
- cc-by-4.0 | |
multilinguality: | |
- monolingual | |
size_categories: | |
- 1K<n<10K | |
source_datasets: [] | |
task_categories: | |
- token-classification | |
task_ids: | |
- named-entity-recognition | |
paperswithcode_id: ipm-nel | |
pretty_name: IPM NEL (Derczynski) | |
tags: | |
- named-entity-linking | |
# Dataset Card for "ipm-nel" | |
## Table of Contents | |
- [Dataset Description](#dataset-description) | |
- [Dataset Summary](#dataset-summary) | |
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) | |
- [Languages](#languages) | |
- [Dataset Structure](#dataset-structure) | |
- [Data Instances](#data-instances) | |
- [Data Fields](#data-fields) | |
- [Data Splits](#data-splits) | |
- [Dataset Creation](#dataset-creation) | |
- [Curation Rationale](#curation-rationale) | |
- [Source Data](#source-data) | |
- [Annotations](#annotations) | |
- [Personal and Sensitive Information](#personal-and-sensitive-information) | |
- [Considerations for Using the Data](#considerations-for-using-the-data) | |
- [Social Impact of Dataset](#social-impact-of-dataset) | |
- [Discussion of Biases](#discussion-of-biases) | |
- [Other Known Limitations](#other-known-limitations) | |
- [Additional Information](#additional-information) | |
- [Dataset Curators](#dataset-curators) | |
- [Licensing Information](#licensing-information) | |
- [Citation Information](#citation-information) | |
- [Contributions](#contributions) | |
## Dataset Description | |
- **Homepage:** []() | |
- **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
- **Paper:** [http://www.derczynski.com/papers/ner_single.pdf](http://www.derczynski.com/papers/ner_single.pdf) | |
- **Point of Contact:** [Leon Derczynski](https://github.com/leondz) | |
- **Size of downloaded dataset files:** 120 KB | |
- **Size of the generated dataset:** | |
- **Total amount of disk used:** | |
### Dataset Summary | |
This data is for the task of named entity recognition and linking/disambiguation over tweets. It comprises | |
the addition of an entity URI layer on top of an NER-annotated tweet dataset. The task is to detect entities | |
and then provide a correct link to them in DBpedia, thus disambiguating otherwise ambiguous entity surface | |
forms; for example, this means linking "Paris" to the correct instance of a city named that (e.g. Paris, | |
France vs. Paris, Texas). | |
The data concentrates on ten types of named entities: company, facility, geographic location, movie, musical | |
artist, person, product, sports team, TV show, and other. | |
The file is tab separated, in CoNLL format, with line breaks between tweets. | |
* Data preserves the tokenisation used in the Ritter datasets. | |
* PoS labels are not present for all tweets, but where they could be found in the Ritter data, they're given. | |
* In cases where a URI could not be agreed, or was not present in DBpedia, the linking URI is `NIL`. | |
See the paper, [Analysis of Named Entity Recognition and Linking for Tweets](http://www.derczynski.com/papers/ner_single.pdf) for a full description of the methodology. | |
### Supported Tasks and Leaderboards | |
* Dataset leaderboard on PWC: [Entity Linking on Derczynski](https://paperswithcode.com/sota/entity-linking-on-derczynski-1) | |
### Languages | |
English of unknown region (`bcp47:en`) | |
## Dataset Structure | |
### Data Instances | |
#### ipm_nel | |
- **Size of downloaded dataset files:** 120 KB | |
- **Size of the generated dataset:** | |
- **Total amount of disk used:** | |
An example of 'train' looks as follows. | |
``` | |
{ | |
'id': '0', | |
'tokens': ['#Astros', 'lineup', 'for', 'tonight', '.', 'Keppinger', 'sits', ',', 'Downs', 'plays', '2B', ',', 'CJ', 'bats', '5th', '.', '@alysonfooter', 'http://bit.ly/bHvgCS'], | |
'ner_tags': [9, 0, 0, 0, 0, 7, 0, 0, 7, 0, 0, 0, 7, 0, 0, 0, 0, 0], | |
'uris': "['http://dbpedia.org/resource/Houston_Astros', '', '', '', '', 'http://dbpedia.org/resource/Jeff_Keppinger', '', '', 'http://dbpedia.org/resource/Brodie_Downs', '', '', '', 'NIL', '', '', '', '', '']" | |
} | |
``` | |
### Data Fields | |
- `id`: a `string` feature. | |
- `tokens`: a `list` of `string` features. | |
- `ner_tags`: a `list` of classification labels (`int`). Full tagset with indices: | |
- `uris`: a `list` of URIs (`string`) that disambiguate entities. Set to `NIL` when an entity has no DBpedia entry, or blank for outside-of-entity tokens. | |
### Data Splits | |
| name |train| | |
|---------|----:| | |
|ipm_nel|183 sentences| | |
## Dataset Creation | |
### Curation Rationale | |
To gather a social media benchmark for named entity linking that is sufficiently different from newswire data. | |
### Source Data | |
#### Initial Data Collection and Normalization | |
The data is partly harvested from that distributed by [Ritter / Named Entity Recognition in Tweets: An Experimental Study](https://aclanthology.org/D11-1141/), | |
and partly taken from Twitter by the authors. | |
#### Who are the source language producers? | |
English-speaking Twitter users, between October 2011 and September 2013 | |
### Annotations | |
#### Annotation process | |
The authors were allocated documents and marked them for named entities (where these were not already present) and then attempted to find | |
the best-fitting DBpedia entry for each entity found. Each entity mention was labelled by a random set of three volunteers. | |
The annotation task was mediated using Crowdflower (Biewald, 2012). Our interface design was to show each volunteer the text of the tweet, any URL links contained | |
therein, and a set of candidate targets from DBpedia. The volunteers were encouraged to click on the URL links from the | |
tweet, to gain addition context and thus ensure that the correct DBpedia URI is chosen by them. Candidate entities were | |
shown in random order, using the text from the corresponding DBpedia abstracts (where available) or the actual DBpedia | |
URI otherwise. In addition, the options ‘‘none of the above’’, ‘‘not an entity’’ and ‘‘cannot decide’’ were added, to allow the | |
volunteers to indicate that this entity mention has no corresponding DBpedia URI (none of the above), the highlighted text | |
is not an entity, or that the tweet text (and any links, if available) did not provide sufficient information to reliably disambiguate the entity mention. | |
#### Who are the annotators? | |
The annotators are 10 volunteer NLP researchers, from the authors and the authors' institutions. | |
### Personal and Sensitive Information | |
The data was public at the time of collection. User names are preserved. | |
## Considerations for Using the Data | |
### Social Impact of Dataset | |
There's a risk of user-deleted content being in this data. The data has NOT been vetted for any content, so there's a risk of harmful text. | |
### Discussion of Biases | |
The data is annotated by NLP researchers; we know that this group has high agreement but low recall on English twitter text [C16-1111](https://aclanthology.org/C16-1111/). | |
### Other Known Limitations | |
The above limitations apply. | |
## Additional Information | |
### Dataset Curators | |
The dataset is curated by the paper's authors. | |
### Licensing Information | |
The authors distribute this data under Creative Commons attribution license, CC-BY 4.0. You must | |
acknowledge the author if you use this data, but apart from that, you're quite | |
free to do most things. See https://creativecommons.org/licenses/by/4.0/legalcode . | |
### Citation Information | |
``` | |
@article{derczynski2015analysis, | |
title={Analysis of named entity recognition and linking for tweets}, | |
author={Derczynski, Leon and Maynard, Diana and Rizzo, Giuseppe and Van Erp, Marieke and Gorrell, Genevieve and Troncy, Rapha{\"e}l and Petrak, Johann and Bontcheva, Kalina}, | |
journal={Information Processing \& Management}, | |
volume={51}, | |
number={2}, | |
pages={32--49}, | |
year={2015}, | |
publisher={Elsevier} | |
} | |
``` | |
### Contributions | |
Author-added dataset [@leondz](https://github.com/leondz) | |