|
--- |
|
annotations_creators: |
|
- expert-generated |
|
language_creators: |
|
- machine-generated |
|
language: |
|
- code |
|
license: |
|
- unknown |
|
multilinguality: |
|
- monolingual |
|
size_categories: |
|
- unknown |
|
source_datasets: |
|
- original |
|
task_categories: |
|
- structure-prediction |
|
task_ids: [] |
|
pretty_name: Binkley |
|
tags: |
|
- word-segmentation |
|
--- |
|
|
|
# Dataset Card for Binkley |
|
|
|
## Dataset Description |
|
|
|
- **Paper:** [Normalizing Source Code Vocabulary](https://www.researchgate.net/publication/224198190_Normalizing_Source_Code_Vocabulary) |
|
|
|
### Dataset Summary |
|
|
|
In programming languages, identifiers are tokens (also called symbols) which name language entities. |
|
Some of the kinds of entities an identifier might denote include variables, types, labels, subroutines, and packages. |
|
|
|
Binkley is a dataset for identifier segmentation, i.e. the task of adding spaces between the words on a identifier. |
|
|
|
### Languages |
|
|
|
- C |
|
- C++ |
|
- Java |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
|
|
``` |
|
{ |
|
"index": 0, |
|
"identifier": "init_g16_i", |
|
"segmentation": "init _ g 16 _ i" |
|
} |
|
``` |
|
|
|
### Data Fields |
|
|
|
- `index`: a numerical index. |
|
- `identifier`: the original identifier. |
|
- `segmentation`: the gold segmentation for the identifier. |
|
|
|
## Dataset Creation |
|
|
|
- All hashtag segmentation and identifier splitting datasets on this profile have the same basic fields: `hashtag` and `segmentation` or `identifier` and `segmentation`. |
|
|
|
- The only difference between `hashtag` and `segmentation` or between `identifier` and `segmentation` are the whitespace characters. Spell checking, expanding abbreviations or correcting characters to uppercase go into other fields. |
|
|
|
- There is always whitespace between an alphanumeric character and a sequence of any special characters ( such as `_` , `:`, `~` ). |
|
|
|
- If there are any annotations for named entity recognition and other token classification tasks, they are given in a `spans` field. |
|
|
|
## Additional Information |
|
|
|
### Citation Information |
|
|
|
``` |
|
@inproceedings{inproceedings, |
|
author = {Lawrie, Dawn and Binkley, David and Morrell, Christopher}, |
|
year = {2010}, |
|
month = {11}, |
|
pages = {3 - 12}, |
|
title = {Normalizing Source Code Vocabulary}, |
|
journal = {Proceedings - Working Conference on Reverse Engineering, WCRE}, |
|
doi = {10.1109/WCRE.2010.10} |
|
} |
|
``` |
|
|
|
### Contributions |
|
|
|
This dataset was added by [@ruanchaves](https://github.com/ruanchaves) while developing the [hashformers](https://github.com/ruanchaves/hashformers) library. |