Update README.md
Browse files
README.md
CHANGED
@@ -65,7 +65,7 @@ license: mit
|
|
65 |
|
66 |
💻 Code: [https://github.com/sebischair/FusionSent](https://github.com/sebischair/FusionSent)
|
67 |
|
68 |
-
This is a dataset of scientific documents derived from [arXiv metadata](https://www.kaggle.com/datasets/Cornell-University/arxiv). The arXiv metadata provides information about more than 2 million scholarly articles published in arXiv from various scientific fields. We use this metadata to create a dataset of
|
69 |
|
70 |
* The `default` subset contains the dataset with the document categories as classes in the form of lists of strings. The categories are ordered hierarchically according to the [arXiv category taxonomy](https://arxiv.org/category_taxonomy). In this dataset, the `->` symbols indicate a `parent->child` relationship between categories that can be linked and create a path from the root to the leaf node. For classification, you can either use the complete paths as classes or just parse the respective leaf nodes as classes, resulting in the same (abbreviated) categories.
|
71 |
* The `arxiv_category_descriptions` subset contains the tags, names, and textual descriptions of the leaf nodes from the [arXiv category taxonomy](https://arxiv.org/category_taxonomy).
|
|
|
65 |
|
66 |
💻 Code: [https://github.com/sebischair/FusionSent](https://github.com/sebischair/FusionSent)
|
67 |
|
68 |
+
This is a dataset of scientific documents derived from [arXiv metadata](https://www.kaggle.com/datasets/Cornell-University/arxiv). The arXiv metadata provides information about more than 2 million scholarly articles published in arXiv from various scientific fields. We use this metadata to create a dataset of 203,961 titles and abstracts categorized into 130 different classes. To this end, we first perform stratified downsampling of the metadata to only 10% of all articles while retaining the original class distribution. Afterward, articles assigned to categories occurring less than 100 times in the downsampled dataset are removed. To obtain the final dataset, we then perform a stratified train/validation/test split of the processed dataset in an 80:10:10 ratio. The number of examples in each set is shown in the table below.
|
69 |
|
70 |
* The `default` subset contains the dataset with the document categories as classes in the form of lists of strings. The categories are ordered hierarchically according to the [arXiv category taxonomy](https://arxiv.org/category_taxonomy). In this dataset, the `->` symbols indicate a `parent->child` relationship between categories that can be linked and create a path from the root to the leaf node. For classification, you can either use the complete paths as classes or just parse the respective leaf nodes as classes, resulting in the same (abbreviated) categories.
|
71 |
* The `arxiv_category_descriptions` subset contains the tags, names, and textual descriptions of the leaf nodes from the [arXiv category taxonomy](https://arxiv.org/category_taxonomy).
|