File size: 2,863 Bytes
7151493 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 |
---
license: cc
task_categories:
- summarization
- feature-extraction
language:
- as
- bh
- bn
- en
- gu
- hi
- kn
- ml
- mr
- ne
- or
- pa
- ta
- te
- ur
pretty_name: varta
size_categories:
- 1B<n<10B
---
## Dataset Description
- **Repository:** https://github.com/rahular/varta
- **Paper:** https://arxiv.org/abs/2305.05858
### Dataset Summary
Varta is a diverse, challenging, large-scale, multilingual, and high-quality headline-generation dataset containing 41.8 million news articles in 14 Indic languages and English.
The data is crawled from DailyHunt, a popular news aggregator in India that pulls high-quality articles from multiple trusted and reputed news publishers.
### Languages
Assamese, Bhojpuri, Bengali, English, Gujarati, Hindi, Kannada, Malayalam, Marathi, Nepali, Oriya, Punjabi, Tamil, Telugu, and Urdu.
## Dataset Structure
### Data Fields
- id: unique identifier for the artilce on DailyHunt. This id will be used to recreate the dataset.
- langCode: ISO 639-1 language code
- source_url: the url that points to the article on the website of the original publisher
- dh_url: the url that points to the article on DailyHunt
- id: unique identifier for the artilce on DailyHunt.
- url: the url that points to the article on DailyHunt
- headline: headline of the article
- publication_date: date of publication
- text: main body of the article
- tags: main topics related to the article
- reactions: user likes, dislikes, etc.
- source_media: original publisher name
- source_url: the url that points to the article on the website of the original publisher
- word_count: number of words in the article
- langCode: language of the article
### Data Splits
From every language, we randomly sample 10,000 articles each for validation and testing. We also ensure that at least 80% of a language’s data is available for training.
Therefore, if a language has less than 100,000 articles, we restrict its validation and test splits to 10% of its size.
We also create a `small` training set by limiting the number of articles from each language to 100K.
This `small` training set with a size of 1.3M is used in all our fine-tuning experiments.
You can find the `small` training set [here](https://huggingface.co/datasets/rahular/varta/blob/main/varta/train/train_100k.json)
## Data Recreation
To recreate the dataset, follow this [README file](https://github.com/rahular/varta/tree/main/crawler#README.md).
## Misc
- Original source: https://m.dailyhunt.in/
- License: CC-BY 4.0
## Citation Information
```
@misc{aralikatte2023varta,
title={V\=arta: A Large-Scale Headline-Generation Dataset for Indic Languages},
author={Rahul Aralikatte and Ziling Cheng and Sumanth Doddapaneni and Jackie Chi Kit Cheung},
year={2023},
eprint={2305.05858},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
|