Divyanshusingh01 commited on
Commit
5ca5f3b
1 Parent(s): 6c79838

Update README.md

Browse files

Dataset Name -Hindi- niBand (Massive Hindi Text Dataset )

Dataset Overview :-

This dataset is a comprehensive collection of text data consisting of more than 10 billion tokens. It encompasses a wide range of sources, including Wikipedia articles, news articles, email transcripts, and generated prompt text. Specific Hindi language data columns have been extracted from the CulturaX dataset, which is a large, cleaned, and multilingual dataset for large language models. We acknowledge and cite the CulturaX dataset using the provided citation.

Data Sources :-

1. Wikipedia Articles: A large corpus of text extracted from Wikipedia articles covering various topics and domains.
2. News Articles: Textual data sourced from news articles from diverse sources and regions.
3. Email Transcripts: Transcripts of email communications, providing insights into natural language usage in electronic correspondence.
4. Prompt Text Generation: Text generated from prompts or prompts used to generate text, contributing to the dataset's diversity and complexity.
5. Hindi Data from CulturaX Dataset: Specific Hindi language data columns have been extracted from the CulturaX dataset, which is a large, cleaned, and multilingual dataset for large language models.

Potential Uses :-

- Training and evaluating natural language generation models in the Hindi language domain.
- Exploring the capabilities of models in narrative generation tasks.
- Conducting research on narrative understanding and generation in Hindi.

#### Importance for Indian Native Languages

This dataset can be crucial for the training of LLM (Large Language Model) models and aiming to explore the capabilities of those natural language generation models in Hindi. It serves as a foundation for training and evaluating models capable of producing coherent and contextually relevant narratives or explanations. Additionally, this dataset aligns with our commitment to promoting Indian native languages on a global scale. We recognize the limited availability of such datasets as a major challenge for innovation within the local Indian community. As part of our contribution to the Indian open-source community, we are planning to release a very large database covering various Indian native languages. This initiative aims to empower researchers, practitioners, and developers to explore and innovate in Indian language processing and generation tasks.

#### Citation :-

We acknowledge and cite the CulturaX dataset using the following citation:

```


@misc

{nguyen2023culturax,
title={CulturaX: A Cleaned, Enormous, and Multilingual Dataset for Large Language Models in 167 Languages},
author={Thuat Nguyen and Chien Van Nguyen and Viet Dac Lai and Hieu Man and Nghia Trung Ngo and Franck Dernoncourt and Ryan A. Rossi and Thien Huu Nguyen},
year={2023},
eprint={2309.09400},
archivePrefix={arXiv},
primaryClass={cs.CL}.
```



Additionally, the dataset includes news article data, and we acknowledge and cite the source of this data using the following citations:

@inproceedings{see-etal-2017-get,
title = "Get To The Point: Summarization with Pointer-Generator Networks",
author = "See, Abigail and
Liu, Peter J. and
Manning, Christopher D.",
booktitle = "Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2017",
address = "Vancouver, Canada",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/P17-1099",
doi = "10.18653/v1/P17-1099",
pages = "1073--1083",
abstract = "Neural sequence-to-sequence models have provided a viable new approach for abstractive text summarization (meaning they are not restricted to simply selecting and rearranging passages from the original text). However, these models have two shortcomings: they are liable to reproduce factual details inaccurately, and they tend to repeat themselves. In this work we propose a novel architecture that augments the standard sequence-to-sequence attentional model in two orthogonal ways. First, we use a hybrid pointer-generator network that can copy words from the source text via pointing, which aids accurate reproduction of information, while retaining the ability to produce novel words through the generator. Second, we use coverage to keep track of what has been summarized, which discourages repetition. We apply our model to the CNN / Daily Mail summarization task, outperforming the current abstractive state-of-the-art by at least 2 ROUGE points.",
}

@inproceedings{DBLP:conf/nips/HermannKGEKSB15,
author={Karl Moritz Hermann and Tomás Kociský and Edward Grefenstette and Lasse Espeholt and Will Kay and Mustafa Suleyman and Phil Blunsom},
title={Teaching Machines to Read and Comprehend},
year={2015},
cdate={1420070400000},
pages={1693-1701},
url={http://papers.nips.cc/paper/5945-teaching-machines-to-read-and-comprehend},
booktitle={NIPS},
crossref={conf/nips/2015}
}

If you use this dataset in your research or applications, please consider citing the CulturaX dataset using the provided citation.

#### License

Please refer to the licensing terms specified by the dataset creators.

#### Disclaimer

The views expressed in the dataset do not necessarily reflect the views of the dataset creators or contributors. Users are advised to use the data responsibly and in accordance with ethical guidelines.

This dataset card provides an overview of the massive multilingual text dataset, highlighting its sources, potential uses, citation, and disclaimer.

Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -1,3 +1,11 @@
1
  ---
2
  license: mit
3
- ---
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ task_categories:
4
+ - text-classification
5
+ - table-question-answering
6
+ - summarization
7
+ - text-generation
8
+ - text2text-generation
9
+ size_categories:
10
+ - 100M<n<1B
11
+ ---