--- dataset_info: features: - name: post dtype: string - name: newsgroup dtype: string - name: embedding sequence: float64 - name: map sequence: float64 splits: - name: train num_bytes: 129296327 num_examples: 18170 download_size: 102808058 dataset_size: 129296327 configs: - config_name: default data_files: - split: train path: data/train-* --- # Dataset Card for 20-Newsgroups Embedded This provides a subset of 20-Newsgroup posts, along with sentence embeddings, and a dimension reduced 2D data map. This provides a basic setup for experimentation with various neural topic modelling approaches. ## Dataset Details ### Dataset Description This is a dataset containing posts from the classic 20-Newsgroups dataset, along with sentence embeddings, and a dimension reduced 2D data map. Per the [source](http://qwone.com/~jason/20Newsgroups/): > The 20-Newsgroups dataset is a collection of approximately 20,000 newsgroup documents, partitioned (nearly) evenly across 20 different newsgroups. > As far as in known it was originally collected by Ken Lang, probably for his Newsweeder: Learning to filter netnews paper. > The 20 newsgroups collection has become a popular data set for experiments in text applications of machine learning techniques, > such as text classification and text clustering. This has then been enriched with sentence embeddings via sentence-transformers using the `all-mpnet-base-v2` model. Further enrichment is provided in the form of a 2D representation of the sentence embeddings generated using UMAP. - **Curated by:** Leland McInnes - **Language(s) (NLP):** English - **License:** Public Domain ### Dataset Sources The post and newsgroup data was collected using the `sckit-learn` function `fetch_20newsgroups` and then processed to exclude very short and excessively long posts in the following manner: ```python newsgroups = sklearn.datasets.fetch_20newsgroups(subset="all", remove=("headers", "footers", "quotes")) useable_content = np.asarray([len(x) > 8 and len(x) < 16384 for x in newsgroups.data]) documents = [ doc for doc, worth_keeping in zip(newsgroups.data, useable_content) if worth_keeping ] newsgroup_categories = [ newsgroups.target_names[newsgroup_id] for newsgroup_id, worth_keeping in zip(newsgroups.target, useable_content) if worth_keeping ] ``` - **Repository:** The original source datasets can be found at [http://qwone.com/~jason/20Newsgroups/](http://qwone.com/~jason/20Newsgroups/) ## Uses This datasets is intended to be used for simple experiments and demonstrations of topic modelling and related tasks. #### Personal and Sensitive Information This data may contain personal information that was posted publicly to NNTP servers in the mid 1990's. It is not believed to contain any senstive information. ## Bias, Risks, and Limitations This dataset is a product of public discussion forums in the 1990s. As such it contains debate, potentially inflammatory and/or derogatory language, etc. It does not provide a representative sampling of opinion from the era. This data should only be used for experiments or demonstration and educational purposes.