--- license: apache-2.0 language: - en pretty_name: HackerNews comments dataset dataset_info: config_name: default features: - name: id dtype: int64 - name: deleted dtype: bool - name: type dtype: string - name: by dtype: string - name: time dtype: int64 - name: text dtype: string - name: dead dtype: bool - name: parent dtype: int64 - name: poll dtype: int64 - name: kids sequence: int64 - name: url dtype: string - name: score dtype: int64 - name: title dtype: string - name: parts sequence: int64 - name: descendants dtype: int64 configs: - config_name: default data_files: - split: train path: items/*.jsonl.zst --- # Hackernews Comments Dataset A dataset of all [HN API](https://github.com/HackerNews/API) items from `id=0` till `id=41422887` (so from 2006 till 02 Sep 2024). The dataset is build by scraping the HN API according to its official [schema and docs](https://github.com/HackerNews/API). Scraper code is also available on github: [nixiesearch/hnscrape](https://github.com/nixiesearch/hnscrape) ## Dataset contents No cleaning, validation or filtering was performed. The resulting data files are raw JSON API response dumps in zstd-compressed JSONL files. An example payload: ```json { "by": "goldfish", "descendants": 0, "id": 46, "score": 4, "time": 1160581168, "title": "Rentometer: Check How Your Rent Compares to Others in Your Area", "type": "story", "url": "http://www.rentometer.com/" } ``` ## Usage You can directly load this dataset with a [Huggingface Datasets](https://github.com/huggingface/datasets/) library. ```shell pip install datasets zstandard ``` ```python from datasets import load_dataset ds = load_dataset("nixiesearch/hackernews-comments", split="train") print(ds.features) ``` ## License Apache License 2.0.