Datasets:
Tasks:
Text Classification
Languages:
English
Size:
10K<n<100K
ArXiv:
Tags:
hate-speech-detection
License:
Commit
•
f6a8b7d
1
Parent(s):
5a137fd
Delete legacy JSON metadata (#3)
Browse files- Delete legacy JSON metadata (6d880b841fa9e3e0f6034eb9872f7bde8e2d5cf3)
- dataset_infos.json +0 -1
dataset_infos.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"plain_text": {"description": "Hatexplain is the first benchmark hate speech dataset covering multiple aspects of the issue. Each post in the dataset is annotated from three different perspectives: the basic, commonly used 3-class classification (i.e., hate, offensive or normal), the target community (i.e., the community that has been the victim of hate speech/offensive speech in the post), and the rationales, i.e., the portions of the post on which their labelling decision (as hate, offensive or normal) is based.\n", "citation": "@misc{mathew2020hatexplain,\n title={HateXplain: A Benchmark Dataset for Explainable Hate Speech Detection}, \n author={Binny Mathew and Punyajoy Saha and Seid Muhie Yimam and Chris Biemann and Pawan Goyal and Animesh Mukherjee},\n year={2020},\n eprint={2012.10289},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n", "homepage": "", "license": "cc-by-4.0", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "annotators": {"feature": {"label": {"num_classes": 3, "names": ["hatespeech", "normal", "offensive"], "names_file": null, "id": null, "_type": "ClassLabel"}, "annotator_id": {"dtype": "int32", "id": null, "_type": "Value"}, "target": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}, "rationales": {"feature": {"feature": {"dtype": "int32", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "length": -1, "id": null, "_type": "Sequence"}, "post_tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "builder_name": "hatexplain", "config_name": "plain_text", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 7114730, "num_examples": 15383, "dataset_name": "hatexplain"}, "validation": {"name": "validation", "num_bytes": 884940, "num_examples": 1922, "dataset_name": "hatexplain"}, "test": {"name": "test", "num_bytes": 884784, "num_examples": 1924, "dataset_name": "hatexplain"}}, "download_checksums": {"https://raw.githubusercontent.com/punyajoy/HateXplain/master/Data/dataset.json": {"num_bytes": 12256170, "checksum": "63bb3340fee0ec469b09690d04cb68f7c187787dd8b83807f071892c084967fb"}, "https://raw.githubusercontent.com/punyajoy/HateXplain/master/Data/post_id_divisions.json": {"num_bytes": 591921, "checksum": "c2fb0d89862e7897b11ea3e9380753f15a793482b4b70ad0532dfb1212212835"}}, "download_size": 12848091, "post_processing_size": null, "dataset_size": 8884454, "size_in_bytes": 21732545}}
|
|
|
|