Datasets:
Tasks:
Text Classification
Languages:
English
task_categories: | |
- text-classification | |
language: | |
- en | |
# Movie Review Data | |
* Original source: sentence polarity dataset v1.0 http://www.cs.cornell.edu/people/pabo/movie-review-data/ | |
* Seems to same as https://huggingface.co/datasets/rotten_tomatoes, but different split. | |
## Original README | |
======= | |
Introduction | |
This README v1.0 (June, 2005) for the v1.0 sentence polarity dataset comes | |
from the URL | |
http://www.cs.cornell.edu/people/pabo/movie-review-data . | |
======= | |
Citation Info | |
This data was first used in Bo Pang and Lillian Lee, | |
``Seeing stars: Exploiting class relationships for sentiment categorization | |
with respect to rating scales.'', Proceedings of the ACL, 2005. | |
@InProceedings{Pang+Lee:05a, | |
author = {Bo Pang and Lillian Lee}, | |
title = {Seeing stars: Exploiting class relationships for sentiment | |
categorization with respect to rating scales}, | |
booktitle = {Proceedings of the ACL}, | |
year = 2005 | |
} | |
======= | |
Data Format Summary | |
- rt-polaritydata.tar.gz: contains this readme and two data files that | |
were used in the experiments described in Pang/Lee ACL 2005. | |
Specifically: | |
* rt-polarity.pos contains 5331 positive snippets | |
* rt-polarity.neg contains 5331 negative snippets | |
Each line in these two files corresponds to a single snippet (usually | |
containing roughly one single sentence); all snippets are down-cased. | |
The snippets were labeled automatically, as described below (see | |
section "Label Decision"). | |
Note: The original source files from which the data in | |
rt-polaritydata.tar.gz was derived can be found in the subjective | |
part (Rotten Tomatoes pages) of subjectivity_html.tar.gz (released | |
with subjectivity dataset v1.0). | |
======= | |
Label Decision | |
We assumed snippets (from Rotten Tomatoes webpages) for reviews marked with | |
``fresh'' are positive, and those for reviews marked with ``rotten'' are | |
negative. | |
## Preprocessing | |
To make csv with text and label field, we use the following script. | |
```python3 | |
import csv | |
import random | |
# NOTE: The encoding of original file is "latin_1". We will change it to "utf8". | |
with open("rt-polarity.pos", encoding="latin_1") as f: | |
texts_pos = [line.strip() for line in f] | |
with open("rt-polarity.neg", encoding="latin_1") as f: | |
texts_neg = [line.strip() for line in f] | |
rows_pos = [{"text": text, "label": 1} for text in texts_pos] | |
rows_neg = [{"text": text, "label": 0} for text in texts_pos] | |
# NOTE: For fair validation, we split it into train and test. Also, for the research who wants to use different setting, we provide whole setting. | |
# NOTE: We follow the split setting in LM-BFF paper. | |
rows_whole = rows_pos + rows_neg | |
random.Random(42).shuffle(rows_whole) | |
rows_test, rows_train = rows_whole[:2000], rows_whole[2000:] | |
with open("whole.csv", "w", encoding="utf8") as f: | |
writer = csv.DictWriter(f, fieldnames=["text", "label"]) | |
writer.writerows(rows_train) | |
with open("train.csv", "w", encoding="utf8") as f: | |
writer = csv.DictWriter(f, fieldnames=["text", "label"]) | |
writer.writerows(rows_train) | |
with open("test.csv", "w", encoding="utf8") as f: | |
writer = csv.DictWriter(f, fieldnames=["text", "label"]) | |
writer.writerows(rows_test) | |
``` | |