File size: 4,723 Bytes
56feadb
 
 
 
0acf86a
e0b55da
9d962b5
0acf86a
fca3715
 
 
0acf86a
56feadb
 
 
 
0acf86a
 
56feadb
 
 
 
 
 
0acf86a
 
56feadb
 
 
 
 
 
 
 
0acf86a
 
56feadb
0acf86a
 
56feadb
0acf86a
 
56feadb
0acf86a
56feadb
 
fad1577
 
56feadb
 
 
0acf86a
56feadb
0acf86a
 
56feadb
 
0acf86a
56feadb
0acf86a
 
 
 
56feadb
 
 
 
 
 
 
 
 
 
0acf86a
 
 
 
56feadb
 
 
 
 
 
 
 
 
 
 
 
 
0acf86a
56feadb
0acf86a
56feadb
0acf86a
 
 
e0b55da
56feadb
2fa4479
56feadb
 
98e6bcc
56feadb
5952276
98e6bcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
"""minipileoflaw"""


import gzip
import json
import csv
import pandas as pd

import json
import logging

import datasets
try:
    import lzma as xz
except ImportError:
    import pylzma as xz


datasets.logging.set_verbosity_info()
logger = datasets.logging.get_logger(__name__)


_DESCRIPTION = """
This is minipileoflaw
"""

_CITATION = """
@misc{hendersonkrass2022pileoflaw,
  url = {https://arxiv.org/abs/2207.00220},
  author = {Henderson, Peter and Krass, Mark S. and Zheng, Lucia and Guha, Neel and Manning, Christopher D. and Jurafsky, Dan and Ho, Daniel E.},
  title = {Pile of Law: Learning Responsible Data Filtering from the Law and a 256GB Open-Source Legal Dataset},
  publisher = {arXiv},
  year = {2022}
}
"""

_URL = "https://huggingface.co/datasets/tomrb/minipileoflaw"


BASE_URL = "https://huggingface.co/datasets/tomrb/minipileoflaw/blob/main/data/minipileoflaw_"


subsets_names = ['r_legaladvice', 'courtlistener_docket_entry_documents', 'atticus_contracts', 'courtlistener_opinions', 'federal_register', 'bva_opinions', 'us_bills', 'cc_casebooks', 'tos', 'euro_parl', 'nlrb_decisions', 'scotus_oral_arguments', 'cfr', 'state_codes', 'scotus_filings', 'exam_outlines', 'edgar', 'cfpb_creditcard_contracts', 'constitutions', 'congressional_hearings', 'oig', 'olc_memos', 'uscode', 'founding_docs', 'ftc_advisory_opinions', 'echr', 'eurlex', 'tax_rulings', 'un_debates', 'fre', 'frcp', 'canadian_decisions', 'eoir', 'dol_ecab', 'icj-pcij', 'uspto_office_actions', 'ed_policy_guidance', 'acus_reports', 'hhs_alj_opinions', 'sec_administrative_proceedings', 'fmshrc_bluebooks', 'resource_contracts', 'medicaid_policy_guidance', 'irs_legal_advice_memos', 'doj_guidance_documents']

_DATA_URL = {
    key: {
        "train": [f"{BASE_URL}{key}_train.json"],
        "validation": [f"{BASE_URL}{key}_validation.json"]
    }
    for key in subsets_names
}

_VARIANTS = ["all"] + list(_DATA_URL.keys())


class MiniPileOfLaw(datasets.GeneratorBasedBuilder):
    """TODO"""

    BUILDER_CONFIGS = [datasets.BuilderConfig(name) for name in _VARIANTS]

    def _info(self):
        return datasets.DatasetInfo(
            description=_DESCRIPTION,
            features=datasets.Features(
                {
                    "text": datasets.Value("string"),
                    "created_timestamp": datasets.Value("string"),
                    "downloaded_timestamp": datasets.Value("string"),
                    "url": datasets.Value("string"),
                }
            ),
            supervised_keys=None,
            homepage=_URL,
            citation=_CITATION,
        )

    def _split_generators(self, dl_manager):
        data_urls = {}
        if self.config.name == "all":
            data_sources = list(_DATA_URL.keys())
        else:
            data_sources = [self.config.name]
        for split in ["train", "validation"]:
            data_urls[split] = []
            for source in data_sources:
                for chunk in _DATA_URL[source][split]:
                    data_urls[split].append(chunk)

        train_downloaded_files = dl_manager.download(data_urls["train"])
        validation_downloaded_files = dl_manager.download(data_urls["validation"])
        return [
            datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepaths": train_downloaded_files}),
            datasets.SplitGenerator(
                name=datasets.Split.VALIDATION, gen_kwargs={"filepaths": validation_downloaded_files}
            ),
        ]


    def _generate_examples(self, filepaths):
        """This function returns the examples in the raw (text) form by iterating on all the files."""
        id_ = 0
        for filepath in filepaths:
            logging.info("Generating examples from = %s", filepath)
            try:
                with open(filepath, "r", encoding="utf-8") as f:
                    for line in f:
                        try:
                            example = json.loads(line)
                            if example is not None and isinstance(example, dict):
                                yield id_, {
                                    "text": example.get("text", ""),
                                    "created_timestamp": example.get("created_timestamp", ""),
                                    "downloaded_timestamp": example.get("downloaded_timestamp", ""),
                                    "url": example.get("url", "")
                                }
                                id_ += 1
                        except json.JSONDecodeError:
                            print(f"Error decoding JSON from line in file: {filepath}")
            except IOError:
                print(f"Error reading file: {filepath}")