You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

Dataset Card for "IL-TUR"

Dataset Description

Dataset Summary

"IL-TUR": Benchmark for Indian Legal Text Understanding and Reasoning is a collaborative effort to establish a modern benchmark for training and evaluating AI/NLP models on Indian Law. IL-TUR consists of 8 foundational tasks, requiring different types of understanding and skills. Apart from English, some tasks involve Indic languages.

This dataset repository has been created to unify the data formats of all tasks in a singular style. Users will have the option to both utilize the Dataset Viewer to inspect each dataset and data split, as well as use Python commands to download and access the data directly. We hope that this will encourage more researchers to take up these tasks in legal AI, and help them adapt easily.

Supported Tasks

The following tasks (and datasets) are part of IL-TUR:

Task Dataset Source Language Task Type
L-NER 105 docs - en Sequence Classification
RR 21,184 sents Malik et al. (2022) en Multi-label Classification
CJPE ILDC Malik et al. (2021) en Binary Classification, Extraction
BAIL HLDC Kapoor et al. (2022) hi Binary Classification
LSI ILSI Paul et al. (2022) en Multi-label Classification
PCR IL-PCR Joshi et al. (2023) en Retrieval
SUMM In-Abs Shukla et al. (2022) en Generation
L-MT MiLPAC Mahapatra et al. (2023) en, hi, bn, gu, mr, ml, or, pa, ta, te Generation

L-NER (Legal Named Entity Recognition)

L-NER aims to automatically predict Named Entities (e.g. Judge, Appellant, Respondent, etc.) in a legal document, along with the class. Unlike standard NER, legal entities are more fine-grained, e.g., PER can be categorized as JUDGE, APPELLANT, RESPONDENT, etc.

Data Source: The dataset consists of 105 Indian Supreme Court case documents. All documents are annotated with every instance of NERs.

Data Splits: The data is divided into 3 folds, 'fold_1', 'fold_2' and 'fold_3' each having 35 documents.

Data Instance: An example from 'fold_1' looks as follows:

{
  "id": "115651329"
  "text": "REPORTABLE IN THE SUPREME COURT OF INDIA CRIMINAL APPELLATE JURISDICTION CRIMINAL APPEAL NO. 92/2015 JAGE RAM & ORS. ..."
  "spans": [
    {"start": 137, "end": 153, "label": 1},
    {"start": 252, "end": 261, "label": 10}
    ...
  ]
}

Instances from all splits have a similar structure.

Data Fields:

  • 'id': string → IndianKanoon Case ID

  • 'text': string → Full document text

  • 'spans': List(

    • 'start': int → starting char index
    • 'end': int → ending char index + 1
    • 'label': class_label → NER label

    )

List of NER labels "APP", "RESP", "A.COUNSEL", "R.COUNSEL", "JUDGE", "WIT", "AUTH", "COURT", "STAT", "PREC", "DATE", "CASENO"


RR (Rhetorical Role Prediction)

RR aims to segment a legal document into topically coherent units such as Facts, Arguments, Rulings, etc. Unlike documents of many countries, Indian court case documents often do not have explicit segmentation or headings, etc., necessiating the automation of this task.

Data Source: The dataset was created from 100 cases involving Competition Law and Income Tax from the Indian Supreme Court.

Data Splits: The data is divided into 6 splits; 'CL_train' (40 docs), 'CL_dev' (5 docs) and 'CL_test' (5 docs) for Competition Law, 'IT_train' (40 docs), 'IT_dev' (5 docs) and 'IT_test' (5 docs) for Income Tax

Data Instance: An example from 'fold_1' looks as follows:

{
  "id": "CCI_Confederation_of_Real_Estate_Developers_AssociatioCO201807081816081635COM964576"
  "text": ["The instant matter filed by the Confederation of Real Estate Developers Association of India ...", "In the said order, the Commission had held ...", ...]
  "labels": [0, 9, ...]
  "expert_1": {
    "primary": [0, 9, ...]
    "secondary": [12, 12, ...]
    "tertiary": [12, 12, ...]
    "overall": [0, 9, ...]
  }
  "expert_2": {...}
  "expert_3": {...}
  }
}

Instances from all splits have a similar structure.

Data Fields:

  • 'id': string → Case ID
  • 'text': List(string) → Full document text divided into sentences
  • 'labels': List(class_label) → Final rhetorical role of each sentence as per majority decision
  • 'expert_1':
    • 'primary': List(class_label) → Primary rhetorical role of each sentence according to expert_1
    • 'secondary': List(class_label) → Secondary rhetorical role of each sentence according to expert_1
    • 'tertiary': List(class_label) → Tertiary rhetorical role of each sentence according to expert_1
    • 'overall': List(class_label) → Final rhetorical role of each sentence according to expert_1
  • 'expert_2' → Similar to expert_1
  • 'expert_3' → Similar to expert_1
List of RR labels "Fact", "Issue", "ArgumentPetitioner", "ArgumentRespondent", "Statute", "Dissent", "PrecedentReliedUpon", "PrecedentNotReliedUpon", "PrecedentOverruled", "RulingByLowerCourt", "RatioOfTheDecision", "RulingByPresentCourt", "None"


CJPE (Court Judgment Prediction and Explanation)

CJPE requires, given the facts and other details of a court case, predicting the final outcome, i.e., appeal accepted/rejected (Prediction), as well as identifying the salient sentences leading to the decision (Explanation). In cases containing multiple appeals, the final outcome is accepted if at least one of the appeals are accepted, else rejected.

Data Source: The dataset consists of 34k Indian Supreme Court case documents. Out of these, 56 documents are annotated with the salient sentences by five legal experts.

Data Splits: The data is divided into 6 parts. For prediction, 'single_train' (5k docs) and 'single_dev' (2.5k docs) consist of cases having single appeals, while 'multi_train' (32k docs) and 'multi_dev' (1k docs) consist of cases having multiple appeals. The 'test' split (1.5k) is the singular test set for both of these settings. The 'expert' split (56 docs) contains documents annotated with the explanations (salient sentences).

Data Instance: An example from 'expert' looks as follows:

{
  "id": "1951_10"
  "text": "CIVIL APPELLATE JURISDICTION Appeal (Civil Appeal No. 57 of 1950) from a judgment and decree of the High Court of Judicature at Bombay dated 1st April, 1948, in Appeal No. 365 of 1947 reversing a judgment of the Joint Civil Judge at Ahmedabad, dated 14th October, 1947 ..."
  "label": 1
  "expert_1": {
    "label": 1
    "rank1": ["Mr. Daphthary contended that the whole object ...", ...]
    "rank2": ["On a plain reading of the language of sections 12 and 50 ...", ...]
    "rank3": ["In our opinion, the decision of the appeal depends ...", ...]
    "rank4": ["It was contended before the High Court that ...", ...]
    "rank5": ["The appellants are owners of a property known as ...", ...]
  }
  "expert_2": {...}
  "expert_3": {...}
  "expert_4": {...}
  "expert_5": {...} 
}

Instances from all other splits have a similar structure, except that the "expert_k" keys have None values.

Data Fields:

  • 'id': string → Case ID
  • 'text': string → Full document text
  • 'label': class_label → final decision
  • 'expert_1':
    • 'label': class_label → final decision according to expert_1
    • 'rank1': List(string) → sentences leading to decision according to expert_1
    • 'rank2': List(string) → sentences contributing to decision according to expert_1
    • 'rank2': List(string) → sentences indicating the decision according to expert_1
    • 'rank2': List(string) → sentences without direct contribution but important for the case according to expert_1
    • 'rank2': List(string) → sentences without direct contribution but important for the case according to expert_1
  • 'expert_2' → Similar to expert_1
  • 'expert_3' → Similar to expert_1
  • 'expert_4' → Similar to expert_1
  • 'expert_5' → Similar to expert_1
List of CJPE labels "REJECTED", "ACCEPTED"


BAIL (Bail Prediction)

BAIL aims to automatically predict whether the accused should be granted bail or not, given a criminal case document.

Data Source: The dataset consists of a total of 176k docs (in Hindi) from different District Courts of Uttar Pradesh. All documents are segmented facts/arguments and judge opinion.

Data Splits: The data is divided into train, dev and test in two different ways. 'train_all', 'dev_all' and 'test_all' are created by distributing the cases randomly. 'train_specific', 'dev_specific' and 'test_specific' are created by grouping docs from certain districts together.

Data Instance: An example from 'train_all' looks as follows:

{
  "id": "Bail Application_2008_202101-04-2021407"
  "district": "agra"
  "text": {
    "facts-and-arguments": ["अग्रिम जमानत प्रार्थनापत्र के समर्थन में प्रार्थी, ...", "अभियोजन कथानक सक्षेंप में इस प्रकार ...", ...]
    "judge-opinion": ["प्रार्थी / अभियुक्तगण के विद्धान अधिवक्त ...", "पत्रावली के अवलोकन से स्पष्ट है कि अभियुक्तगण ...", ...]
  }
  "label": 0
}

Instances from all splits have a similar structure.

Data Fields:

  • 'id': string → Case ID
  • 'district': string → Name of the District Court
  • 'text':
    • 'facts-and-arguments': List(string) → sentences containing facts and arguments by lawyers
    • 'judge-opinion': List(string) → sentences containing the opinion of the judge
  • 'label': class_label → final bail decision
List of BAIL labels "DENIED", "GRANTED"


LSI (Legal Statute Identification)

LSI involves identifying the relevant statute(s) (written laws) given the facts of a court case document, from a predetermined set of target statutes.

Data Source: The dataset was created from 66k Indian Supreme Court and prominent High Court documents, each citing one or more statutes from a target set of 100 Indian Penal Code statutes.

Data Splits: The data is divided into 'train' (43k docs), 'dev' (10k docs) and 'test' (13k docs) splits. There is an additional 'statutes' split (100 docs) containing the description of the statutes.

Data Instance: An example from 'train' looks as follows:

{
  "id": "100120460"
  "text": ["It is further alleged that present applicant ...", "Complainant also alleges in the FIR that ...", ...]
  "labels": [77, 71, 15, 74, 72, 16]
}

Instances from all 'dev' and 'test' have a similar structure. But, an example from 'statutes' looks as follows:

{
  "id": "Section 120B"
  "text": ["Punishment of criminal conspiracy:", "(1) Whoever is a party to a criminal conspiracy ...", ...]
  "labels": None
}

Data Fields:

  • 'id': string → IndianKanoon Case ID / IPC Section Number
  • 'text': List(string) → sentences containing only the facts
  • 'labels': List(class_label) → list of relevant statutes
List of LSI labels Section 2, Section 3, Section 4, Section 5, Section 13, Section 34, Section 107, Section 109, Section 114, Section 120, Section 120B, Section 143, Section 147, Section 148, Section 149, Section 155, Section 156, Section 161, Section 164, Section 173, Section 174A, Section 186, Section 188, Section 190, Section 193, Section 200, Section 201, Section 228, Section 229A, Section 279, Section 294, Section 294(b), Section 299, Section 300, Section 302, Section 304, Section 304A, Section 304B, Section 306, Section 307, Section 308, Section 313, Section 320, Section 323, Section 324, Section 325, Section 326, Section 332, Section 336, Section 337, Section 338, Section 341, Section 342, Section 353, Section 354, Section 363, Section 364, Section 365, Section 366, Section 366A, Section 375, Section 376, Section 376(2), Section 379, Section 380, Section 384, Section 389, Section 392, Section 394, Section 395, Section 397, Section 406, Section 409, Section 411, Section 415, Section 417, Section 419, Section 420, Section 427, Section 436, Section 437, Section 438, Section 447, Section 448, Section 450, Section 452, Section 457, Section 465, Section 467, Section 468, Section 471, Section 482, Section 494, Section 498, Section 498A, Section 500, Section 504, Section 506, Section 509, Section 511


PCR (Prior Case Retrieval)

PCR requires identifying relevant prior cases (based on facts and precedents) from a set of candidate case documents, given a query case document. The task is modeled as a pure retrieval task instead of multi-label classification since the candidate pool is large and dynamic.

Data Source: Both candidate and query documents (total 8k docs) are from the Indian Supreme Court.

Data Splits: Both queries and candidates are split for train/dev/test, resulting in 6 splits: 'train_queries' (827 docs), 'dev_queries' (118 docs), 'test_queries' (237 docs), 'train_candidates' (4.3k docs), 'dev_candidates' (1k docs), 'test_candidates' (1.7k docs).

Data Instance: An example from 'train_queries' looks as follows:

{
  "id": "100120460"
  "text": ["CASE NO.: Appeal (civil) 2387 of 2001 ...", "THOMAS, J.", ...]
  "relevant_candidates": ["0000586994", "0000772259", "0001182839", "0001610057"]
}

Instances from 'dev_queries' and 'test_queries' have a similar structure. For '*_candidates' splits, the 'relevant_candidates' is None.

Data Fields:

  • 'id': string → IndianKanoon Case ID
  • 'text': List(string) → Full document text divided into sentences
  • 'relevant_candidates': List(string) → list of relevant document IDs


SUMM (Summarization)

SUMM automates the process of generating a gist (abstractive summary) of a legal case document that captures the critical aspects of the case. These summaries are similar to headnotes provided in many case documents.

Data Source: The dataset was created from 7.1k Indian Supreme Court case documents.

Data Splits: The data is divided into two splits: 'train' (7k docs) and 'test' (100 docs).

Data Instance: An example from 'train' looks as follows:

{
  "id": "2858"
  "num_doc_tokens": 4867
  "num_summ_tokens": 708
  "document": ["Appeal No. 36 of 1967.", "Appeal by special leave from the judgment and order dated August 25, 1966 ...", ...]
  "summary": ["The appellant executed a usufructuary mortgage ...", "In 1949, the mortgagee left for Pakistan.", ...]
}

Instances from all splits have a similar structure.

Data Fields:

  • 'id': string → IndianKanoon Case ID
  • 'num_doc_tokens': int → the number of words in the full document
  • 'num_summ_tokens': int → the number of words in the summary document
  • 'document': List(string) → Full document text divided into sentences
  • 'summary': List(string) → Summary text divided into sentences


L-MT (Legal Machine Translation)

L-MT automates the process of converting a legal text in English to different Indic languages (Hindi, Bengali, etc.).

Data Source: The dataset was created from Central Government Acts, Intellectual Property primers and FAQs by the Competition Commission of India. There are a total of 6.6k en-xx pairs.

Data Splits: The data is divided into 'acts' (4k pairs), 'ip' (1k pairs), 'cci_faq' (1.4k pairs).

Data Instance: An example from 'acts' looks as follows:

{
  "id": "2006_4_1_0/HI",
  "src_lang": "EN",
  "src": "1. Short title, extent and commencement.—",
  "tgt_lang": "HI",
  "tgt": "संक्षिप्त नाम, विस्तार और प्रारंभ--"
}

Instances from all splits have a similar structure.

Data Fields:

  • 'id': string → Instance ID
  • 'src_lang': string → Source language
  • 'src': string → Source text
  • 'tgt_lang': string → Target language
  • 'tgt': string → Target text

Benchmark Results

TaskSOTAMetricModel
L-NER48.58%strict m-F1InLegalBERT + CRF
RR69.01%m-F1MTL-BERT
CJPE81.31%m-F1InLegalBERT + BiLSTM
0.56ROUGE-L
0.32BLEU
BAIL81%m-F1TF-IDF + IndicBERT
LSI28.08%m-F1LeSICiN (Graph-based)
PCR39.15%µ-F1@KEvent-based
SUMM0.33ROUGE-LLegal LED
0.86BERTScore
L-MT0.28BLEUMSFT (Microsoft Translation)
0.32GLEU
0.57chrF++

Citation Information

@inproceedings{iltur-2024,
      title = "IL-TUR: Benchmark for Indian Legal Text Understanding and Reasoning",
      author = "Joshi, Abhinav and Paul, Shounak and Sharma, Akshat and Goyal, Pawan and Ghosh, Saptarshi and Modi, Ashutosh"
      booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
      month = aug,
      year = "2024",
      address = "Bangkok, Thailand",
      publisher = "Association for Computational Linguistics",
  }
Downloads last month
28