| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Fairlex: A multilingual benchmark for evaluating fairness in legal text processing.""" |
|
|
| import json |
| import os |
| import textwrap |
|
|
| import datasets |
|
|
|
|
| MAIN_CITATION = """\ |
| @inproceedings{chalkidis-etal-2022-fairlex, |
| author={Chalkidis, Ilias and Passini, Tommaso and Zhang, Sheng and |
| Tomada, Letizia and Schwemer, Sebastian Felix and Søgaard, Anders}, |
| title={FairLex: A Multilingual Benchmark for Evaluating Fairness in Legal Text Processing}, |
| booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics}, |
| year={2022}, |
| address={Dublin, Ireland} |
| } |
| """ |
|
|
| _DESCRIPTION = """\ |
| Fairlex: A multilingual benchmark for evaluating fairness in legal text processing. |
| """ |
|
|
| ECTHR_ARTICLES = ["2", "3", "5", "6", "8", "9", "10", "11", "14", "P1-1"] |
|
|
| SCDB_ISSUE_AREAS = [ |
| "Criminal Procedure", |
| "Civil Rights", |
| "First Amendment", |
| "Due Process", |
| "Privacy", |
| "Attorneys", |
| "Unions", |
| "Economic Activity", |
| "Judicial Power", |
| "Federalism", |
| "Federal Taxation", |
| ] |
|
|
| FSCS_LABELS = ["dismissal", "approval"] |
|
|
| CAIL_LABELS = ["0", "<=12", "<=36", "<=60", "<=120", ">120"] |
|
|
|
|
| class FairlexConfig(datasets.BuilderConfig): |
| """BuilderConfig for Fairlex.""" |
|
|
| def __init__( |
| self, |
| label_column, |
| url, |
| data_url, |
| citation, |
| label_classes=None, |
| multi_label=None, |
| attributes=None, |
| **kwargs, |
| ): |
| """BuilderConfig for Fairlex. |
| |
| Args: |
| label_column: `string`, name of the column in the jsonl file corresponding |
| to the label |
| url: `string`, url for the original project |
| data_url: `string`, url to download the zip file from |
| data_file: `string`, filename for data set |
| citation: `string`, citation for the data set |
| url: `string`, url for information about the data set |
| label_classes: `list[string]`, the list of classes if the label is |
| categorical. If not provided, then the label will be of type |
| `datasets.Value('float32')`. |
| multi_label: `boolean`, True if the task is multi-label |
| attributes: `List<string>`, names of the protected attributes |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super(FairlexConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) |
| self.label_column = label_column |
| self.label_classes = label_classes |
| self.multi_label = multi_label |
| self.attributes = attributes |
| self.url = url |
| self.data_url = data_url |
| self.citation = citation |
|
|
|
|
| class Fairlex(datasets.GeneratorBasedBuilder): |
| """Fairlex: A multilingual benchmark for evaluating fairness in legal text processing. Version 1.0""" |
|
|
| BUILDER_CONFIGS = [ |
| FairlexConfig( |
| name="ecthr", |
| description=textwrap.dedent( |
| """\ |
| The European Court of Human Rights (ECtHR) hears allegations that a state has breached human rights |
| provisions of the European Convention of Human Rights (ECHR). We use the dataset of Chalkidis et al. |
| (2021), which contains 11K cases from ECtHR's public database. Each case is mapped to articles of the ECHR |
| that were violated (if any). This is a multi-label text classification task. Given the facts of a case, |
| the goal is to predict the ECHR articles that were violated, if any, as decided (ruled) by the court.""" |
| ), |
| label_column="labels", |
| label_classes=ECTHR_ARTICLES, |
| multi_label=True, |
| attributes=[ |
| ("applicant_age", ["n/a", "<=35", "<=65", ">65"]), |
| ("applicant_gender", ["n/a", "male", "female"]), |
| ("defendant_state", ["C.E. European", "Rest of Europe"]), |
| ], |
| data_url="https://zenodo.org/record/6322643/files/ecthr.zip", |
| url="https://huggingface.co/datasets/ecthr_cases", |
| citation=textwrap.dedent( |
| """\ |
| @inproceedings{chalkidis-etal-2021-paragraph, |
| title = "Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases", |
| author = "Chalkidis, Ilias and |
| Fergadiotis, Manos and |
| Tsarapatsanis, Dimitrios and |
| Aletras, Nikolaos and |
| Androutsopoulos, Ion and |
| Malakasiotis, Prodromos", |
| booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
| month = jun, |
| year = "2021", |
| address = "Online", |
| publisher = "Association for Computational Linguistics", |
| url = "https://aclanthology.org/2021.naacl-main.22", |
| doi = "10.18653/v1/2021.naacl-main.22", |
| pages = "226--241", |
| } |
| }""" |
| ), |
| ), |
| FairlexConfig( |
| name="scotus", |
| description=textwrap.dedent( |
| """\ |
| The US Supreme Court (SCOTUS) is the highest federal court in the United States of America and generally |
| hears only the most controversial or otherwise complex cases which have not been sufficiently well solved |
| by lower courts. We combine information from SCOTUS opinions with the Supreme Court DataBase (SCDB) |
| (Spaeth, 2020). SCDB provides metadata (e.g., date of publication, decisions, issues, decision directions |
| and many more) for all cases. We consider the available 14 thematic issue areas (e.g, Criminal Procedure, |
| Civil Rights, Economic Activity, etc.). This is a single-label multi-class document classification task. |
| Given the court opinion, the goal is to predict the issue area whose focus is on the subject matter |
| of the controversy (dispute). """ |
| ), |
| label_column="label", |
| label_classes=SCDB_ISSUE_AREAS, |
| multi_label=False, |
| attributes=[ |
| ("decision_direction", ["conservative", "liberal"]), |
| ("respondent_type", ["other", "person", "organization", "public entity", "facility"]), |
| ], |
| url="http://scdb.wustl.edu/data.php", |
| data_url="https://zenodo.org/record/6322643/files/scotus.zip", |
| citation=textwrap.dedent( |
| """\ |
| @misc{spaeth2020, |
| author = {Harold J. Spaeth and Lee Epstein and Andrew D. Martin, Jeffrey A. Segal |
| and Theodore J. Ruger and Sara C. Benesh}, |
| year = {2020}, |
| title ={{Supreme Court Database, Version 2020 Release 01}}, |
| url= {http://Supremecourtdatabase.org}, |
| howpublished={Washington University Law} |
| }""" |
| ), |
| ), |
| FairlexConfig( |
| name="fscs", |
| description=textwrap.dedent( |
| """\ |
| The Federal Supreme Court of Switzerland (FSCS) is the last level of appeal in Switzerland and similarly |
| to SCOTUS, the court generally hears only the most controversial or otherwise complex cases which have |
| not been sufficiently well solved by lower courts. The court often focus only on small parts of previous |
| decision, where they discuss possible wrong reasoning by the lower court. The Swiss-Judgment-Predict |
| dataset (Niklaus et al., 2021) contains more than 85K decisions from the FSCS written in one of three |
| languages (50K German, 31K French, 4K Italian) from the years 2000 to 2020. The dataset is not parallel, |
| i.e., all cases are unique and decisions are written only in a single language. The dataset provides labels |
| for a simplified binary (approval, dismissal) classification task. Given the facts of the case, the goal |
| is to predict if the plaintiff's request is valid or partially valid.""" |
| ), |
| label_column="label", |
| label_classes=FSCS_LABELS, |
| multi_label=False, |
| attributes=[ |
| ("decision_language", ["de", "fr", "it"]), |
| ("legal_area", ["other", "public law", "penal law", "civil law", "social law", "insurance law"]), |
| ( |
| "court_region", |
| [ |
| "n/a", |
| "Région lémanique", |
| "Zürich", |
| "Espace Mittelland", |
| "Northwestern Switzerland", |
| "Eastern Switzerland", |
| "Central Switzerland", |
| "Ticino", |
| "Federation", |
| ], |
| ), |
| ], |
| url="https://github.com/JoelNiklaus/SwissCourtRulingCorpus", |
| data_url="https://zenodo.org/record/6322643/files/fscs.zip", |
| citation=textwrap.dedent( |
| """\ |
| @InProceedings{niklaus-etal-2021-swiss, |
| author = {Niklaus, Joel |
| and Chalkidis, Ilias |
| and Stürmer, Matthias}, |
| title = {Swiss-Court-Predict: A Multilingual Legal Judgment Prediction Benchmark}, |
| booktitle = {Proceedings of the 2021 Natural Legal Language Processing Workshop}, |
| year = {2021}, |
| location = {Punta Cana, Dominican Republic}, |
| }""" |
| ), |
| ), |
| FairlexConfig( |
| name="cail", |
| description=textwrap.dedent( |
| """\ |
| The Supreme People's Court of China (CAIL) is the last level of appeal in China and considers cases that |
| originated from the high people's courts concerning matters of national importance. The Chinese AI and Law |
| challenge (CAIL) dataset (Xiao et al., 2018) is a Chinese legal NLP dataset for judgment prediction and |
| contains over 1m criminal cases. The dataset provides labels for relevant article of criminal code |
| prediction, charge (type of crime) prediction, imprisonment term (period) prediction, and monetary penalty |
| prediction. The updated (soft) version of the CAIL dataset has 104K criminal court cases. The tasks is |
| crime severity prediction task, a multi-class classification task, where given the facts of a case, |
| the goal is to predict how severe was the committed crime with respect to the imprisonment term. |
| We approximate crime severity by the length of imprisonment term, split in 6 clusters |
| (0, >=12, >=36, >=60, >=120, >120 months).""" |
| ), |
| label_column="label", |
| label_classes=CAIL_LABELS, |
| multi_label=False, |
| attributes=[ |
| ("defendant_gender", ["male", "female"]), |
| ("court_region", ["Beijing", "Liaoning", "Hunan", "Guangdong", "Sichuan", "Guangxi", "Zhejiang"]), |
| ], |
| url="https://github.com/thunlp/LegalPLMs", |
| data_url="https://zenodo.org/record/6322643/files/cail.zip", |
| citation=textwrap.dedent( |
| """\ |
| @article{wang-etal-2021-equality, |
| title={Equality before the Law: Legal Judgment Consistency Analysis for Fairness}, |
| author={Yuzhong Wang and Chaojun Xiao and Shirong Ma and Haoxi Zhong and Cunchao Tu and Tianyang Zhang and Zhiyuan Liu and Maosong Sun}, |
| year={2021}, |
| journal={Science China - Information Sciences}, |
| url={https://arxiv.org/abs/2103.13868} |
| }""" |
| ), |
| ), |
| ] |
|
|
| def _info(self): |
| features = {"text": datasets.Value("string")} |
| if self.config.multi_label: |
| features["labels"] = datasets.features.Sequence(datasets.ClassLabel(names=self.config.label_classes)) |
| else: |
| features["label"] = datasets.ClassLabel(names=self.config.label_classes) |
| for attribute_name, attribute_groups in self.config.attributes: |
| features[attribute_name] = datasets.ClassLabel(names=attribute_groups) |
| return datasets.DatasetInfo( |
| description=self.config.description, |
| features=datasets.Features(features), |
| homepage=self.config.url, |
| citation=self.config.citation + "\n" + MAIN_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| data_dir = dl_manager.download_and_extract(self.config.data_url) |
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TRAIN, |
| |
| gen_kwargs={ |
| "filepath": os.path.join(data_dir, "train.jsonl"), |
| "split": "train", |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| |
| gen_kwargs={ |
| "filepath": os.path.join(data_dir, "test.jsonl"), |
| "split": "test", |
| }, |
| ), |
| datasets.SplitGenerator( |
| name=datasets.Split.VALIDATION, |
| |
| gen_kwargs={ |
| "filepath": os.path.join(data_dir, "val.jsonl"), |
| "split": "val", |
| }, |
| ), |
| ] |
|
|
| def _generate_examples(self, filepath, split): |
| """This function returns the examples in the raw (text) form.""" |
| with open(filepath, encoding="utf-8") as f: |
| for id_, row in enumerate(f): |
| data = json.loads(row) |
| example = { |
| "text": data["text"], |
| self.config.label_column: data[self.config.label_column], |
| } |
| for attribute_name, _ in self.config.attributes: |
| example[attribute_name] = data["attributes"][attribute_name] |
| yield id_, example |
|
|