Datasets:
Tasks:
Text Classification
Modalities:
Text
Formats:
json
Sub-tasks:
multi-class-classification
Languages:
German
Size:
10K - 100K
License:
| annotations_creators: | |
| - expert-generated | |
| - found | |
| language_creators: | |
| - found | |
| language: | |
| - de | |
| license: | |
| - cc-by-4.0 | |
| multilinguality: | |
| - monolingual | |
| pretty_name: Annotated German Legal Decision Corpus | |
| size_categories: | |
| - 10K<n<100K | |
| source_datasets: | |
| - original | |
| task_categories: | |
| - text-classification | |
| task_ids: | |
| - multi-class-classification | |
| # Dataset Card for Annotated German Legal Decision Corpus | |
| ## Table of Contents | |
| - [Table of Contents](#table-of-contents) | |
| - [Dataset Description](#dataset-description) | |
| - [Dataset Summary](#dataset-summary) | |
| - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) | |
| - [Languages](#languages) | |
| - [Dataset Structure](#dataset-structure) | |
| - [Data Instances](#data-instances) | |
| - [Data Fields](#data-fields) | |
| - [Data Splits](#data-splits) | |
| - [Dataset Creation](#dataset-creation) | |
| - [Curation Rationale](#curation-rationale) | |
| - [Source Data](#source-data) | |
| - [Annotations](#annotations) | |
| - [Personal and Sensitive Information](#personal-and-sensitive-information) | |
| - [Considerations for Using the Data](#considerations-for-using-the-data) | |
| - [Social Impact of Dataset](#social-impact-of-dataset) | |
| - [Discussion of Biases](#discussion-of-biases) | |
| - [Other Known Limitations](#other-known-limitations) | |
| - [Additional Information](#additional-information) | |
| - [Dataset Curators](#dataset-curators) | |
| - [Licensing Information](#licensing-information) | |
| - [Citation Information](#citation-information) | |
| - [Contributions](#contributions) | |
| ## Dataset Description | |
| - **Homepage:** | |
| - **Repository:** https://zenodo.org/record/3936490#.X1ed7ovgomK | |
| - **Paper:** Urchs., S., Mitrović., J., & Granitzer., M. (2021). Design and Implementation of German Legal Decision | |
| Corpora. Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART, | |
| 515–521. https://doi.org/10.5220/0010187305150521 | |
| - **Leaderboard:** | |
| - **Point of Contact:** [Joel Niklaus](mailto:joel.niklaus.2@bfh.ch) | |
| ### Dataset Summary | |
| This dataset consists of 200 randomly chosen judgments. In these judgments a legal expert annotated the components | |
| conclusion, definition and subsumption of the German legal writing style Urteilsstil. | |
| *"Overall 25,075 sentences are annotated. 5% (1,202) of these sentences are marked as conclusion, 21% (5,328) as | |
| definition, 53% (13,322) are marked as subsumption and the remaining 21% (6,481) as other. The length of judgments in | |
| sentences ranges from 38 to 862 sentences. The median of judgments have 97 sentences, the length of most judgments is on | |
| the shorter side."* (Urchs. et al., 2021) | |
| *"Judgments from 22 of the 131 courts are selected for the corpus. Most judgments originate from the VG Augsburg (59 / | |
| 30%) followed by the VG Ansbach (39 / 20%) and LSG Munich (33 / 17%)."* (Urchs. et al., 2021) | |
| *"29% (58) of all selected judgments are issued in the year 2016, followed by 22% (44) from the year 2017 and 21% (41) | |
| issued in the year 2015. [...] The percentages of selected judgments and decisions issued in 2018 and 2019 are roughly | |
| the same. No judgments from 2020 are selected."* (Urchs. et al., 2021) | |
| ### Supported Tasks and Leaderboards | |
| The dataset can be used for multi-class text classification tasks, more specifically, for argument mining. | |
| ### Languages | |
| The language in the dataset is German as it is used in Bavarian courts in Germany. | |
| ## Dataset Structure | |
| ### Data Instances | |
| Each sentence is saved as a json object on a line in one of the three files `train.jsonl`, `validation.jsonl` | |
| or `test.jsonl`. The file `meta.jsonl` contains meta information for each court. The `file_number` is present in all | |
| files for identification. Each sentence of the court decision was categorized according to its function. | |
| ### Data Fields | |
| The file `meta.jsonl` contains for each row the following fields: | |
| - `meta_title`: Title provided by the website, it is used for saving the decision | |
| - `court`: Issuing court | |
| - `decision_style`: Style of the decision; the corpus contains either *Urteil* (='judgment') or *Endurteil* ( | |
| ='end-judgment') | |
| - `date`: Date when the decision was issued by the court | |
| - `file_number`: Identification number used for this decision by the court | |
| - `title`: Title provided by the court | |
| - `norm_chains`: Norms related to the decision | |
| - `decision_guidelines`: Short summary of the decision | |
| - `keywords`: Keywords associated with the decision | |
| - `lower_court`: Court that decided on the decision before | |
| - `additional_information`: Additional Information | |
| - `decision_reference`: References to the location of the decision in beck-online | |
| - `tenor`: Designation of the legal consequence ordered by the court (list of paragraphs) | |
| - `legal_facts`: Facts that form the base for the decision (list of paragraphs) | |
| The files `train.jsonl`, `validation.jsonl` and `test.jsonl` contain the following fields: | |
| - `file_number`: Identification number for linkage with the file `meta.jsonl` | |
| - `input_sentence`: The sentence to be classified | |
| - `label`: In depth explanation of the court decision. Each sentence is assigned to one of the major components of | |
| German *Urteilsstil* (Urchs. et al., 2021) (list of paragraphs, each paragraph containing list of sentences, each | |
| sentence annotated with one of the following four labels): | |
| - `conclusion`: Overall result | |
| - `definition`: Abstract legal facts and consequences | |
| - `subsumption`: Determination sentence / Concrete facts | |
| - `other`: Anything else | |
| - `context_before`: Context in the same paragraph before the input_sentence | |
| - `context_after`: Context in the same paragraph after the input_sentence | |
| ### Data Splits | |
| No split provided in the original release. | |
| Splits created by Joel Niklaus. We randomly split the dataset into 80% (160 decisions, 19271 sentences) train, 10% | |
| validation (20 decisions, 2726 sentences) and 10% test (20 decisions, 3078 sentences). We made sure, that a decision | |
| only occurs in one split and is not dispersed over multiple splits. | |
| Label Distribution | |
| | label | train | validation | test | | |
| |:---------------|-----------:|-------------:|----------:| | |
| | conclusion | 975 | 115 | 112 | | |
| | definition | 4105 | 614 | 609 | | |
| | subsumption | 10034 | 1486 | 1802 | | |
| | other | 4157 | 511 | 555 | | |
| | total | **19271** | **2726** | **3078** | | |
| ## Dataset Creation | |
| ### Curation Rationale | |
| Creating a publicly available German legal text corpus consisting of judgments that have been annotated by a legal | |
| expert. The annotated components consist of *conclusion*, *definition* and *subsumption* of the German legal writing | |
| style *Urteilsstil*. | |
| ### Source Data | |
| #### Initial Data Collection and Normalization | |
| *“The decision corpus is a collection of the decisions published on the website www.gesetze-bayern.de. At the time of | |
| the crawling the website offered 32,748 decisions of 131 Bavarian courts, dating back to 2015. The decisions are | |
| provided from the Bavarian state after the courts agreed to a publication. All decisions are processed by the publisher | |
| C.H.BECK, commissioned by the Bavarian state. This processing includes anonymisation, key-wording, and adding of | |
| editorial guidelines to the decisions.”* (Urchs. et al., 2021) | |
| #### Who are the source language producers? | |
| German courts from Bavaria | |
| ### Annotations | |
| #### Annotation process | |
| *“As stated above, the judgment corpus consist of 200 randomly chosen judgments that are annotated by a legal expert, | |
| who holds a first legal state exam. Due to financial, staff and time reasons the presented iteration of the corpus was | |
| only annotated by a single expert. In a future version several other experts will annotate the corpus and the | |
| inter-annotator agreement will be calculated.”* (Urchs. et al., 2021) | |
| #### Who are the annotators? | |
| A legal expert, who holds a first legal state exam. | |
| ### Personal and Sensitive Information | |
| *"All decisions are processed by the publisher C.H.BECK, commissioned by the Bavarian state. This processing includes ** | |
| anonymisation**, key-wording, and adding of editorial guidelines to the decisions.”* (Urchs. et al., 2021) | |
| ## Considerations for Using the Data | |
| ### Social Impact of Dataset | |
| [More Information Needed] | |
| ### Discussion of Biases | |
| [More Information Needed] | |
| ### Other Known Limitations | |
| The SoMaJo Sentence Splitter has been used. Upon manual inspection of the dataset, we could see that the sentence | |
| splitter had poor accuracy in some cases (see ```analyze_dataset()``` in ```convert_to_hf_dataset.py```). When creating | |
| the splits, we thought about merging small sentences with their neighbors or removing them all together. However, since | |
| we could not find an straightforward way to do this, we decided to leave the dataset content untouched. | |
| Note that the information given in this dataset card refer to the dataset version as provided by Joel Niklaus and Veton | |
| Matoshi. The dataset at hand is intended to be part of a bigger benchmark dataset. Creating a benchmark dataset | |
| consisting of several other datasets from different sources requires postprocessing. Therefore, the structure of the | |
| dataset at hand, including the folder structure, may differ considerably from the original dataset. In addition to that, | |
| differences with regard to dataset statistics as give in the respective papers can be expected. The reader is advised to | |
| have a look at the conversion script ```convert_to_hf_dataset.py``` in order to retrace the steps for converting the | |
| original dataset into the present jsonl-format. For further information on the original dataset structure, we refer to | |
| the bibliographical references and the original Github repositories and/or web pages provided in this dataset card. | |
| ## Additional Information | |
| ### Dataset Curators | |
| The names of the original dataset curators and creators can be found in references given below, in the section *Citation | |
| Information*. Additional changes were made by Joel Niklaus ([Email](mailto:joel.niklaus.2@bfh.ch) | |
| ; [Github](https://github.com/joelniklaus)) and Veton Matoshi ([Email](mailto:veton.matoshi@bfh.ch) | |
| ; [Github](https://github.com/kapllan)). | |
| ### Licensing Information | |
| [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/legalcode) | |
| ### Citation Information | |
| ``` | |
| @dataset{urchs_stefanie_2020_3936490, | |
| author = {Urchs, Stefanie and | |
| Mitrović, Jelena}, | |
| title = {{German legal jugements annotated with judement | |
| style components}}, | |
| month = jul, | |
| year = 2020, | |
| publisher = {Zenodo}, | |
| doi = {10.5281/zenodo.3936490}, | |
| url = {https://doi.org/10.5281/zenodo.3936490} | |
| } | |
| ``` | |
| ``` | |
| @conference{icaart21, | |
| author = {Urchs., Stefanie and Mitrovi{\'{c}}., Jelena and Granitzer., Michael}, | |
| booktitle = {Proceedings of the 13th International Conference on Agents and Artificial Intelligence - Volume 2: ICAART,}, | |
| doi = {10.5220/0010187305150521}, | |
| isbn = {978-989-758-484-8}, | |
| issn = {2184-433X}, | |
| organization = {INSTICC}, | |
| pages = {515--521}, | |
| publisher = {SciTePress}, | |
| title = {{Design and Implementation of German Legal Decision Corpora}}, | |
| year = {2021} | |
| } | |
| ``` | |
| ### Contributions | |
| Thanks to [@kapllan](https://github.com/kapllan) and [@joelniklaus](https://github.com/joelniklaus) for adding this | |
| dataset. | |