drelhaj commited on
Commit
d1849db
·
verified ·
1 Parent(s): 4f91cdf

Upload 11 files

Browse files
.gitattributes CHANGED
@@ -57,3 +57,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
 
 
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
60
+ single_docs_train.csv filter=lfs diff=lfs merge=lfs -text
61
+ single_docs.csv filter=lfs diff=lfs merge=lfs -text
62
+ single_docs.jsonl filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,178 @@
1
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: cc-by-4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - ar
4
+ - zh
5
+ - cs
6
+ - en
7
+ - fr
8
+ - el
9
+ - he
10
+ - hi
11
+ - ro
12
+ - es
13
+
14
+ pretty_name: "MultiLing Multilingual Summarisation Corpus (Single- and Multi-Document)"
15
+ tags:
16
+ - summarisation
17
+ - multi-document
18
+ - multilingual
19
+ - abstractive
20
+ - news
21
+ - evaluation
22
+ task_categories:
23
+ - summarization
24
  license: cc-by-4.0
25
+ size_categories:
26
+ - 1K<n<10K
27
+ dataset_info:
28
+ features:
29
+ - name: language
30
+ dtype: string
31
+ - name: doc_id
32
+ dtype: string
33
+ - name: documents_text
34
+ dtype: string
35
+ - name: summary
36
+ dtype: string
37
+ - name: summary_1
38
+ dtype: string
39
+ - name: summary_2
40
+ dtype: string
41
+ - name: summary_3
42
+ dtype: string
43
+ - name: doc_ids
44
+ dtype: string
45
+ splits:
46
+ - train
47
+ - validation
48
+ - test
49
+ task_templates:
50
+ - text2text-generation
51
  ---
52
+
53
+ # MultiLing Multilingual Summarisation Corpus
54
+
55
+ The **MultiLing Multilingual Summarisation Corpus** is a comprehensive multilingual benchmark for **single-document** and **multi-document abstractive summarisation**, originally created for the *MultiLing 2011* and *MultiLing 2013* shared tasks held under ACL.
56
+ This release consolidates, cleans, and reformats the original resources into a standard, machine-readable dataset suitable for modern sequence-to-sequence and large language model research.
57
+
58
+ The corpus covers **ten languages**:
59
+
60
+ **Arabic, Chinese, Czech, English, French, Greek, Hebrew, Hindi, Romanian, Spanish**
61
+
62
+ and contains both:
63
+
64
+ - **Single-document summarisation pairs** (source article, 1 gold summary)
65
+ - **Multi-document summarisation clusters** (10 related articles, 3 human-written abstractive summaries)
66
+
67
+ All texts originate from **WikiNews**, following the Creative Commons BY licence. See http://multiling.iit.demokritos.gr/
68
+
69
+ ---
70
+
71
+ ## ✨ Key Features
72
+
73
+ ### **Single-Document Summarisation**
74
+ - 40 languages in the original collection; the present release includes the main ten used for MultiLing 2013.
75
+ - For each language, every document has:
76
+ - One source article
77
+ - One human abstractive summary
78
+ - Fully parallel across languages.
79
+
80
+ ### **Multi-Document Summarisation**
81
+ - Each topic consists of **10 articles describing an event sequence**.
82
+ - Topics appear consistently across all languages that contributed to that year’s task.
83
+ - Each cluster includes **three human-written abstractive summaries**, produced independently.
84
+ - Human summaries were constrained to **240–250 words** (or equivalent byte limits for Chinese).
85
+
86
+ ### **Parallel & Comparable Structure**
87
+ The corpus was originally designed to allow:
88
+ - Cross-lingual and multilingual summarisation
89
+ - Comparative analyses of summarisation difficulty across languages
90
+ - Multilingual evaluation of automatic summarisation metrics (ROUGE, AutoSummENG-MeMoG, NPowER)
91
+
92
+ ---
93
+
94
+ ## 📘 Source and Citation
95
+
96
+ This dataset is derived from the corpus described in:
97
+
98
+ **Li L, Forăscu C, El-Haj M, Giannakopoulos G. (2013)**
99
+ *Multi-Document Multilingual Summarization Corpus Preparation, Part 1: Arabic, English, Greek, Chinese, Romanian.*
100
+ In **Proceedings of the MultiLing 2013 Workshop on Multilingual Multi-Document Summarization**, pp. 1–12.
101
+ ACL 2013, Sofia, Bulgaria.
102
+ PDF: https://aclanthology.org/W13-3101.pdf
103
+
104
+ Please cite the paper above when using this dataset.
105
+
106
+ ---
107
+
108
+ ## 🗂 Dataset Structure
109
+
110
+ ### **Single-Document Format**
111
+ Each sample includes:
112
+
113
+ - `language` — ISO folder name (`ar`, `en`, `fr`, etc.)
114
+ - `doc_id`
115
+ - `document_text`
116
+ - `summary`
117
+
118
+ ### **Multi-Document Format**
119
+ Each sample includes:
120
+
121
+ - `cluster_id` — e.g. `M000`, `M014`, `M103`
122
+ - `language`
123
+ - `doc_ids` — list of the ten document identifiers
124
+ - `documents_text` — concatenated with `<DOC id=…>` wrappers
125
+ - `summary_1`, `summary_2`, `summary_3` — three reference human summaries
126
+
127
+ All files are provided in **CSV** and **JSONL**, with **train/dev/test** splits.
128
+
129
+ ---
130
+
131
+ ## 🔧 Recommended Use Cases
132
+
133
+ - Multilingual abstractive summarisation
134
+ - Cross-lingual evaluation of LLMs
135
+ - Multi-document summarisation research
136
+ - Training summarisation models on parallel news texts
137
+ - Research on multilingual evaluation metrics
138
+ - Cross-lingual transfer learning
139
+ - Low-resource summarisation investigations
140
+
141
+ ---
142
+
143
+ ## 📊 Splits
144
+
145
+ The dataset is released with deterministic:
146
+
147
+ - `train`
148
+ - `validation`
149
+ - `test`
150
+
151
+ splits for **single-document** and **multi-document** subsets.
152
+ For multi-document summarisation, splits are **cluster-based** to prevent data leakage.
153
+
154
+ ---
155
+
156
+ ## 📥 Loading the Dataset
157
+
158
+ ```python
159
+ from datasets import load_dataset
160
+
161
+ ds = load_dataset("YOUR_DATASET_NAME", "multi")
162
+ # or
163
+ ds = load_dataset("YOUR_DATASET_NAME", "single")
164
+
165
+
166
+ ---
167
+
168
+ ## 🔒 Licence
169
+
170
+ All texts originate from WikiNews under Creative Commons BY 2.5/3.0 licences.
171
+ This consolidated dataset is released under CC-BY-4.0.
172
+
173
+ ---
174
+
175
+ ## 🙏 Acknowledgements
176
+
177
+ MultiLing is the result of a large international community effort involving contributors from more than ten universities and research centres.
178
+ This cleaned and repackaged release builds on that original work to make the corpus more accessible for modern NLP research.
multi_docs.csv ADDED
The diff for this file is too large to render. See raw diff
 
multi_docs.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
multi_docs_dev.csv ADDED
The diff for this file is too large to render. See raw diff
 
multi_docs_test.csv ADDED
The diff for this file is too large to render. See raw diff
 
multi_docs_train.csv ADDED
The diff for this file is too large to render. See raw diff
 
single_docs.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9feed643987b5faa0acca021227faacf48c350f7a85d12191b8b0ee4d215f30
3
+ size 40992719
single_docs.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b8a6717cee67d62b45f4834a19b9aeeacee906a6b62ec3112fe54c867171faf
3
+ size 41147596
single_docs_dev.csv ADDED
The diff for this file is too large to render. See raw diff
 
single_docs_test.csv ADDED
The diff for this file is too large to render. See raw diff
 
single_docs_train.csv ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dd8d408fe510b844839e398cd42524c2a7397ad50d5dec6453529bfd947bd36
3
+ size 32795647