Datasets:

Modalities:
Text
Formats:
parquet
ArXiv:
License:
ilacunza commited on
Commit
cc06892
·
verified ·
1 Parent(s): ea5923c

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -57,3 +57,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
60
+ acadbench/instruct_acadbench.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,283 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ task_categories:
4
+ - translation
5
+ language:
6
+ - es
7
+ - en
8
+ - ca
9
+ - pt
10
+ - fr
11
+ - eu
12
+ - gl
13
+ - de
14
+ - nl
15
+ - el
16
+ - it
17
+ size_categories:
18
+ - 1M<n<10M
19
+ configs:
20
+ - config_name: default
21
+ data_files:
22
+ - split: train
23
+ path: acadtrain.parquet
24
+ - split: test
25
+ path: acadbench/instruct_acadbench.json
26
+ ---
27
+
28
+ # Dataset Card for ACAData
29
+
30
+ ## Dataset Description
31
+
32
+ - **Point of Contact:** langtech@bsc.es
33
+
34
+
35
+ ### Dataset Summary
36
+
37
+ ACAData is a multilingual instruction tuning dataset containing parallel text paragraphs from the academic domain.
38
+
39
+
40
+ ### Supported Tasks and Leaderboards
41
+
42
+ The dataset is meant to be used for fine-tuning and benchmarking general purpose LLM's on Machine Translation tasks.
43
+
44
+ ### Languages
45
+
46
+ The dataset contains (mainly long) paragraph of scientific texts from the academic domain in many European language pairs.
47
+ The language coverage and distribution of the dataset is represented in the following tables. For further details, we refer to the paper [**add paper ref as soon as available*]() (to be published).
48
+
49
+
50
+ ## Dataset Structure
51
+
52
+ ACAData is composed of two different subsets: **ACAD-Train** and **ACAD-Bench**. The first is intended for training while the second serves as the benchmarking split.
53
+
54
+ **IMPORTANT**:
55
+
56
+ **ACAD-Train** is released in raw format as a Parquet file where each row contains a paragraph aligned across multiple languages, with one language per column, with a total number of **739,211** raw instances.
57
+ This corresponds to the dataset before conversion into the instruction format described in [CITE PAPER]. During conversion, each parallel pair is used to generate two instruction instances (one per translation direction), resulting in **1,478,422** training instances.
58
+
59
+ **ACAD-Bench**, on the other hand, is released directly in instruction format, where each pair has already been duplicated and swapped to cover both translation directions.
60
+ Total number of instances: **5,944**. ACAD-Bench is ready to be used for model evaluation, as detailed in the Evaluation section of [PUT PAPER HERE].
61
+
62
+
63
+ ### Data Instances
64
+
65
+ The key characteristics of ACAD-Train are the following:
66
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66703ae2d5c5d8fd8d999a44/0X5uUYLFYPok6r-5eNxe0.png)
67
+
68
+ The key characteristics of ACAD-Bench are the following:
69
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66703ae2d5c5d8fd8d999a44/fXuOo9P-ohZC1RDstXBpC.png)
70
+
71
+
72
+ The [`acadtrain.parquet` file](https://huggingface.co/datasets/LangTech-MT/ACAData/blob/main/acadtrain.parquet) has the following structure:
73
+
74
+
75
+ ```markdown
76
+ lang1_code lang2_code lang1 lang2 lang1_prob lang2_prob alignment
77
+ 0 ast ca Introducción al analisis forense con distribuc... Introducció a l'anàlisi forense... 0.982162 0.999984 0.963158
78
+ 1 ast ca Creación de un almacén de datos... Creació d'un magatzem de dades ... 0.847277 0.999991 0.990240
79
+ 2 ast ca Monografía ilustrada sobre la i... Monografia il·lustrada sobre la... 0.848437 0.990025 0.980378
80
+ 3 ast ca Entrevista con el escritor alba... Entrevista amb l'escriptor alba... 0.803880 0.979105 0.995741
81
+ 4 ast en Afondamos nesti trabayu con abo... Following the short essay Topon... 0.999985 0.983456 0.912953
82
+ ```
83
+
84
+ On the other hand, the instances of the [`instruct_acadbench.parquet` file](https://huggingface.co/datasets/LangTech-MT/ACAData/blob/main/acadbench/instruct_acadbench.json) have the following structure (Catalan → English example):
85
+
86
+ ```json
87
+ {
88
+ "id": "test_ca-en_abstract_dataset_{idx}",
89
+ "task": "abstract_dataset",
90
+ "lang": "ca-en",
91
+ "conversations": [
92
+ {
93
+ "from": "human",
94
+ "value": "Translate the following text from Catalan to English.\nCatalan: {lang1}\:"
95
+ },
96
+ {
97
+ "from": "gpt",
98
+ "value": "{lang2}"
99
+ }
100
+ ]
101
+ },
102
+ ```
103
+
104
+ In this case, given the human input, the model outputs the translation. The model's output is then compared against the target ({lang2}) for benchmarking.
105
+
106
+
107
+
108
+ ### Data Fields
109
+
110
+ - **lang1_code**: ISO language code of the text in **lang1** (the first text in the pair).
111
+ - **lang2_code**: ISO language code of the text in **lang2** (the second text in the pair).
112
+ - **lang1**: The first text in the bilingual instance.
113
+ - **lang2**: The second text in the bilingual instance.
114
+ - **lang1_prob**: Language identification probability for **lang1** (GlotLid).
115
+ - **lang2_prob**: Language identification probability for **lang2** (GlotLid).
116
+ - **alignment**: Cosine similarity between the embeddings of **lang1** and **lang2** (LaBSE).
117
+
118
+
119
+ ### Data Splits
120
+
121
+ The dataset contains two splits: `train`(ACAD-Train) and `benchmarking` (ACAD-Bench).
122
+
123
+ ## Dataset Creation
124
+
125
+ ### Curation Rationale
126
+
127
+ This dataset is aimed at improving the Machine Translation performance of LLM's in the academic domain.
128
+
129
+ ### Source Data
130
+
131
+ Translation pairs were harvested from the metadata of multiple European Academic repositories using the [OAI-PMH protocol](https://www.openarchives.org/OAI/openarchivesprotocol.html). For each harvested metadata record we extracted the textual content from the record’s "description" field and used those texts as the source for candidate segments.
132
+
133
+ #### Initial Data Collection and Normalization
134
+
135
+ Using OAI-PMH, we inspected each record’s description field to detect multiple entries. When multiple entries were present, we extracted embeddings for each entry with [LaBSE](https://huggingface.co/sentence-transformers/LaBSE), computed pairwise cosine similarities, and selected translation pairs with similarity ≥ 0.80. Language identification was then performed using [GlotLID](https://github.com/cisnlp/GlotLID).
136
+
137
+ For normalization, we applied preprocessing before embedding and language ID: stripped leading language markers (e.g., “(Spanish)”, “(eng)”); normalized punctuation and typography (converted all quotation marks and apostrophes to ASCII equivalents, replaced masculine ordinals “º” with degree symbols “°”, and converted superscript/subscript digits to regular digits); removed common inline markers (short bracketed/parenthesized codes, leading // or :); collapsed simple HTML tags; and collapsed repeated whitespace into single spaces. We also applied Unicode NFKC normalization and, where appropriate, lowercasing to ensure consistent tokenization and more stable embeddings.
138
+
139
+ #### Who are the source language producers?
140
+
141
+ In the following table, we provide a complete list of the source repositories from where the data were extracted (**the shown number of instances is before deduplication**).
142
+
143
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/646772c2b990713c5031e8f9/0fnlMtUMw8eFxLt78PZAs.png)
144
+
145
+ ### Annotations
146
+
147
+ #### Annotation process
148
+
149
+ The dataset does not contain any annotations.
150
+
151
+ #### Who are the annotators?
152
+
153
+ [N/A]
154
+
155
+ ### Personal and Sensitive Information
156
+
157
+ No specific anonymisation process has been applied. Personal and sensitive information might be present in the data. This needs to be considered when using the data for fine-tuning models.
158
+
159
+
160
+ ## Evaluation
161
+
162
+
163
+ Aggregated results for the XX ↔ EN and XX ↔ ES translation directions in ACAD-Bench dataset. Baselines are grouped into **large-scale proprietary general models**, **medium- to small-sized open-weights models** and **dedicated MMNMT models**. For every metric, the top-scoring system is shown in **bold**. For a more detailed evaluation analysis, please refer to the paper.
164
+
165
+
166
+ <details>
167
+ <summary>xx → en</summary>
168
+
169
+ | Direction | Model | d-BLEU | BP | Blonde | Comet | Comet-Kiwi |
170
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: |
171
+ | XX → EN | GPT-mini | 46.03 | **1.00** | 0.60 | **0.84** | 0.77 |
172
+ | | GPT-nano | 41.30 | 0.97 | 0.55 | **0.84** | **0.78** |
173
+ | | Gemini-2 | 48.65 | **1.00** | 0.61 | **0.84** | 0.77 |
174
+ | | Gemini-2.5 | 45.10 | 0.98 | 0.58 | **0.84** | 0.77 |
175
+ | | Llama-3-8B | 43.12 | 0.99 | 0.56 | 0.83 | 0.76 |
176
+ | | Gemma-3-27B | 46.37 | 0.98 | 0.59 | **0.84** | 0.77 |
177
+ | | MADLAD-7B | 38.69 | 0.86 | 0.51 | 0.81 | 0.77 |
178
+ | | Salamandra-2B | 37.09 | 0.92 | 0.52 | 0.82 | 0.75 |
179
+ | | &nbsp;&nbsp;+ ACADTRAIN | 48.45 | **1.00** | 0.61 | 0.83 | 0.76 |
180
+ | | Salamandra-7B | 45.87 | 0.99 | 0.59 | 0.83 | 0.76 |
181
+ | | &nbsp;&nbsp;+ ACADTRAIN | **50.07** | **1.00** | **0.62** | **0.84** | 0.76 |
182
+
183
+ </details>
184
+
185
+
186
+ <details>
187
+ <summary>en → xx</summary>
188
+
189
+ | Direction | Model | d-BLEU | BP | Blonde | Comet | Comet-Kiwi |
190
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: |
191
+ | EN → XX | GPT-mini | 45.01 | 0.99 | - | 0.86 | **0.82** |
192
+ | | GPT-nano | 43.78 | **1.00** | - | 0.86 | **0.82** |
193
+ | | Gemini-2 | 48.00 | 0.99 | - | **0.87** | **0.82** |
194
+ | | Gemini-2.5 | 47.75 | 0.99 | - | **0.87** | **0.82** |
195
+ | | Llama-3-8B | 39.87 | 0.99 | - | 0.85 | 0.81 |
196
+ | | Gemma-3-27B | 46.29 | 0.99 | - | 0.86 | **0.82** |
197
+ | | MADLAD-7B | 36.08 | 0.82 | - | 0.83 | 0.80 |
198
+ | | Salamandra-2B | 32.91 | 0.90 | - | 0.83 | 0.78 |
199
+ | | &nbsp;&nbsp;+ ACADTRAIN | 46.86 | 0.98 | - | 0.86 | 0.81 |
200
+ | | Salamandra-7B | 42.55 | 0.98 | - | 0.86 | 0.81 |
201
+ | | &nbsp;&nbsp;+ ACADTRAIN | **49.20** | 0.98 | - | 0.86 | 0.81 |
202
+
203
+ </details>
204
+
205
+
206
+ <details>
207
+ <summary>xx → es</summary>
208
+
209
+ | Direction | Model | d-BLEU | BP | Blonde | Comet | Comet-Kiwi |
210
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: |
211
+ | XX → ES | GPT-mini | 60.60 | 0.98 | - | 0.86 | **0.82** |
212
+ | | GPT-nano | 57.88 | **0.99** | - | 0.86 | **0.82** |
213
+ | | Gemini-2 | 62.02 | 0.99 | - | 0.86 | **0.82** |
214
+ | | Gemini-2.5 | 61.43 | 0.98 | - | **0.87** | **0.82** |
215
+ | | Llama-3-8B | 55.4 | 0.98 | - | 0.86 | 0.81 |
216
+ | | Gemma-3-27B | 60.71 | 0.98 | - | 0.86 | **0.82** |
217
+ | | MADLAD-7B | 43.44 | 0.76 | - | 0.83 | 0.81 |
218
+ | | Salamandra-2B | 50.09 | 0.92 | - | 0.85 | 0.80 |
219
+ | | &nbsp;&nbsp;+ ACADTRAIN | 61.97 | 0.98 | - | 0.86 | **0.82** |
220
+ | | Salamandra-7B | 57.55 | 0.98 | - | 0.86 | **0.82** |
221
+ | | &nbsp;&nbsp;+ ACADTRAIN | **63.60** | 0.98 | - | 0.86 | **0.82** |
222
+
223
+ </details>
224
+
225
+
226
+ <details>
227
+ <summary>es → xx</summary>
228
+
229
+ | Direction | Model | d-BLEU | BP | Blonde | Comet | Comet-Kiwi |
230
+ | :--- | :--- | :---: | :---: | :---: | :---: | :---: |
231
+ | ES → XX | GPT-mini | 54.19 | **0.99** | - | **0.86** | **0.81** |
232
+ | | GPT-nano | 51.95 | **0.99** | - | **0.86** | **0.81** |
233
+ | | Gemini-2 | 60.28 | **0.99** | - | **0.86** | **0.81** |
234
+ | | Gemini-2.5 | 57.61 | **0.99** | - | **0.86** | **0.81** |
235
+ | | Llama-3-8B | 52.12 | **0.99** | - | 0.85 | 0.80 |
236
+ | | Gemma-3-27B | 57.31 | **0.99** | - | **0.86** | **0.81** |
237
+ | | MADLAD-7B | 40.13 | 0.79 | - | 0.83 | **0.81** |
238
+ | | Salamandra-2B | 47.84 | 0.94 | - | 0.84 | 0.80 |
239
+ | | &nbsp;&nbsp;+ ACADTRAIN | 60.09 | **0.99** | - | **0.86** | **0.81** |
240
+ | | Salamandra-7B | 55.65 | 0.98 | - | **0.86** | 0.80 |
241
+ | | &nbsp;&nbsp;+ ACADTRAIN | **61.61** | **0.99** | - | **0.86** | **0.81** |
242
+
243
+ </details>
244
+
245
+
246
+
247
+ ## Considerations for Using the Data
248
+
249
+ ### Discussion of Biases
250
+
251
+ No specific bias mitigation strategies were applied to this dataset.
252
+ Inherent biases may exist within the data.
253
+
254
+ ### Other Known Limitations
255
+
256
+ The dataset contains data of the academic domain. Applications of this dataset in domains or languages not included in the training set would be of limited use.
257
+
258
+ ## Additional Information
259
+
260
+ ### Dataset Curators
261
+
262
+ Language Technologies Unit at the Barcelona Supercomputing Center (langtech@bsc.es).
263
+
264
+ ### Funding
265
+
266
+ This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the project Modelos del Lenguaje.
267
+
268
+ This work has been promoted and financed by the Government of Catalonia through the [Aina project](https://projecteaina.cat/).
269
+
270
+ This work is funded by the Ministerio para la Transformación Digital y de la Función Pública - Funded by EU – NextGenerationEU within the framework of the [project ILENIA](https://proyectoilenia.es/) with reference 2022/TL22/00215337.
271
+
272
+
273
+ ### Licensing Information
274
+
275
+ This work is licensed under an [Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/) license.
276
+
277
+ ### Citation Information
278
+
279
+ [N/A]
280
+
281
+ ### Contributions
282
+
283
+ [N/A]
acadbench/instruct_acadbench.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5702dd81dfd6ad43f9e279f30334457236077cec7d1a6685d28efa28a6614572
3
+ size 15716499
acadtrain.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:604f68c8857c2f4a945c7dc77755af18fa7f47d4a662d844827bff590e8ab4ee
3
+ size 938700155