Datasets:

Modalities:
Text
Formats:
json
Libraries:
Datasets
pandas
License:
File size: 7,380 Bytes
d6de700
76816d3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38afaee
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
76816d3
 
 
 
 
 
 
a488551
 
76816d3
 
 
 
 
3816ccb
 
 
 
 
76816d3
 
 
3816ccb
 
 
 
 
 
 
 
 
 
 
76816d3
a488551
76816d3
 
 
 
 
 
3816ccb
 
76816d3
3816ccb
76816d3
3816ccb
76816d3
 
3816ccb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
---
license: cc-by-sa-4.0
task_categories:
  - text-generation
language:
  - en
  - ar
  - de
  - es
  - fr
  - it
  - ja
  - ko
  - th
  - tr
  - zh
size_categories:
  - 10K<n<100K
configs:
- config_name: en-it
  data_files:
    - split: sample
      path: data/sample/it_IT.jsonl
    - split: validation
      path: data/validation/it_IT.jsonl
    - split: test
      path: data/test/it_IT.jsonl
- config_name: en-ar
  data_files:
    - split: sample
      path: data/sample/ar_AE.jsonl
    - split: validation
      path: data/validation/ar_AE.jsonl
    - split: test
      path: data/test/ar_AE.jsonl
- config_name: en-de
  data_files:
    - split: sample
      path: data/sample/de_DE.jsonl
    - split: validation
      path: data/validation/de_DE.jsonl
    - split: test
      path: data/test/de_DE.jsonl
- config_name: en-es
  data_files:
    - split: sample
      path: data/sample/es_ES.jsonl
    - split: validation
      path: data/validation/es_ES.jsonl
    - split: test
      path: data/test/es_ES.jsonl
- config_name: en-fr
  data_files:
    - split: sample
      path: data/sample/fr_FR.jsonl
    - split: validation
      path: data/validation/fr_FR.jsonl
    - split: test
      path: data/test/fr_FR.jsonl
- config_name: en-ja
  data_files:
    - split: sample
      path: data/sample/ja_JP.jsonl
    - split: validation
      path: data/validation/ja_JP.jsonl
    - split: test
      path: data/test/ja_JP.jsonl
- config_name: en-ko
  data_files:
    - split: sample
      path: data/sample/ko_KR.jsonl
    - split: validation
      path: data/validation/ko_KR.jsonl
    - split: test
      path: data/test/ko_KR.jsonl
- config_name: en-th
  data_files:
    - split: sample
      path: data/sample/th_TH.jsonl
    - split: validation
      path: data/validation/th_TH.jsonl
    - split: test
      path: data/test/th_TH.jsonl
- config_name: en-tr
  data_files:
    - split: sample
      path: data/sample/tr_TR.jsonl
    - split: validation
      path: data/validation/tr_TR.jsonl
    - split: test
      path: data/test/tr_TR.jsonl
- config_name: en-zh
  data_files:
    - split: sample
      path: data/sample/zh_TW.jsonl
    - split: validation
      path: data/validation/zh_TW.jsonl
    - split: test
      path: data/test/zh_TW.jsonl
---

# Dataset Card for EA-MT

EA-MT (Entity-Aware Machine Translation) is a multilingual benchmark for evaluating the capabilities of Large Language Models (LLMs) and Machine Translation (MT) models in translating simple sentences with potentially challenging entity mentions, e.g., entities for which a word-for-word translation may not be accurate.

Here is an example of a simple sentence with a challenging entity mention:
* *English*: "What is the plot of **The Catcher in the Rye**?"
* *Italian*:
  * Word-for-word translation (incorrect): "Qual è la trama del **Cacciatore nella segale**?"
  * Correct translation: "Qual è la trama de **Il giovane Holden**?"

> Note: In the example above, the correct translation of "The Catcher in the Rye" is "Il giovane Holden" in Italian, which roughly translates to "The Young Holden."

You can find more information about this task here:
* Paper (coming soon!)
* [Website](https://sapienzanlp.github.io/ea-mt/)
* [Leaderboard](https://huggingface.co/spaces/sapienzanlp/ea-mt-leaderboard)

## Languages

The dataset is available in the following languages pairs:
- `en-ar`: English - Arabic
- `en-zh`: English - Chinese
- `en-fr`: English - French
- `en-de`: English - German
- `en-it`: English - Italian
- `en-ja`: English - Japanese
- `en-ko`: English - Korean
- `en-es`: English - Spanish
- `en-th`: English - Thai
- `en-tr`: English - Turkish
- `en-zh`: English - Chinese (Traditional)

## How To Use

You can use this benchmark in Hugging Face Datasets by specifying the language pair you want to use. For example, to load the English-Italian dataset, you can use the following configuration:

```python
from datasets import load_dataset

# Load the English-Italian dataset ("en-it")
dataset = load_dataset("sapienzanlp/ea-mt-benchmark", "en-it")

# Iterate over the "sample" split and print the source sentence and the first target translation.
for example in dataset["sample"]:
    print(example["source"])
    print(example["targets"][0])
    print()
```

This will load the English-Italian dataset and print the source sentence and the target translation.

### Data format

The dataset is available in the following splits:
* `sample`: A small sample of the dataset for quick testing and debugging.
* `validation`: A validation set for fine-tuning and hyperparameter tuning.
* `test`: A test set for evaluating the model's performance.

Each example in the dataset has the following format:
```json
{
  "id": "Q1422318_1",
  "wikidata_id": "Q1422318",
  "entity_types": [
    "Artwork",
    "Book"
  ],
  "source": "Who is the author of the novel The Dark Tower: The Gunslinger?",
  "targets": [
    {
      "translation": "Chi è l'autore del romanzo L'ultimo cavaliere?",
      "mention": "L'ultimo cavaliere"
    }
  ],
  "source_locale": "en",
  "target_locale": "it"
}
```

Each example contains the following fields:
- `id`: A unique identifier for the example.
- `wikidata_id`: The Wikidata ID of the entity mentioned in the source sentence.
- `entity_types`: The types of the entity mentioned in the source sentence.
- `source`: The source sentence in English.
- `targets`: A list of target translations in the target language. Each target translation contains the following fields:
  - `translation`: The target translation.
  - `mention`: The entity mention in the target translation.
- `source_locale`: The source language code.
- `target_locale`: The target language code.

> Note: This is a multi-reference translation dataset, meaning that each example has multiple valid translations. The translations are provided as a list of target translations in the `targets` field. A model's output is considered correct if it generates any of the valid translations for a given example.

## License

The dataset is released under the [Creative Commons Attribution-ShareAlike 4.0 International License](https://creativecommons.org/licenses/by-sa/4.0/).

## Citation

If you use this benchmark in your work, please cite the following papers:

```bibtex
@inproceedings{ea-mt-benchmark,
    title       = "{S}em{E}val-2025 Task 2: Entity-Aware Machine Translation",
    author      = "Simone Conia and Min Li and Roberto Navigli and Saloni Potdar",
    booktitle   = "Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)",
    year        = "2025",
    publisher   = "Association for Computational Linguistics",
}
```

```bibtex
@inproceedings{conia-etal-2024-towards,
    title = "Towards Cross-Cultural Machine Translation with Retrieval-Augmented Generation from Multilingual Knowledge Graphs",
    author = "Conia, Simone  and
      Lee, Daniel  and
      Li, Min  and
      Minhas, Umar Farooq  and
      Potdar, Saloni  and
      Li, Yunyao",
    booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2024",
    address = "Miami, Florida, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.emnlp-main.914/",
    doi = "10.18653/v1/2024.emnlp-main.914",
    pages = "16343--16360",
}
```