Datasets:
File size: 5,389 Bytes
f18332f 01b759e 52027c2 01b759e f18332f 9d429e2 621458c d71a742 0362f64 d71a742 9d429e2 f18332f 8687ea6 f18332f 9d429e2 d71a742 f18332f 01b759e f18332f 01b759e 11a98bf 01b759e 11a98bf 01b759e 11a98bf 01b759e 52027c2 01b759e 52027c2 01b759e 52027c2 01b759e 11a98bf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 | ---
annotations_creators:
- expert-annotated
language:
- nld
license: cc-by-nc-sa-4.0
multilinguality: monolingual
source_datasets:
- clips/mteb-nl-bbsard
task_categories:
- text-retrieval
task_ids: []
dataset_info:
- config_name: corpus
features:
- name: id
dtype: string
- name: text
dtype: string
- name: title
dtype: string
splits:
- name: test
num_bytes: 21656574
num_examples: 22415
download_size: 8789748
dataset_size: 21656574
- config_name: qrels
features:
- name: query-id
dtype: string
- name: corpus-id
dtype: string
- name: score
dtype: int64
splits:
- name: test
num_bytes: 24338
num_examples: 1059
download_size: 6847
dataset_size: 24338
- config_name: queries
features:
- name: id
dtype: string
- name: text
dtype: string
splits:
- name: test
num_bytes: 23204
num_examples: 222
download_size: 13641
dataset_size: 23204
configs:
- config_name: corpus
data_files:
- split: test
path: corpus/test-*
- config_name: qrels
data_files:
- split: test
path: qrels/test-*
- config_name: queries
data_files:
- split: test
path: queries/test-*
tags:
- mteb
- text
---
<!-- adapted from https://github.com/huggingface/huggingface_hub/blob/v0.30.2/src/huggingface_hub/templates/datasetcard_template.md -->
<div align="center" style="padding: 40px 20px; background-color: white; border-radius: 12px; box-shadow: 0 2px 10px rgba(0, 0, 0, 0.05); max-width: 600px; margin: 0 auto;">
<h1 style="font-size: 3.5rem; color: #1a1a1a; margin: 0 0 20px 0; letter-spacing: 2px; font-weight: 700;">bBSARDNLRetrieval</h1>
<div style="font-size: 1.5rem; color: #4a4a4a; margin-bottom: 5px; font-weight: 300;">An <a href="https://github.com/embeddings-benchmark/mteb" style="color: #2c5282; font-weight: 600; text-decoration: none;" onmouseover="this.style.textDecoration='underline'" onmouseout="this.style.textDecoration='none'">MTEB</a> dataset</div>
<div style="font-size: 0.9rem; color: #2c5282; margin-top: 10px;">Massive Text Embedding Benchmark</div>
</div>
Building on the Belgian Statutory Article Retrieval Dataset (BSARD) in French, we introduce the bilingual version of this dataset, bBSARD. The dataset contains parallel Belgian statutory articles in both French and Dutch, along with legal questions from BSARD and their Dutch translation.
| | |
|---------------|---------------------------------------------|
| Task category | t2t |
| Domains | Legal, Written |
| Reference | https://aclanthology.org/2025.regnlp-1.3.pdf |
Source datasets:
- [clips/bBSARD](https://huggingface.co/datasets/clips/bBSARD)
## How to evaluate on this task
You can evaluate an embedding model on this dataset using the following code:
```python
import mteb
task = mteb.get_task("bBSARDNLRetrieval")
evaluator = mteb.MTEB([task])
model = mteb.get_model(YOUR_MODEL)
evaluator.run(model)
```
<!-- Datasets want link to arxiv in readme to autolink dataset with paper -->
To learn more about how to run models on `mteb` task check out the [GitHub repository](https://github.com/embeddings-benchmark/mteb).
## Citation
If you use this dataset, please cite the dataset as well as MTEB-NL, as this dataset includes additional processing.
```bibtex
@article{lotfi2025bilingual,
author = {Lotfi, Ehsan and Banar, Nikolay and Yuzbashyan, Nerses and Daelemans, Walter},
journal = {COLING 2025},
pages = {10},
title = {Bilingual BSARD: Extending Statutory Article Retrieval to Dutch},
year = {2025},
}
@misc{banar2025mtebnle5nlembeddingbenchmark,
title={MTEB-NL and E5-NL: Embedding Benchmark and Models for Dutch},
author={Nikolay Banar and Ehsan Lotfi and Jens Van Nooten and Cristina Arhiliuc and Marija Kliocaite and Walter Daelemans},
year={2025},
eprint={2509.12340},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2509.12340},
}
```
# Dataset Statistics
<details>
<summary> Dataset Statistics</summary>
The following code contains the descriptive statistics from the task. These can also be obtained using:
```python
import mteb
task = mteb.get_task("bBSARDNLRetrieval")
desc_stats = task.metadata.descriptive_stats
```
```json
{
"test": {
"num_samples": 22637,
"number_of_characters": 21218611,
"documents_text_statistics": {
"total_text_length": 21197901,
"min_text_length": 7,
"average_text_length": 945.7015837608744,
"max_text_length": 37834,
"unique_texts": 22415
},
"documents_image_statistics": null,
"queries_text_statistics": {
"total_text_length": 20710,
"min_text_length": 22,
"average_text_length": 93.28828828828829,
"max_text_length": 250,
"unique_texts": 222
},
"queries_image_statistics": null,
"relevant_docs_statistics": {
"num_relevant_docs": 1059,
"min_relevant_docs_per_query": 1,
"average_relevant_docs_per_query": 4.77027027027027,
"max_relevant_docs_per_query": 57,
"unique_relevant_docs": 491
},
"top_ranked_statistics": null
}
}
```
</details>
--- |