Datasets:

Modalities:
Audio
Text
Formats:
parquet
ArXiv:
lau-eval / README.md
diarray's picture
Update README.md
a8f05b0 verified
---
dataset_info:
features:
- name: audio
dtype: audio
- name: duration
dtype: float64
- name: bam
dtype: string
- name: french
dtype: string
- name: asr-ctc
dtype: string
- name: asr-tdt
dtype: string
- name: asr-mt-ctc
dtype: string
- name: asr-mt-tdt
dtype: string
- name: st-ctc
dtype: string
- name: st-tdt
dtype: string
- name: lau-tdt-k1
dtype: string
- name: lau-ctc-k1
dtype: string
- name: lau-tdt-k5
dtype: string
- name: lau-ctc-k5
dtype: string
- name: lau-tdt-k0.2
dtype: string
- name: lau-ctc-k0.2
dtype: string
- name: lau-tdt-mse-k1
dtype: string
- name: lau-ctc-mse-k1
dtype: string
- name: cluster_label
dtype: string
splits:
- name: test
num_bytes: 121749438
num_examples: 1218
download_size: 117816418
dataset_size: 121749438
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
---
# LAU eval dataset
This dataset was created while evaluating and comparing the models trained with Listen Attend Understand regularization and our [E2E-ST model](https://huggingface.co/RobotsMali/st-soloni-114m-tdt-ctc).
The audio is from jeli-asr test set; the regularization loss weight lambda in the paper is represented by the character "k" in the fields of this dataset, each field represent a model with a specific decoding strategy (CTC or TDT)
---
## Citation
```bibtex
@misc{diarra2026listenattendunderstandregularization,
title={Listen, Attend, Understand: a Regularization Technique for Stable E2E Speech Translation Training on High Variance labels},
author={Yacouba Diarra and Michael Leventhal},
year={2026},
eprint={2601.01121},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2601.01121},
}
```