File size: 5,000 Bytes
3f8a4d0
 
9d21390
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a5bb3ef
9d21390
 
a5bb3ef
9d21390
 
a5bb3ef
9d21390
 
a5bb3ef
9d21390
a5bb3ef
 
9d21390
 
 
 
 
 
 
 
 
 
 
3f8a4d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9d21390
3f8a4d0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
---
license: cc-by-4.0
dataset_info:
- config_name: librispeech_long
  features:
  - name: audio
    dtype:
      audio:
        sampling_rate: 16000
  - name: dataset
    dtype: string
  - name: text
    dtype: string
  - name: id
    dtype: string
  - name: audio_length_s
    dtype: float64
  splits:
  - name: dev_clean
    num_bytes: 1809224578.0
    num_examples: 295
  - name: dev_other
    num_bytes: 1070577217.0
    num_examples: 188
  - name: test_clean
    num_bytes: 1677717130.0
    num_examples: 270
  - name: test_other
    num_bytes: 1195726954.0
    num_examples: 207
  download_size: 5753245879
  dataset_size: 5753245879
configs:
- config_name: librispeech_long
  data_files:
  - split: dev_clean
    path: librispeech_long/dev_clean-*
  - split: dev_other
    path: librispeech_long/dev_other-*
  - split: test_clean
    path: librispeech_long/test_clean-*
  - split: test_other
    path: librispeech_long/test_other-*
---

# LibriSpeech-Long

LibriSpeech-Long is a benchmark dataset for long-form speech generation and processing. Released as part of "Long-Form Speech Generation with Spoken Language Models" (arXiv 2024).

## Prerequisites

1. [uv](https://docs.astral.sh/uv/)

## Development

1. Clone the repo
    ```sh
    git clone git@hf.co:datasets/ilyakam/librispeech-long && cd librispeech-long
    ```

1. Activate the virtual environment:
    ```sh
    source .venv/bin/activate
    ```

1. Install dependencies:
    ```sh
    uv pip install -r requirements.txt
    ```

1. Process the dataset:
    ```sh
    python process.py [SOURCE_ROOT] . --limit-speakers
    ```

    Notes:
    - The `SOURCE_ROOT` is a folder where you've extracted the dataset containing the splits `dev-clean`, `dev-other`, `test-clean`, and `test-other`.
    - The optional `--limit-speakers` flag will only process the first speaker directory in each split for quick testing.

1. Copy the metadata to the top of this `README.md` file.

1. Commit and push the generated `*.parquet` files and the updated `README.md` file.
    

## Purpose

To migrate the [google-deepmind/librispeech-long](https://github.com/google-deepmind/librispeech-long) dataset to Hugging Face datasets.

# LibriSpeech-Long (Original)

**[Download audio, ground-truth transcripts, and per-file durations for all splits (3GB).](https://storage.googleapis.com/librispeech_long/v0_1.tar.gz)** 

This is a benchmark dataset for evaluating long-form variants of speech processing tasks such as speech continuation, speech recognition, and text-to-speech synthesis. It is derived from the [LibriSpeech](https://www.openslr.org/12) dev and test sets, whose utterances are reprocessed into contiguous examples of up to 4 minutes in length (in the manner of [LibriLight's `cut_by_vad.py` script](https://github.com/facebookresearch/libri-light/blob/main/data_preparation/cut_by_vad.py)).

**For more details:** see "Long-Form Speech Generation with Spoken Language Models" ([paper](https://arxiv.org/abs/2412.18603), [website](https://google.github.io/tacotron/publications/speechssm/)).

*This is part of a preprint that is work-in-progress; dataset may be subject to change.*

## Citation

When using this dataset, please cite the associated [paper](https://arxiv.org/abs/2412.18603):

```
@article{park2024long,
  author       = {Se Jin Park and
                  Julian Salazar and
                  Aren Jansen and
                  Keisuke Kinoshita and
                  Yong Man Ro and
                  R. J. Skerry{-}Ryan},
  title        = {Long-Form Speech Generation with Spoken Language Models},
  journal      = {CoRR},
  volume       = {abs/2412.18603},
  year         = {2024}
}
```

## License and disclaimer

Copyright 2024 DeepMind Technologies Limited

The software and materials, except for the underlying LibriSpeech data, are licensed under the Creative Commons Attribution 4.0 International License (CC-BY). You may obtain a copy of the CC-BY license at: https://creativecommons.org/licenses/by/4.0/legalcode, or in the LICENSE file.

The materials contain adapted material from the LibriSpeech dataset. LibriSpeech is also licensed under the Creative Commons Attribution 4.0 International License (CC-BY). You may obtain a copy of the CC-BY license at: https://creativecommons.org/licenses/by/4.0/legalcode, or in the LICENSE file. LibriSpeech is available at https://www.openslr.org/12 and created by Vassil Panayotov, Guoguo Chen, Daniel Povey and Sanjeev Khudanpur, pursuant to the paper “LibriSpeech: an ASR corpus based on public domain audio books", ICASSP 2015 (https://ieeexplore.ieee.org/document/7178964).

Unless required by applicable law or agreed to in writing, all software and materials distributed here under the CC-BY licenses are distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the licenses for the specific language governing permissions and limitations under those licenses.

This is not an official Google product.