prrao87's picture
Update README with LanceDB examples
37e28be verified
---
license: cc-by-4.0
task_categories:
- automatic-speech-recognition
- audio-classification
- text-retrieval
language:
- en
tags:
- librispeech
- asr
- audio
- speech
- lance
- sentence-transformers
pretty_name: librispeech-clean-lance
size_categories:
- 10K<n<100K
---
# LibriSpeech `clean` (Lance Format)
Lance-formatted version of the LibriSpeech ASR `clean` configuration (sourced from [`openslr/librispeech_asr`](https://huggingface.co/datasets/openslr/librispeech_asr)). Audio is stored inline as FLAC bytes (no re-encoding); transcripts are sentence-embedded so semantic transcript search works out of the box.
## Splits
| Split | Lance file | Rows | Description |
|-------|------------|------|-------------|
| `dev_clean.lance` | dev.clean | 2,703 | Standard ASR validation set |
| `test_clean.lance` | test.clean | 2,620 | Standard ASR test set |
| `train_clean_100.lance` | train.clean.100 | 28,539 | 100-hour clean training subset |
> The 360-hour and 500-hour LibriSpeech subsets (`train.360`, `train.other.500`) are **not** bundled here. To extend the dataset, point `librispeech/dataprep.py` at additional splits.
## Schema
| Column | Type | Notes |
|---|---|---|
| `id` | `string` | Utterance id (e.g. `1272-128104-0000`) |
| `audio` | `large_binary` | Inline FLAC bytes (16 kHz mono) |
| `sampling_rate` | `int32` | Always 16,000 |
| `text` | `string` | Reference transcript |
| `speaker_id` | `int64` | LibriVox speaker id |
| `chapter_id` | `int64` | LibriVox chapter id |
| `num_chars` | `int32` | Length of `text` in characters |
| `text_emb` | `fixed_size_list<float32, 384>` | sentence-transformers `all-MiniLM-L6-v2` (cosine-normalized) |
## Pre-built indices
- `IVF_PQ` on `text_emb``metric=cosine`
- `INVERTED` (FTS) on `text`
- `BTREE` on `id`, `speaker_id`, `chapter_id`
## Quick start
```python
import lance
ds = lance.dataset("hf://datasets/lance-format/librispeech-clean-lance/data/test_clean.lance")
print(ds.count_rows(), ds.schema.names, ds.list_indices())
```
## Load with LanceDB
These tables can also be consumed by [LanceDB](https://lancedb.github.io/lancedb/), the multimodal lakehouse and embedded search library built on top of Lance, for simplified vector search and other queries. Each `.lance` file in `data/` is a table — open by name (e.g., `test_clean`, `train_clean_100`).
```python
import lancedb
db = lancedb.connect("hf://datasets/lance-format/librispeech-clean-lance/data")
tbl = db.open_table("test_clean")
print(f"LanceDB table opened with {len(tbl)} utterances")
```
## Read one utterance and play it
```python
from pathlib import Path
import lance
ds = lance.dataset("hf://datasets/lance-format/librispeech-clean-lance/data/test_clean.lance")
row = ds.take([0], columns=["id", "audio", "text", "speaker_id"]).to_pylist()[0]
Path(f"{row['id']}.flac").write_bytes(row["audio"])
print("speaker:", row["speaker_id"])
print("transcript:", row["text"])
```
You can decode the FLAC bytes in-memory with `soundfile` and feed them straight into a model:
```python
import io
import soundfile as sf
samples, sr = sf.read(io.BytesIO(row["audio"]))
print(samples.shape, sr)
```
## Semantic transcript retrieval
```python
import lance
import pyarrow as pa
from sentence_transformers import SentenceTransformer
encoder = SentenceTransformer("sentence-transformers/all-MiniLM-L6-v2", device="cuda")
q = encoder.encode(["a person talking about astronomy"], normalize_embeddings=True)[0]
ds = lance.dataset("hf://datasets/lance-format/librispeech-clean-lance/data/train_clean_100.lance")
emb_field = ds.schema.field("text_emb")
hits = ds.scanner(
nearest={"column": "text_emb", "q": pa.array([q.tolist()], type=emb_field.type)[0], "k": 5},
columns=["id", "speaker_id", "text"],
).to_table().to_pylist()
for h in hits:
print(h)
```
### LanceDB semantic transcript retrieval
```python
import lancedb
from sentence_transformers import SentenceTransformer
encoder = SentenceTransformer("sentence-transformers/all-MiniLM-L6-v2", device="cuda")
q = encoder.encode(["a person talking about astronomy"], normalize_embeddings=True)[0]
db = lancedb.connect("hf://datasets/lance-format/librispeech-clean-lance/data")
tbl = db.open_table("train_clean_100")
results = (
tbl.search(q.tolist(), vector_column_name="text_emb")
.metric("cosine")
.select(["id", "speaker_id", "text"])
.limit(5)
.to_list()
)
```
## Full-text and per-speaker filtering
```python
ds = lance.dataset("hf://datasets/lance-format/librispeech-clean-lance/data/train_clean_100.lance")
# Word search via the FTS index.
hits = ds.scanner(full_text_query="universe stars", columns=["id", "text"], limit=10).to_table()
# All utterances by a given speaker.
sp = ds.scanner(filter="speaker_id = 1272", columns=["id", "chapter_id", "text"], limit=10).to_table()
```
### LanceDB full-text search and per-speaker filtering
```python
import lancedb
db = lancedb.connect("hf://datasets/lance-format/librispeech-clean-lance/data")
tbl = db.open_table("train_clean_100")
# Word search via the FTS index.
hits = (
tbl.search("universe stars")
.select(["id", "text"])
.limit(10)
.to_list()
)
# All utterances by a given speaker.
sp = (
tbl.search()
.where("speaker_id = 1272")
.select(["id", "chapter_id", "text"])
.limit(10)
.to_list()
)
```
## Why Lance?
- One dataset for audio + transcripts + embeddings + indices — no parallel folder of FLAC files plus a transcript JSON.
- On-disk vector and full-text indices live next to the data, so search works on local copies and on the Hub.
- Schema evolution: add columns (alternate transcripts, speaker embeddings, model predictions) without rewriting the data.
## Source & license
Converted from [`openslr/librispeech_asr`](https://huggingface.co/datasets/openslr/librispeech_asr). LibriSpeech is released under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) and is built from the public-domain LibriVox audiobook corpus.
## Citation
```
@inproceedings{panayotov2015librispeech,
title={LibriSpeech: An ASR corpus based on public domain audiobooks},
author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
booktitle={IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)},
year={2015}
}
```