File size: 2,060 Bytes
52f783f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
license: apache-2.0
task_categories:
- automatic-speech-recognition
language:
- uz
---
# Speech-to-Text Evaluation Dataset

## Dataset Overview

This dataset is designed for evaluating Uzbek speech-to-text (STT) models on real-world conversational speech data. The audio samples were collected from various open Telegram groups, capturing natural voice messages in diverse acoustic conditions and speaking styles.

### Key Statistics

- **Total Samples**: 745 audio files
- **Total Duration**: 1 hour 40 minutes (~100 minutes)
- **Average Duration**: ~8 seconds per sample
- **Source**: Voice messages from various open Telegram groups
- **Transcriptions**: Manually annotated

## Dataset Structure

The dataset is saved as a `datasets.Dataset` object in Arrow format, containing the following fields:

- `name`: Name of audio file
- `audio`: Audio file data (dict with `array`, and `sampling_rate`)
- `transcription`: Ground truth text transcription (manually annotated)

## Loading the Dataset

### Installation

To use this dataset, you need to install the Hugging Face `datasets` library:

```bash
pip install datasets
```

### Basic Loading

```python
from datasets import load_dataset

# Load the dataset from the Arrow files
dataset = load_dataset("OvozifyLabs/asr_evaluate_set")

# View dataset information
print(dataset)
print(f"Number of samples: {len(dataset)}")
```

## Data Characteristics

### Audio Properties

- **Source Domain**: Conversational voice messages from Telegram
- **Variability**: Multiple speakers, diverse acoustic environments
- **Recording Conditions**: Real-world
- **Language**: Uzbek

### Transcription Details

- **Annotation Method**: Manual transcription
- **Quality**: Human-verified ground truth labels
- **Convention**: punctuation removed, lowercased

## Use Cases

This dataset is suitable for:

- Evaluating speech-to-text model performance on conversational speech
- Benchmarking ASR systems on real-world voice messages
- Testing model robustness to varied acoustic conditions
- Comparing different STT models