Datasets:
File size: 5,222 Bytes
6a014f3 d559e76 a670f2c e0f03e3 6a014f3 45a1bd9 6a014f3 4d8f650 6a014f3 187b264 190a22b 187b264 6a014f3 4e63031 6a014f3 4e63031 c8b8e34 6a014f3 4e63031 c8b8e34 4e63031 6a014f3 187b264 6a014f3 caa87ec fa5195b caa87ec 6a014f3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 |
---
tags:
- speech
- speech-transcription
- romanian
language:
- ro
license: mit
task_categories:
- automatic-speech-recognition
- audio-classification
- text-to-speech
- text-to-audio
pretty_name: RO_CV20
size_categories:
- 10K<n<100K
---
# Common Voices Corpus 20.0 (Romanian)
<h5 style="font-family: 'Calibri'; margin-bottom: 20px;">
Common Voices is an open-source dataset of speech recordings created by
<a href="https://commonvoice.mozilla.org" target="_blank">Mozilla</a> to improve speech recognition technologies.
It consists of crowdsourced voice samples in multiple languages, contributed by volunteers worldwide.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 20px;">
<strong>Challenges:</strong> The raw dataset included numerous recordings with incorrect transcriptions
or those requiring adjustments, such as sampling rate modifications, conversion to .wav format, and other refinements
essential for optimal use in developing and fine-tuning various models.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 20px;">
<strong>Processing:</strong> Our team, led by project manager Ionuț Vișan, carefully reviewed and manually corrected the
transcriptions of all audio segments, ensuring their conversion into the required format for modern models
(16k Hz sampling rate, mono channel, .wav format).
</h5>
---
<h2>Dataset Summary<h2>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<strong>common_voices20_audio.zip: </strong> The archive containing all processed audio segments.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
Total number of audio segments: <strong>41,431</strong>.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 20px;">
Total duration of all audio segments combined: approximately <strong>47 hours</strong>.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<strong>common_voices20.csv: </strong> Contains metadata for all segments from the common_voices20_audio.
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
The file contains 41,431 rows and 2 columns:
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<ul>
<li><em>audio_file</em>: File names of the processed audio segments from common_voices20_audio.</li>
<li><em>transcript</em>: Corresponding text transcriptions for each audio file from common_voices20_audio.</li>
</ul>
</h5>
---
<h2>Split<h2>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
To split the dataset (common_voices20.csv), we performed an 80-20 split into training and test sets using a seed value of 42, resulting in:
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<ul>
<li><em>train_common_voices20.csv</em>: It contains 33,144 of the audio segments.</li>
<li><em>test_common_voices20.csv</em>: It contains 8,287 of the audio segments.</li>
</ul>
</h5>
---
<h2>How to use<h2>
<h5 style="font-family: 'Calibri'; margin-bottom: 20px;">
<strong></strong>If you want to download all the files from the dataset, use the following code:
</h5>
<details>
<summary><strong>Click to expand the code</strong></summary>
```python
from huggingface_hub import hf_hub_download
import zipfile
import os
# Repo and files
Dataset = "TransferRapid/CommonVoices20_ro"
filenames = [
"common_voices20.csv",
"test_common_voices20.csv",
"train_common_voices20.csv",
"common_voices20_audio.zip"
]
# Download files
for filename in filenames:
print(f"Downloading {filename}...")
file_path = hf_hub_download(repo_id=Dataset, filename=filename, repo_type="dataset", local_dir="./")
print(f"Downloaded {filename} to: {file_path}")
# Extract ZIP files
if filename.endswith('.zip'):
extracted_dir = filename.replace('.zip', '')
with zipfile.ZipFile(file_path, 'r') as zip_ref:
zip_ref.extractall(extracted_dir)
print(f"Extracted files to: {extracted_dir}")
print(os.listdir(extracted_dir))
else:
print(f"{filename} is available.")
```
</details>
---
<h2>Usage<h2>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
The dataset can be used for:
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<ul>
<li><em>Speech-to-Text (STT) – Automatic Transcription</em></li>
<li><em>Text-to-Speech (TTS) – Synthetic Voice Generation</em></li>
<li><em>Speech Enhancement & Noise Reduction</em></li>
<li><em>Speaker Recognition & Verification</em></li>
<li><em>Sentiment Analysis & Emotion Recognition</em></li>
<li><em>AI-Powered Voice Assistants & Smart Devices</em></li>
</ul>
</h5>
---
<h2>Communication<h2>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
For any questions regarding this dataset or to explore collaborations on ambitious AI/ML projects, please feel free to contact us at:
</h5>
<h5 style="font-family: 'Calibri'; margin-bottom: 2px;">
<ul>
<li><em>ionut.visan@transferrapid.com</em></li>
<li><em><a href="https://www.linkedin.com/in/ionut-visan/" target="_blank">Ionuț Vișan's Linkedin</a></em></li>
<li><em><a href="https://www.linkedin.com/company/transfer-rapid" target="_blank">Transfer Rapid's Linkedin</a></em></li>
</ul>
</h5>
</ul>
</h5> |