parquet-converter commited on
Commit
33fe9b5
·
1 Parent(s): b0a274f

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,51 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.lz4 filter=lfs diff=lfs merge=lfs -text
11
- *.model filter=lfs diff=lfs merge=lfs -text
12
- *.msgpack filter=lfs diff=lfs merge=lfs -text
13
- *.npy filter=lfs diff=lfs merge=lfs -text
14
- *.npz filter=lfs diff=lfs merge=lfs -text
15
- *.onnx filter=lfs diff=lfs merge=lfs -text
16
- *.ot filter=lfs diff=lfs merge=lfs -text
17
- *.parquet filter=lfs diff=lfs merge=lfs -text
18
- *.pb filter=lfs diff=lfs merge=lfs -text
19
- *.pickle filter=lfs diff=lfs merge=lfs -text
20
- *.pkl filter=lfs diff=lfs merge=lfs -text
21
- *.pt filter=lfs diff=lfs merge=lfs -text
22
- *.pth filter=lfs diff=lfs merge=lfs -text
23
- *.rar filter=lfs diff=lfs merge=lfs -text
24
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
25
- *.tar.* filter=lfs diff=lfs merge=lfs -text
26
- *.tflite filter=lfs diff=lfs merge=lfs -text
27
- *.tgz filter=lfs diff=lfs merge=lfs -text
28
- *.wasm filter=lfs diff=lfs merge=lfs -text
29
- *.xz filter=lfs diff=lfs merge=lfs -text
30
- *.zip filter=lfs diff=lfs merge=lfs -text
31
- *.zst filter=lfs diff=lfs merge=lfs -text
32
- *tfevents* filter=lfs diff=lfs merge=lfs -text
33
- # Audio files - uncompressed
34
- *.pcm filter=lfs diff=lfs merge=lfs -text
35
- *.sam filter=lfs diff=lfs merge=lfs -text
36
- *.raw filter=lfs diff=lfs merge=lfs -text
37
- # Audio files - compressed
38
- *.aac filter=lfs diff=lfs merge=lfs -text
39
- *.flac filter=lfs diff=lfs merge=lfs -text
40
- *.mp3 filter=lfs diff=lfs merge=lfs -text
41
- *.ogg filter=lfs diff=lfs merge=lfs -text
42
- *.wav filter=lfs diff=lfs merge=lfs -text
43
- # Image files - uncompressed
44
- *.bmp filter=lfs diff=lfs merge=lfs -text
45
- *.gif filter=lfs diff=lfs merge=lfs -text
46
- *.png filter=lfs diff=lfs merge=lfs -text
47
- *.tiff filter=lfs diff=lfs merge=lfs -text
48
- # Image files - compressed
49
- *.jpg filter=lfs diff=lfs merge=lfs -text
50
- *.jpeg filter=lfs diff=lfs merge=lfs -text
51
- *.webp filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
audio.tgz → all/test01-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba58e91bc9a87c57cb226a75a81734811ad274d176d9210e6549184822c4ae76
3
- size 23521945
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c29183236ae585220bf5317eadcd75561a3ea8f19bbe66ed3b79f3d1e255647
3
+ size 23821260
data/metadata_test.csv.gz → bal/test01-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e976ad18d6b019dc12886c644f53c10defcfdbb056bbb655efe236c459e5ff5
3
- size 24347
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1f37973d92159a4c675b3dc86df12970eb78b8070f141d99c7a86a401bba5b1
3
+ size 9933147
languages.py DELETED
@@ -1,10 +0,0 @@
1
- LANGUAGES = {
2
- 'ace': 'Acehnese',
3
- 'bal': 'Balinese',
4
- 'bug': 'Bugisnese',
5
- 'id': 'Indonesian',
6
- 'min': 'Minangkabau',
7
- 'jav': 'Javanese',
8
- 'sun': 'Sundanese',
9
- 'all': 'All'
10
- }
 
 
 
 
 
 
 
 
 
 
 
release_stats.py DELETED
@@ -1,13 +0,0 @@
1
- STATS = {
2
- "name": "Librivox-Indonesia",
3
- "bundleURLTemplate": "https://huggingface.co/datasets/cahya/test01/resolve/main/audio.tgz",
4
- "version": "1.0.0",
5
- "date": "2022-09-04",
6
- "multilingual": True,
7
- "locales": {
8
- "bal": {'duration': 10632910099, 'buckets': {'dev': 16335, 'invalidated': 242945, 'other': 267228, 'reported': 3903, 'test': 16335, 'train': 890116, 'validated': 1556254}, 'reportedSentences': 3842, 'clips': 2066427, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'twenties': 0.24, 'sixties': 0.04, 'thirties': 0.13, 'teens': 0.06, 'seventies': 0.01, 'fourties': 0.1, 'fifties': 0.05, 'eighties': 0, 'nineties': 0}, 'gender': {'': 0.37, 'male': 0.46, 'female': 0.15, 'other': 0.02}}, 'users': 81085, 'size': 76768058135, 'checksum': '7c789a325e1d9f16137c1aeab9f45a1fde049a39b8fbf49e9ef4a8827ae67eae', 'avgDurationSecs': 5.146, 'validDurationSecs': 8007787.777, 'totalHrs': 2953.58, 'validHrs': 2224.38},
9
- "sun": {'duration': 10632910099, 'buckets': {'dev': 16335, 'invalidated': 242945, 'other': 267228, 'reported': 3903, 'test': 16335, 'train': 890116, 'validated': 1556254}, 'reportedSentences': 3842, 'clips': 2066427, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'twenties': 0.24, 'sixties': 0.04, 'thirties': 0.13, 'teens': 0.06, 'seventies': 0.01, 'fourties': 0.1, 'fifties': 0.05, 'eighties': 0, 'nineties': 0}, 'gender': {'': 0.37, 'male': 0.46, 'female': 0.15, 'other': 0.02}}, 'users': 81085, 'size': 76768058135, 'checksum': '7c789a325e1d9f16137c1aeab9f45a1fde049a39b8fbf49e9ef4a8827ae67eae', 'avgDurationSecs': 5.146, 'validDurationSecs': 8007787.777, 'totalHrs': 2953.58, 'validHrs': 2224.38},
10
- "all": {'duration': 10632910099, 'buckets': {'dev': 16335, 'invalidated': 242945, 'other': 267228, 'reported': 3903, 'test': 16335, 'train': 890116, 'validated': 1556254}, 'reportedSentences': 3842, 'clips': 2066427, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'twenties': 0.24, 'sixties': 0.04, 'thirties': 0.13, 'teens': 0.06, 'seventies': 0.01, 'fourties': 0.1, 'fifties': 0.05, 'eighties': 0, 'nineties': 0}, 'gender': {'': 0.37, 'male': 0.46, 'female': 0.15, 'other': 0.02}}, 'users': 81085, 'size': 76768058135, 'checksum': '7c789a325e1d9f16137c1aeab9f45a1fde049a39b8fbf49e9ef4a8827ae67eae', 'avgDurationSecs': 5.146, 'validDurationSecs': 8007787.777, 'totalHrs': 2953.58, 'validHrs': 2224.38},
11
- },
12
- 'totalDuration': 72782088097, 'totalHrs': 20217, 'totalValidDurationSecs': 53904443, 'totalValidHrs': 14973,
13
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/metadata_train.csv.gz → sun/test01-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c74c39006e76f50a9aa8aa49fa9d1c44683f82a50c9dbf7d579640607f263e5
3
- size 189364
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b493d63e543a23850e01d4995cf3b98e9fad01fac92fe70bf7c39b0e8e83cacc
3
+ size 13893759
test01.py DELETED
@@ -1,157 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2022 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """ LibriVox-Indonesia Dataset"""
16
-
17
-
18
- import csv
19
- import os
20
-
21
- import datasets
22
- from datasets.utils.py_utils import size_str
23
-
24
- from .languages import LANGUAGES
25
- from .release_stats import STATS
26
-
27
- _CITATION = """\
28
- """
29
-
30
- _HOMEPAGE = "https://huggingface.co/indonesian-nlp/librivox-indonesia"
31
-
32
- _LICENSE = "https://creativecommons.org/publicdomain/zero/1.0/"
33
-
34
- _AUDIO_URL = "https://huggingface.co/datasets/cahya/test01/resolve/main/audio.tgz"
35
-
36
-
37
- class Test01Config(datasets.BuilderConfig):
38
- """BuilderConfig for Test01."""
39
-
40
- def __init__(self, name, version, **kwargs):
41
- print("## Kwargs", kwargs)
42
- self.language = kwargs.pop("language", None)
43
- self.release_date = kwargs.pop("release_date", None)
44
- self.num_clips = kwargs.pop("num_clips", None)
45
- self.num_speakers = kwargs.pop("num_speakers", None)
46
- self.total_hr = kwargs.pop("total_hr", None)
47
- self.size_bytes = kwargs.pop("size_bytes", None)
48
- self.size_human = size_str(self.size_bytes)
49
- description = (
50
- f"LibriVox-Indonesia speech to text dataset in {self.language} released on {self.release_date}. "
51
- f"The dataset comprises {self.total_hr} hours of transcribed speech data"
52
- )
53
- super(Test01Config, self).__init__(
54
- name=name,
55
- version=datasets.Version(version),
56
- description=description,
57
- **kwargs,
58
- )
59
-
60
-
61
- class Test01(datasets.GeneratorBasedBuilder):
62
- DEFAULT_CONFIG_NAME = "all"
63
-
64
- BUILDER_CONFIGS = [
65
- Test01Config(
66
- name=lang,
67
- version=STATS["version"],
68
- language=LANGUAGES[lang],
69
- release_date=STATS["date"],
70
- num_clips=lang_stats["clips"],
71
- num_speakers=lang_stats["users"],
72
- total_hr=float(lang_stats["totalHrs"]) if lang_stats["totalHrs"] else None,
73
- size_bytes=int(lang_stats["size"]) if lang_stats["size"] else None,
74
- )
75
- for lang, lang_stats in STATS["locales"].items()
76
- ]
77
-
78
- def _info(self):
79
- total_languages = len(STATS["locales"])
80
- total_hours = self.config.total_hr
81
- description = (
82
- "LibriVox-Indonesia is a speech dataset generated from LibriVox with only languages from Indonesia."
83
- f"The dataset currently consists of {total_hours} hours of speech "
84
- f" in {total_languages} languages, but more voices and languages are always added."
85
- )
86
- features = datasets.Features(
87
- {
88
- "path": datasets.Value("string"),
89
- "language": datasets.Value("string"),
90
- "reader": datasets.Value("string"),
91
- "sentence": datasets.Value("string"),
92
- "audio": datasets.features.Audio(sampling_rate=48_000)
93
- }
94
- )
95
-
96
- return datasets.DatasetInfo(
97
- description=description,
98
- features=features,
99
- supervised_keys=None,
100
- homepage=_HOMEPAGE,
101
- license=_LICENSE,
102
- citation=_CITATION,
103
- version=self.config.version,
104
- )
105
-
106
- def _split_generators(self, dl_manager):
107
- """Returns SplitGenerators."""
108
- dl_manager.download_config.ignore_url_params = True
109
-
110
- archive_path = dl_manager.download(_AUDIO_URL)
111
- local_extracted_archive = dl_manager.extract(archive_path) if not dl_manager.is_streaming else None
112
- path_to_clips = "audio"
113
-
114
- return [
115
- datasets.SplitGenerator(
116
- name=datasets.Split.TRAIN,
117
- gen_kwargs={
118
- "local_extracted_archive": local_extracted_archive,
119
- "archive_iterator": dl_manager.iter_archive(archive_path),
120
- "metadata_filepath": "audio_transcription.csv",
121
- "path_to_clips": path_to_clips,
122
- },
123
- ),
124
- ]
125
-
126
- def _generate_examples(
127
- self,
128
- local_extracted_archive,
129
- archive_iterator,
130
- metadata_filepath,
131
- path_to_clips,
132
- ):
133
- """Yields examples."""
134
- data_fields = list(self._info().features.keys())
135
- metadata = {}
136
- filepath = local_extracted_archive + "/audio/audio_transcription.csv"
137
- with open(filepath, "r") as f:
138
- lines = (line for line in f)
139
- utterances = csv.DictReader(lines)
140
- for row in utterances:
141
- if self.config.name == "all" or self.config.name == row["language"]:
142
- row["path"] = os.path.join(path_to_clips, row["path"])
143
- # if data is incomplete, fill with empty values
144
- for field in data_fields:
145
- if field not in row:
146
- row[field] = ""
147
- metadata[row["path"]] = row
148
- for path, f in archive_iterator:
149
- if path in metadata:
150
- result = dict(metadata[path])
151
- # set the audio feature and the path to the extracted file
152
- path = os.path.join(local_extracted_archive, path) if local_extracted_archive else path
153
- result["audio"] = {"path": path, "bytes": f.read()}
154
- # set path to None if the audio file doesn't exist locally (i.e. in streaming mode)
155
- result["path"] = path if local_extracted_archive else None
156
-
157
- yield path, result
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
usage.py DELETED
@@ -1,10 +0,0 @@
1
- from datasets import load_dataset
2
-
3
-
4
- def main():
5
- ds = load_dataset("./test01.py", "bal", ignore_verifications=True)
6
- print(ds)
7
-
8
-
9
- if __name__ == "__main__":
10
- main()