Switch to parquet
Browse files- clustering_files.py +3 -3
- dataset.parquet +3 -0
clustering_files.py
CHANGED
|
@@ -82,7 +82,7 @@ class ClusteringFiles(datasets.GeneratorBasedBuilder):
|
|
| 82 |
|
| 83 |
def _generate_examples(self, file_path):
|
| 84 |
"""Yields examples."""
|
| 85 |
-
df = pd.
|
| 86 |
|
| 87 |
files = {}
|
| 88 |
for idx, row in df.iterrows():
|
|
@@ -93,9 +93,9 @@ class ClusteringFiles(datasets.GeneratorBasedBuilder):
|
|
| 93 |
}
|
| 94 |
files[row["filename"]]["segments"].append({
|
| 95 |
"segment_id": row["segment_id"],
|
| 96 |
-
"speaker": row
|
| 97 |
"duration": row["duration"],
|
| 98 |
-
"segment_clean": row
|
| 99 |
"start": row['start'],
|
| 100 |
"end": row['end'],
|
| 101 |
"readable_start": row['readable_start'],
|
|
|
|
| 82 |
|
| 83 |
def _generate_examples(self, file_path):
|
| 84 |
"""Yields examples."""
|
| 85 |
+
df = pd.read_parquet(file_path)
|
| 86 |
|
| 87 |
files = {}
|
| 88 |
for idx, row in df.iterrows():
|
|
|
|
| 93 |
}
|
| 94 |
files[row["filename"]]["segments"].append({
|
| 95 |
"segment_id": row["segment_id"],
|
| 96 |
+
"speaker": row.get("speaker", 'unknown'),
|
| 97 |
"duration": row["duration"],
|
| 98 |
+
"segment_clean": row.get("segment_clean", True),
|
| 99 |
"start": row['start'],
|
| 100 |
"end": row['end'],
|
| 101 |
"readable_start": row['readable_start'],
|
dataset.parquet
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:16963c709d10f4491adb04b8c2566cd158f6f3966eec34d223103f916f272036
|
| 3 |
+
size 449338
|