plnguyen2908 commited on
Commit
c336ea1
·
verified ·
1 Parent(s): b948ca8

Upload UniTalk.py

Browse files
Files changed (1) hide show
  1. UniTalk.py +161 -21
UniTalk.py CHANGED
@@ -1,23 +1,163 @@
1
- from pathlib import Path
2
-
3
- def _split_generators(self, dl_manager):
4
- # always resolve to the folder where this script lives
5
- base_dir = Path(__file__).parent.resolve()
6
- return [
7
- datasets.SplitGenerator(
8
- name=datasets.Split.TRAIN,
9
- gen_kwargs={
10
- "csv_path": str(base_dir / "csv" / "train_orig.csv"),
11
- "videos_dir": str(base_dir / "clips_videos" / "train"),
12
- "audios_dir": str(base_dir / "clips_audios" / "train"),
13
- },
14
- ),
15
- datasets.SplitGenerator(
16
- name=datasets.Split.VALIDATION,
17
- gen_kwargs={
18
- "csv_path": str(base_dir / "csv" / "val_orig.csv"),
19
- "videos_dir": str(base_dir / "clips_videos" / "val"),
20
- "audios_dir": str(base_dir / "clips_audios" / "val"),
21
- },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  ),
23
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # unitalk_dataset.py
2
+
3
+ import os
4
+ import zipfile
5
+ import tempfile
6
+
7
+ import numpy as np
8
+ import pandas as pd
9
+ from PIL import Image
10
+ from scipy.io import wavfile
11
+ import datasets
12
+ from datasets import Value, Sequence, Features
13
+
14
+ _CITATION = """\
15
+ @misc{unitalk2025,
16
+ title={UniTalk Dataset},
17
+ author={Your Name},
18
+ year={2025}
19
+ }
20
+ """
21
+
22
+ class UniTalk(datasets.GeneratorBasedBuilder):
23
+ """
24
+ UniTalk: grouped frames and audio per entity_id.
25
+ Downloads CSVs and ZIPs from the HF repo, pairs video_ids with downloaded zips directly.
26
+
27
+ Each example returns:
28
+ - entity_id: string
29
+ - images: list of flattened uint8 frame arrays ([H*W*3])
30
+ - audio: list of int16 PCM samples
31
+ - frame_timestamp: list of floats
32
+ - label_id: list of int64 labels (0/1)
33
+ """
34
+
35
+ VERSION = datasets.Version("1.0.0")
36
+ BUILDER_CONFIGS = [
37
+ datasets.BuilderConfig(
38
+ name="default",
39
+ version=VERSION,
40
+ description="Default config for UniTalk dataset"
41
  ),
42
  ]
43
+ DEFAULT_CONFIG_NAME = "default"
44
+
45
+ def _info(self):
46
+ return datasets.DatasetInfo(
47
+ description="UniTalk: frames and audio grouped by entity_id.",
48
+ features=Features({
49
+ "entity_id": Value("string"),
50
+ "images": Sequence(Sequence(Value("uint8"))), # [num_frames][H*W*3]
51
+ "audio": Sequence(Value("int16")), # [num_samples]
52
+ "frame_timestamp": Sequence(Value("float32")), # timestamps per frame
53
+ "label_id": Sequence(Value("int64")), # labels per frame
54
+ }),
55
+ supervised_keys=None,
56
+ homepage="https://huggingface.co/datasets/plnguyen2908/UniTalk",
57
+ citation=_CITATION,
58
+ )
59
+
60
+ def _split_generators(self, dl_manager):
61
+ repo_id = "plnguyen2908/UniTalk"
62
+ revision = "main"
63
+ base_url = f"https://huggingface.co/datasets/{repo_id}/resolve/{revision}"
64
+
65
+ # Download CSVs
66
+ train_csv = dl_manager.download(f"{base_url}/csv/train_orig.csv")
67
+ val_csv = dl_manager.download(f"{base_url}/csv/val_orig.csv")
68
+
69
+ # Read CSVs to get video_ids
70
+ df_train = pd.read_csv(train_csv)
71
+ train_vids = sorted(df_train["video_id"].unique())
72
+ df_val = pd.read_csv(val_csv)
73
+ val_vids = sorted(df_val["video_id"].unique())
74
+
75
+ # Build URLs and download train zips
76
+ train_vid_urls = [f"{base_url}/clips_videos/train/{v}.zip" for v in train_vids]
77
+ train_aud_urls = [f"{base_url}/clips_audios/train/{v}.zip" for v in train_vids]
78
+ train_vid_zips = dl_manager.download(train_vid_urls)
79
+ train_aud_zips = dl_manager.download(train_aud_urls)
80
+
81
+ # Build URLs and download val zips
82
+ val_vid_urls = [f"{base_url}/clips_videos/val/{v}.zip" for v in val_vids]
83
+ val_aud_urls = [f"{base_url}/clips_audios/val/{v}.zip" for v in val_vids]
84
+ val_vid_zips = dl_manager.download(val_vid_urls)
85
+ val_aud_zips = dl_manager.download(val_aud_urls)
86
+
87
+ return [
88
+ datasets.SplitGenerator(
89
+ name=datasets.Split.TRAIN,
90
+ gen_kwargs={
91
+ "csv_path": train_csv,
92
+ "video_ids": train_vids,
93
+ "video_zips": train_vid_zips,
94
+ "audio_zips": train_aud_zips,
95
+ },
96
+ ),
97
+ datasets.SplitGenerator(
98
+ name=datasets.Split.VALIDATION,
99
+ gen_kwargs={
100
+ "csv_path": val_csv,
101
+ "video_ids": val_vids,
102
+ "video_zips": val_vid_zips,
103
+ "audio_zips": val_aud_zips,
104
+ },
105
+ ),
106
+ ]
107
+
108
+ def _generate_examples(self, csv_path, video_ids, video_zips, audio_zips):
109
+ temp_root = tempfile.mkdtemp(prefix="unitalk_")
110
+ df = pd.read_csv(csv_path, dtype={"frame_timestamp": float, "label_id": int})
111
+
112
+ # Pair video_ids to corresponding zips
113
+ vid_map = dict(zip(video_ids, video_zips))
114
+ aud_map = dict(zip(video_ids, audio_zips))
115
+
116
+ for entity_id, group in df.groupby("entity_id"):
117
+ raw_vid = group["video_id"].iloc[0]
118
+ # video_id must match key in video_ids
119
+ if raw_vid not in vid_map:
120
+ raise ValueError(f"Video ID {raw_vid} not found among downloaded zips")
121
+ video_zip = vid_map[raw_vid]
122
+ audio_zip = aud_map[raw_vid]
123
+
124
+ # 1) Extract frames
125
+ with zipfile.ZipFile(video_zip, "r") as z:
126
+ prefix = f"{raw_vid}/{entity_id}/"
127
+ members = sorted(f for f in z.namelist() if f.startswith(prefix) and f.lower().endswith('.jpg'))
128
+ images, timestamps = [], []
129
+ for member in members:
130
+ path = z.extract(member, path=temp_root)
131
+ arr = np.array(Image.open(path).convert('RGB'), dtype=np.uint8)
132
+ images.append(arr.flatten().tolist())
133
+ ts = float(os.path.splitext(os.path.basename(path))[0])
134
+ timestamps.append(ts)
135
+
136
+ # 2) Extract audio
137
+ with zipfile.ZipFile(audio_zip, "r") as z:
138
+ wav_members = [m for m in z.namelist()
139
+ if m.startswith(f"{raw_vid}/{entity_id}") and m.lower().endswith('.wav')]
140
+ if not wav_members:
141
+ raise FileNotFoundError(f"No WAV for {entity_id} in {audio_zip}")
142
+ wav_path = z.extract(wav_members[0], path=temp_root)
143
+ _, audio_arr = wavfile.read(wav_path)
144
+ audio = audio_arr.tolist()
145
+
146
+ # 3) Align labels
147
+ label_map = dict(zip(group["frame_timestamp"].tolist(), group["label_id"].tolist()))
148
+ label_id = [label_map[t] for t in timestamps]
149
+
150
+ yield entity_id, {
151
+ "entity_id": entity_id,
152
+ "images": images,
153
+ "audio": audio,
154
+ "frame_timestamp": timestamps,
155
+ "label_id": label_id,
156
+ }
157
+
158
+ if __name__ == "__main__":
159
+ ds = datasets.load_dataset(
160
+ path=__file__,
161
+ trust_remote_code=True
162
+ )
163
+ print(ds)