orgn3ai commited on
Commit
354062e
·
verified ·
1 Parent(s): bcc1bce

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,307 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-nd-4.0
3
+ task_categories:
4
+ - video-classification
5
+ language:
6
+ - en
7
+ tags:
8
+ - egocentric
9
+ - embodied-ai
10
+ - robotics
11
+ - imu
12
+ - real-world
13
+ - computer-vision
14
+ - dataset
15
+ - sample-dataset
16
+ size_categories:
17
+ - n<1K
18
+ viewer: false
19
+ ---
20
+
21
+ # PIZZA-PREPARE-Dough-Ballforming-sample
22
+
23
+ ## Overview
24
+
25
+ This dataset provides a high-fidelity, ego-centric capture of professional order preparation tasks within a cold-chain logistics environment. It focuses on the rapid, repetitive, and high-precision picking and packing of multi-unit yogurt packs. This resource is specifically designed to train robotic agents in **bimanual coordination**, **spatial reachability**, and **contact-point optimization** for fragile rigid goods.
26
+
27
+ ## Key Technical Features
28
+
29
+ **Temporal Protocol (T1-T4)**: Every micro-action is frame-accurately annotated using our proprietary four-point system:
30
+ * **T1 (Contact)**: Initial physical engagement.
31
+ * **T2 (Lift-off)**: Static-to-dynamic transition.
32
+ * **T3 (Placement)**: Target container engagement.
33
+ * **T4 (Release)**: Total tactile disengagement.
34
+
35
+ ## Use Cases for Research
36
+
37
+ * **Foundation Models & World Models**: Training models to understand the Newtonian physics of rigid payloads and container boundaries.
38
+ * **Bimanual Policy Learning**: Developing algorithms for coordinated two-hand tasks, such as stabilizing a box while inserting a pack.
39
+ * **End-to-End Picking Pipelines**: Optimizing the "approach-to-grasp" phase in unstructured industrial environments where lighting and clutter vary.
40
+
41
+ ## Custom Data Collection Services
42
+
43
+ Our team specializes in high-fidelity data acquisition within real-world professional settings. We provide on-demand data collection services tailored to specific AI and robotics requirements:
44
+ * Professional Network: Direct access to 100+ professional environments, including professional kitchens, bakeries, mechanical workshops, craft studios, and industrial facilities.
45
+ * Multi-Modal Capture: Expertise in collecting synchronized streams including Third-Person views, Ego-centric (FPV), IMU sensors (motion tracking), and Expert Audio Narration.
46
+ * Domain Expertise: We bridge the gap between technical AI needs and authentic professional "tacit knowledge."
47
+
48
+ ## Commercial Licensing and Contact
49
+
50
+ * The complete dataset and our custom collection services are available for commercial licensing and large-scale R&D. Whether you need existing data or a custom setup in a specific professional environment, do not hesitate to reach out for more information.
51
+ * Contact: orgn3ai@gmail.com
52
+
53
+ ## License
54
+
55
+ * This dataset is licensed under cc-by-nc-nd-4.0.
56
+
57
+ ## Dataset Statistics
58
+
59
+ This section provides detailed statistics extracted from `dataset_metadata.json`:
60
+
61
+ ### Overall Statistics
62
+
63
+ - **Dataset Name**: PIZZA-PREPARE-Dough-Ballforming-sample
64
+ - **Batch ID**: pizza
65
+ - **Total Clips**: 10
66
+ - **Number of Sequences**: 15
67
+ - **Number of Streams**: 3
68
+ - **Stream Types**: ego, imu, third
69
+
70
+ ### Duration Statistics
71
+
72
+ - **Total Duration**: 4.82 minutes (289.20 seconds)
73
+ - **Average Clip Duration**: 28.92 seconds (28920.0 ms)
74
+ - **Min Clip Duration**: 27.10 seconds (27100 ms)
75
+ - **Max Clip Duration**: 32.83 seconds (32833 ms)
76
+
77
+ ### Clip Configuration
78
+
79
+ - **Padding**: 1500 ms
80
+
81
+ ### Statistics by Stream Type
82
+
83
+ #### Ego
84
+
85
+ - **Number of clips**: 5
86
+ - **Total duration**: 2.41 minutes (144.60 seconds)
87
+ - **Average clip duration**: 28.92 seconds (28920.0 ms)
88
+ - **Min clip duration**: 27.10 seconds (27100 ms)
89
+ - **Max clip duration**: 32.83 seconds (32833 ms)
90
+
91
+ #### Third
92
+
93
+ - **Number of clips**: 5
94
+ - **Total duration**: 2.41 minutes (144.60 seconds)
95
+ - **Average clip duration**: 28.92 seconds (28920.0 ms)
96
+ - **Min clip duration**: 27.10 seconds (27100 ms)
97
+ - **Max clip duration**: 32.83 seconds (32833 ms)
98
+
99
+ > **Note**: Complete metadata is available in `dataset_metadata.json` in the dataset root directory.
100
+
101
+ ## Dataset Structure
102
+
103
+ The dataset uses a **unified structure** where each example contains all synchronized video streams:
104
+
105
+ ```
106
+ dataset/
107
+ ├── data-*.arrow # Dataset files (Arrow format)
108
+ ├── dataset_info.json # Dataset metadata
109
+ ├── dataset_metadata.json # Complete dataset statistics
110
+ ├── state.json # Dataset state
111
+ ├── README.md # This file
112
+ ├── medias/ # Media files (mosaics, previews, etc.)
113
+ │ └── mosaic.mp4 # Mosaic preview video
114
+ └── videos/ # All video clips
115
+ └── ego/ # Ego video clips
116
+ └── imu/ # Imu video clips
117
+ └── third/ # Third video clips
118
+ ```
119
+
120
+ ### Dataset Format
121
+
122
+ The dataset contains **10 synchronized scenes** in a single `train` split. Each example includes:
123
+
124
+ - **Synchronized video columns**: One column per flux type (e.g., `ego`, `imu`, `third`)
125
+ - **Scene metadata**: `scene_id`, `sync_id`, `duration_ms`, `padding_ms`, `fps`
126
+ - **Rich metadata dictionary**: Task, environment, audio info, and synchronization details
127
+
128
+ All videos in a single example are synchronized and correspond to the same moment in time.
129
+
130
+ ## Usage
131
+
132
+ ### Load and Access Dataset
133
+
134
+ ```python
135
+ import json
136
+ import random
137
+ from pathlib import Path
138
+ import cv2
139
+ from huggingface_hub import snapshot_download
140
+ from datasets import load_from_disk
141
+
142
+ repo = "orgn3ai/PIZZA-PREPARE-Dough-Ballforming-sample"
143
+
144
+ # 1) Download snapshot locally
145
+ local_path = snapshot_download(repo_id=repo, repo_type="dataset")
146
+ base_dir = Path(local_path)
147
+ print("Snapshot path:", base_dir)
148
+
149
+ # 2) Load dataset saved with save_to_disk()
150
+ ds = load_from_disk(str(base_dir))
151
+ train = ds["train"] if isinstance(ds, dict) and "train" in ds else ds
152
+ print("Train rows:", len(train))
153
+ print("Train columns:", train.column_names)
154
+
155
+ # 3) Read root metadata.json and extract "flux"
156
+ metadata_path = base_dir / "dataset_metadata.json"
157
+ if not metadata_path.exists():
158
+ raise FileNotFoundError(
159
+ f"dataset_metadata.json not found at repo root: {metadata_path}\n"
160
+ "Check your repo tree; maybe it's named dataset_metadata.json instead."
161
+ )
162
+
163
+ with metadata_path.open("r", encoding="utf-8") as f:
164
+ root_meta = json.load(f)
165
+
166
+ flux = root_meta.get("flux")
167
+ if not isinstance(flux, list) or not flux:
168
+ raise ValueError(f'Expected metadata.json["flux"] to be a non-empty list, got: {flux}')
169
+
170
+ print("Flux entries:", flux)
171
+
172
+ # 4) Pick a random dataset entry
173
+ idx = random.randrange(len(train))
174
+ ex = train[idx]
175
+
176
+ print("\nRandom example index:", idx)
177
+ print("Example keys:", list(ex.keys()))
178
+
179
+ def resolve_video_path(video_value) -> Path:
180
+ """
181
+ video_value can be:
182
+ - string path (most common case)
183
+ - dict like {"path": "...", "bytes": ...} (for backward compatibility)
184
+ """
185
+ if isinstance(video_value, dict) and "path" in video_value:
186
+ rel = video_value["path"]
187
+ elif isinstance(video_value, str):
188
+ rel = video_value
189
+ else:
190
+ raise TypeError(f"Unsupported video value type: {type(video_value)}; value={video_value}")
191
+
192
+ # Normalize to avoid leading "./"
193
+ rel = str(rel).lstrip("/")
194
+
195
+ # Your dataset may store relative paths like "videos/ego/xxx.mp4"
196
+ # Resolve them inside the snapshot folder.
197
+ return base_dir / rel
198
+
199
+ def inspect_video(path: Path):
200
+ print(f" Local path: {path}")
201
+ print(f" Exists: {path.exists()}")
202
+ if not path.exists():
203
+ return {"ok": False, "reason": "file_not_found"}
204
+
205
+ cap = cv2.VideoCapture(str(path))
206
+ if not cap.isOpened():
207
+ return {"ok": False, "reason": "cannot_open"}
208
+
209
+ frame_count = int(cap.get(cv2.CAP_PROP_FRAME_COUNT))
210
+ fps = float(cap.get(cv2.CAP_PROP_FPS))
211
+ width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH))
212
+ height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))
213
+
214
+ # Some codecs report fps=0; guard it
215
+ duration = (frame_count / fps) if fps and fps > 0 else None
216
+
217
+ # Try read first frame
218
+ ret, frame0 = cap.read()
219
+ cap.release()
220
+
221
+ info = {
222
+ "ok": True,
223
+ "width": width,
224
+ "height": height,
225
+ "fps": fps,
226
+ "frame_count": frame_count,
227
+ "duration_sec": duration,
228
+ "first_frame_ok": bool(ret),
229
+ "first_frame_shape": tuple(frame0.shape) if ret and frame0 is not None else None,
230
+ "first_frame_dtype": str(frame0.dtype) if ret and frame0 is not None else None,
231
+ }
232
+ return info
233
+
234
+ # 5) For each flux key, inspect the associated video
235
+ print("\n=== VIDEO CHECK ===")
236
+ for key in flux:
237
+ print(f"\nFlux key: {key}")
238
+ if key not in ex:
239
+ print(f" ERROR: key '{key}' not in example. Available keys: {list(ex.keys())}")
240
+ continue
241
+
242
+ try:
243
+ video_path = resolve_video_path(ex[key])
244
+ except Exception as e:
245
+ print(f" ERROR resolving path: {e}")
246
+ continue
247
+
248
+ info = inspect_video(video_path)
249
+ if not info["ok"]:
250
+ print(f" ERROR: {info['reason']}")
251
+ continue
252
+
253
+ print(" Video properties:")
254
+ print(f" - Resolution: {info['width']}x{info['height']}")
255
+ print(f" - FPS: {info['fps']:.3f}")
256
+ print(f" - Frames: {info['frame_count']}")
257
+ if info["duration_sec"] is not None:
258
+ print(f" - Duration: {info['duration_sec']:.3f}s")
259
+ else:
260
+ print(" - Duration: (fps unavailable)")
261
+ print(f" - First frame decoded: {info['first_frame_ok']}")
262
+ if info["first_frame_ok"]:
263
+ print(f" - Frame0 shape: {info['first_frame_shape']}")
264
+ print(f" - Frame0 dtype: {info['first_frame_dtype']}")
265
+
266
+ print('\n=== LABELS ===')
267
+ print(f"nbLabels: {len(ex['labels'])}")
268
+ for label in ex['labels']:
269
+ print(f" - {label['time_ms']}ms (withoutPadding): {label['label']}")
270
+
271
+ print("\nDONE.")
272
+ ```
273
+
274
+ ### Dataset Features
275
+
276
+ Each example contains:
277
+
278
+ - **`scene_id`**: Unique scene identifier (e.g., "01_0000")
279
+ - **`sync_id`**: Synchronization ID linking synchronized clips
280
+ - **`duration_ms`**: Duration of the synchronized clip in milliseconds (includes padding)
281
+ - **`padding_ms`**: Padding applied to clips (added at beginning and end, total padding = padding_ms × 2)
282
+ - **`fps`**: Frames per second (extracted from video)
283
+ - **`batch_id`**: Batch identifier
284
+ - **`dataset_name`**: Dataset name from config
285
+ - **One column per flux**: Each flux name from `metadata['flux_names']` has its own column (e.g., `ego`, `imu`, `third`) - String path to video file (relative to dataset root)
286
+ - **`metadata`**: Dictionary containing:
287
+ - `task`: Task identifier
288
+ - `environment`: Environment description
289
+ - `has_audio`: Whether videos contain audio
290
+ - `num_fluxes`: Number of synchronized flux types
291
+ - `flux_names`: List of flux names present
292
+ - `sequence_ids`: List of original sequence IDs
293
+ - `sync_offsets_ms`: List of synchronization offsets
294
+
295
+ ## Additional Notes
296
+
297
+ **Important**: This dataset uses a unified structure where each example contains all synchronized video streams in separate columns. All examples are in the `train` split.
298
+
299
+ **Synchronization**: Videos in the same example (same index in the `train` split) are automatically synchronized. They share the same `sync_id` and correspond to the same moment in time.
300
+
301
+ **Flux Keys**: The available flux keys are listed in `dataset_metadata.json` under the `"flux"` key. Use these keys to programmatically access video columns in each example.
302
+
303
+ **Video Paths**: Video paths are stored as strings (relative to the dataset root directory). Paths can be resolved using the `resolve_video_path` function shown in the usage example above.
304
+
305
+ ## License
306
+
307
+ This dataset is licensed under **cc-by-nc-nd-4.0**.
dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train"]}
dataset_metadata.json ADDED
@@ -0,0 +1,156 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_name": "PIZZA-PREPARE-Dough-Ballforming-sample",
3
+ "batch_id": "pizza",
4
+ "total_clips": 10,
5
+ "num_sequences": 15,
6
+ "num_streams": 3,
7
+ "stream_types": [
8
+ "ego",
9
+ "imu",
10
+ "third"
11
+ ],
12
+ "flux": [
13
+ "ego",
14
+ "imu",
15
+ "third"
16
+ ],
17
+ "padding_ms": 1500,
18
+ "default_duration_ms": 500,
19
+ "clip_duration_ms": {
20
+ "base": 500,
21
+ "with_padding": 3500
22
+ },
23
+ "duration_ms": {
24
+ "average": 28920.0,
25
+ "total": 289200,
26
+ "min": 27100,
27
+ "max": 32833
28
+ },
29
+ "flux_stats": {
30
+ "ego": {
31
+ "num_clips": 5,
32
+ "duration_ms": {
33
+ "average": 28920.0,
34
+ "total": 144600,
35
+ "min": 27100,
36
+ "max": 32833
37
+ }
38
+ },
39
+ "third": {
40
+ "num_clips": 5,
41
+ "duration_ms": {
42
+ "average": 28920.0,
43
+ "total": 144600,
44
+ "min": 27100,
45
+ "max": 32833
46
+ }
47
+ }
48
+ },
49
+ "sequences": [
50
+ {
51
+ "id": "pizza_dough_ballforming_01_ego_25000ms",
52
+ "flux_name": "ego",
53
+ "type": "video",
54
+ "num_clips": 1,
55
+ "duration_ms": 25000
56
+ },
57
+ {
58
+ "id": "pizza_dough_ballforming_01_ego_24100ms",
59
+ "flux_name": "ego",
60
+ "type": "video",
61
+ "num_clips": 1,
62
+ "duration_ms": 24100
63
+ },
64
+ {
65
+ "id": "pizza_dough_ballforming_01_ego_26500ms",
66
+ "flux_name": "ego",
67
+ "type": "video",
68
+ "num_clips": 1,
69
+ "duration_ms": 26500
70
+ },
71
+ {
72
+ "id": "pizza_dough_ballforming_01_ego_24167ms",
73
+ "flux_name": "ego",
74
+ "type": "video",
75
+ "num_clips": 1,
76
+ "duration_ms": 24167
77
+ },
78
+ {
79
+ "id": "pizza_dough_ballforming_01_ego_29833ms",
80
+ "flux_name": "ego",
81
+ "type": "video",
82
+ "num_clips": 1,
83
+ "duration_ms": 29833
84
+ },
85
+ {
86
+ "id": "pizza_dough_ballforming_01_third_25000ms",
87
+ "flux_name": "third",
88
+ "type": "video",
89
+ "num_clips": 1,
90
+ "duration_ms": 25000
91
+ },
92
+ {
93
+ "id": "pizza_dough_ballforming_01_third_24100ms",
94
+ "flux_name": "third",
95
+ "type": "video",
96
+ "num_clips": 1,
97
+ "duration_ms": 24100
98
+ },
99
+ {
100
+ "id": "pizza_dough_ballforming_01_third_26500ms",
101
+ "flux_name": "third",
102
+ "type": "video",
103
+ "num_clips": 1,
104
+ "duration_ms": 26500
105
+ },
106
+ {
107
+ "id": "pizza_dough_ballforming_01_third_24167ms",
108
+ "flux_name": "third",
109
+ "type": "video",
110
+ "num_clips": 1,
111
+ "duration_ms": 24167
112
+ },
113
+ {
114
+ "id": "pizza_dough_ballforming_01_third_29833ms",
115
+ "flux_name": "third",
116
+ "type": "video",
117
+ "num_clips": 1,
118
+ "duration_ms": 29833
119
+ },
120
+ {
121
+ "id": "pizza_dough_ballforming_01_imu_25000ms",
122
+ "flux_name": "imu",
123
+ "type": "imu",
124
+ "num_clips": 1,
125
+ "duration_ms": 25000
126
+ },
127
+ {
128
+ "id": "pizza_dough_ballforming_01_imu_24100ms",
129
+ "flux_name": "imu",
130
+ "type": "imu",
131
+ "num_clips": 1,
132
+ "duration_ms": 24100
133
+ },
134
+ {
135
+ "id": "pizza_dough_ballforming_01_imu_26500ms",
136
+ "flux_name": "imu",
137
+ "type": "imu",
138
+ "num_clips": 1,
139
+ "duration_ms": 26500
140
+ },
141
+ {
142
+ "id": "pizza_dough_ballforming_01_imu_24167ms",
143
+ "flux_name": "imu",
144
+ "type": "imu",
145
+ "num_clips": 1,
146
+ "duration_ms": 24167
147
+ },
148
+ {
149
+ "id": "pizza_dough_ballforming_01_imu_29833ms",
150
+ "flux_name": "imu",
151
+ "type": "imu",
152
+ "num_clips": 1,
153
+ "duration_ms": 29833
154
+ }
155
+ ]
156
+ }
imu/imu/0000.json ADDED
The diff for this file is too large to render. See raw diff
 
train/data-00000-of-00001.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:633200aa3f169d9680e9ddb334f18fd182599adbfb317278eeaf7ad061e9c0e7
3
+ size 5208
train/dataset_info.json ADDED
@@ -0,0 +1,114 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "",
3
+ "description": "",
4
+ "features": {
5
+ "scene_id": {
6
+ "dtype": "string",
7
+ "_type": "Value"
8
+ },
9
+ "sync_id": {
10
+ "dtype": "int32",
11
+ "_type": "Value"
12
+ },
13
+ "duration_ms": {
14
+ "dtype": "int32",
15
+ "_type": "Value"
16
+ },
17
+ "padding_ms": {
18
+ "dtype": "int32",
19
+ "_type": "Value"
20
+ },
21
+ "fps": {
22
+ "dtype": "float32",
23
+ "_type": "Value"
24
+ },
25
+ "batch_id": {
26
+ "dtype": "string",
27
+ "_type": "Value"
28
+ },
29
+ "dataset_name": {
30
+ "dtype": "string",
31
+ "_type": "Value"
32
+ },
33
+ "ego": {
34
+ "dtype": "string",
35
+ "_type": "Value"
36
+ },
37
+ "imu": {
38
+ "dtype": "string",
39
+ "_type": "Value"
40
+ },
41
+ "third": {
42
+ "dtype": "string",
43
+ "_type": "Value"
44
+ },
45
+ "metadata": {
46
+ "task": {
47
+ "dtype": "string",
48
+ "_type": "Value"
49
+ },
50
+ "environment": {
51
+ "dtype": "string",
52
+ "_type": "Value"
53
+ },
54
+ "has_audio": {
55
+ "dtype": "bool",
56
+ "_type": "Value"
57
+ },
58
+ "padding_ms": {
59
+ "dtype": "int32",
60
+ "_type": "Value"
61
+ },
62
+ "num_fluxes": {
63
+ "dtype": "int32",
64
+ "_type": "Value"
65
+ },
66
+ "flux_names": {
67
+ "feature": {
68
+ "dtype": "string",
69
+ "_type": "Value"
70
+ },
71
+ "_type": "List"
72
+ },
73
+ "sequence_ids": {
74
+ "feature": {
75
+ "dtype": "string",
76
+ "_type": "Value"
77
+ },
78
+ "_type": "List"
79
+ },
80
+ "sync_offsets_ms": {
81
+ "feature": {
82
+ "flux_name": {
83
+ "dtype": "string",
84
+ "_type": "Value"
85
+ },
86
+ "offset_ms": {
87
+ "dtype": "int32",
88
+ "_type": "Value"
89
+ }
90
+ },
91
+ "_type": "List"
92
+ }
93
+ },
94
+ "labels": {
95
+ "feature": {
96
+ "time_ms": {
97
+ "dtype": "int32",
98
+ "_type": "Value"
99
+ },
100
+ "label": {
101
+ "dtype": "string",
102
+ "_type": "Value"
103
+ },
104
+ "sequenceType": {
105
+ "dtype": "string",
106
+ "_type": "Value"
107
+ }
108
+ },
109
+ "_type": "List"
110
+ }
111
+ },
112
+ "homepage": "",
113
+ "license": ""
114
+ }
train/state.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "data-00000-of-00001.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "0c4954faa0fc27b3",
8
+ "_format_columns": null,
9
+ "_format_kwargs": {},
10
+ "_format_type": null,
11
+ "_output_all_columns": false,
12
+ "_split": null
13
+ }
videos/ego/0000.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f02406939641a917554bf7dca0e8ece9bf15c86af6d0b68a13dbb4ed11c21bd1
3
+ size 126781437
videos/ego/0001.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c440c720fbdec8f200d087afccc622ba482a845c886926f658aab40ab3c016
3
+ size 121786041
videos/ego/0002.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c6b48950641824a643c8940aabca082445804a8a70eaf36351952b4986e6a23
3
+ size 129113901
videos/ego/0003.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b632574cbe5ddfe92ab885325a2d68ee06394adefa74fb892008270678e0f10
3
+ size 119814737
videos/ego/0004.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b2cc69a6a282c0e694d1dfb9d004879dc643eae8681f87070acab9f8e6874c
3
+ size 146147677
videos/third/0000.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba014e77dfa4a15e40c15d009ffa6cab913c22269567db0ab1b7cab660dbb87f
3
+ size 30950718
videos/third/0001.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54ae60c5ac5fc388b01eae778076c3f59d4f95f6ea670cff7b415b14800a0ccf
3
+ size 30356502
videos/third/0002.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b70e13e3f0b52e5e51bf9d271f0bbfaaeef664301645d22b8927bc24d1ce2e41
3
+ size 32530954
videos/third/0003.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43f0896a238ab2c558dfcd16b8adf0486bad086fecbcd48b3d4f74587964491b
3
+ size 29784698
videos/third/0004.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eeb0213fac44636bc2fafd906bce654b4e26464b0d856b5098e6f707968b868
3
+ size 36393263