Datasets:
File size: 15,551 Bytes
ba80d27 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 | #!/usr/bin/env python3
"""
EgomotionData Dataset Inspector
================================
This script reads and displays the structure, statistics, and field meanings of the EgomotionData dataset.
Usage:
python inspect_dataset.py # View dataset overview + random sample details
python inspect_dataset.py --sample PATH.npz # View details of a specific npz file
python inspect_dataset.py --overview-only # View only dataset overview statistics
python inspect_dataset.py --check-consistency N # Randomly check consistency of N files
"""
import argparse
import os
import sys
import random
import numpy as np
from pathlib import Path
from collections import defaultdict
# ============================================================================
# Data Field Descriptions
# ============================================================================
FIELD_DESCRIPTIONS = {
"input_rgbd": {
"en": "Input RGBD image sequence",
"detail": (
"Shape (T, 4, H, W): T=frames, 4=RGB+Depth channels, H=height, W=width.\n"
" - Channels 0-2: RGB color image, normalized to [0, 1]\n"
" - Channel 3: Depth map"
),
},
"gt_joints_relativeCam_2Dpos": {
"en": "GT joint 2D positions relative to camera (projected)",
"detail": (
"Shape (T, J, 2): T=frames, J=joints(22), 2=pixel coords (x, y).\n"
" 3D body joints projected onto the camera image plane."
),
},
"gt_joints_relativePelvis_3Dpos": {
"en": "GT joint 3D positions relative to pelvis",
"detail": (
"Shape (T, J, 3): T=frames, J=joints(22), 3=(x, y, z).\n"
" 3D joint positions in pelvis-centered local coordinate system.\n"
" Pelvis joint is always at origin (0, 0, 0)."
),
},
"gt_pelvis_camera_3Dpos": {
"en": "GT pelvis 3D position in camera coordinate system",
"detail": (
"Shape (T, 3): T=frames, 3=(x, y, z).\n"
" Absolute 3D position of pelvis in camera space."
),
},
"gt_pelvis_camera_4Drot": {
"en": "GT pelvis rotation in camera space (quaternion)",
"detail": (
"Shape (T, 4): T=frames, 4=quaternion.\n"
" Rotation of pelvis joint in camera coordinate system."
),
},
"hmd_position_global_full_gt_list": {
"en": "HMD global position/orientation data",
"detail": (
"Shape (T, 54): T=frames, 54-dim vector.\n"
" Global pose information from HMD tracking (head + hands)."
),
},
"head_global_trans_list": {
"en": "Head global transformation matrices",
"detail": (
"Shape (T, 4, 4): T=frames, 4x4 homogeneous transformation matrix.\n"
" Head pose (rotation + translation) in global coordinates per frame."
),
},
"body_parms_list": {
"en": "SMPL body model parameters",
"detail": (
"Dictionary with sub-fields:\n"
" - root_orient (T, 3): Root (pelvis) rotation in axis-angle\n"
" - pose_body (T, 63): Body joint poses (21 joints × 3 axis-angle)\n"
" - trans (T, 3): Global translation (x, y, z)"
),
},
"pred_2d": {
"en": "Predicted 2D joint positions",
"detail": (
"Shape (T, J, 2): T=frames, J=joints(22), 2=pixel coords (x, y).\n"
" 2D joint detections from a pretrained pose estimator."
),
},
"pred_3d": {
"en": "Predicted 3D joint positions",
"detail": (
"Shape (T, J, 3): T=frames, J=joints(22), 3=(x, y, z).\n"
" 3D joint predictions from a pretrained pose estimator."
),
},
}
def print_separator(char="=", length=80):
print(char * length)
def print_header(title):
print_separator()
print(f" {title}")
print_separator()
def get_dataset_root():
"""Return dataset root directory"""
return Path(__file__).parent
def load_path_file(filepath):
"""Load path list file"""
if not filepath.exists():
return []
with open(filepath, "r") as f:
return [line.strip() for line in f if line.strip()]
def dataset_overview(root):
"""Print dataset overview"""
print_header("Dataset Overview")
# Scene statistics
scenes = sorted([d for d in os.listdir(root) if d.startswith("Scene") and os.path.isdir(root / d)])
print(f"\nNumber of scenes: {len(scenes)}")
total_seqs = 0
total_npz = 0
source_counter = defaultdict(int)
for scene in scenes:
scene_path = root / scene
seqs = [d for d in os.listdir(scene_path) if os.path.isdir(scene_path / d)]
npz_count = 0
for seq in seqs:
seq_path = scene_path / seq
npz_files = [f for f in os.listdir(seq_path) if f.endswith(".npz")]
npz_count += len(npz_files)
# Data source extraction
if "CMU" in seq:
source_counter["CMU"] += 1
elif "BioMotionLab_NTroje" in seq:
source_counter["BioMotionLab_NTroje"] += 1
else:
source_counter["Other"] += 1
total_seqs += len(seqs)
total_npz += npz_count
print(f" {scene}: {len(seqs):>4} sequences, {npz_count:>5} npz files")
print(f"\nTotal:")
print(f" Sequences: {total_seqs}")
print(f" NPZ files: {total_npz}")
# Data source statistics
print(f"\nData Sources:")
for src, cnt in sorted(source_counter.items(), key=lambda x: -x[1]):
print(f" {src}: {cnt} sequences")
# Train/Val/Test split
print(f"\nTrain/Val/Test Split:")
for split in ["train", "val", "test", "all"]:
path_file = root / f"{split}_npz_paths.txt"
paths = load_path_file(path_file)
print(f" {split:>5}: {len(paths):>6} samples")
print()
def inspect_single_npz(npz_path, verbose=True):
"""Inspect the detailed structure of a single npz file"""
data = np.load(npz_path, allow_pickle=True)
keys = list(data.keys())
if verbose:
print_header(f"NPZ File Details")
print(f"Path: {npz_path}")
print(f"Size: {os.path.getsize(npz_path) / 1024 / 1024:.2f} MB")
print(f"Number of fields: {len(keys)}")
print()
info = {}
for key in keys:
arr = data[key]
field_info = {"key": key, "dtype": str(arr.dtype), "shape": arr.shape}
if arr.dtype == object:
# Handle dictionary type (body_parms_list)
obj = arr.item()
if isinstance(obj, dict):
field_info["type"] = "dict"
field_info["sub_fields"] = {}
for k, v in obj.items():
field_info["sub_fields"][k] = {
"shape": v.shape,
"dtype": str(v.dtype),
"min": float(v.min()),
"max": float(v.max()),
"mean": float(v.mean()),
}
else:
field_info["type"] = "array"
if arr.size > 0 and arr.dtype.kind in ("f", "i", "u"):
field_info["min"] = float(arr.min())
field_info["max"] = float(arr.max())
field_info["mean"] = float(arr.mean())
field_info["std"] = float(arr.std())
info[key] = field_info
if verbose:
desc = FIELD_DESCRIPTIONS.get(key, {})
print_separator("-", 70)
print(f"Field: {key}")
if desc:
print(f" Desc: {desc['en']}")
print(f" Dtype: {arr.dtype}")
print(f" Shape: {arr.shape}")
if field_info["type"] == "dict":
print(f" Sub-fields:")
for k, v_info in field_info["sub_fields"].items():
print(f" - {k}: shape={v_info['shape']}, dtype={v_info['dtype']}, "
f"range=[{v_info['min']:.4f}, {v_info['max']:.4f}], mean={v_info['mean']:.4f}")
elif "min" in field_info:
print(f" Range: [{field_info['min']:.4f}, {field_info['max']:.4f}]")
print(f" Mean: {field_info['mean']:.4f}, Std: {field_info['std']:.4f}")
if desc and "detail" in desc:
print(f" Details:")
for line in desc["detail"].split("\n"):
print(f" {line}")
print()
data.close()
return info
def print_data_schema():
"""Print complete data schema documentation"""
print_header("Data Schema Documentation")
print("""
Each .npz file represents a motion clip with T consecutive frames (typically T=100).
The data is designed for egomotion estimation: recovering 3D human pose from egocentric RGBD views.
Directory Structure:
EgomotionData/
├── Scene{0-6}/ # 7 different virtual scenes
│ └── AllDataPath_{Source}_{split}_{id}/ # motion sequence directory
│ └── {clip_id}.npz # motion clip file
├── train_npz_paths.txt # train set path list
├── val_npz_paths.txt # validation set path list
├── test_npz_paths.txt # test set path list
└── all_npz_paths.txt # all paths list
Data Sources:
- CMU: CMU Graphics Lab Motion Capture Database
- BioMotionLab_NTroje: BioMotionLab (NTroje) Dataset
""")
print("Field Details:")
print_separator("-", 70)
for key, desc in FIELD_DESCRIPTIONS.items():
print(f"\n[{key}]")
print(f" English: {desc['en']}")
print(f" Details:")
for line in desc["detail"].split("\n"):
print(f" {line}")
print()
print("Joint Definition (22 joints):")
print(" SMPL model uses 22 joints, typical order:")
print(" 0:Pelvis 1:L_Hip 2:R_Hip 3:Spine1 4:L_Knee 5:R_Knee")
print(" 6:Spine2 7:L_Ankle 8:R_Ankle 9:Spine3 10:L_Foot 11:R_Foot")
print(" 12:Neck 13:L_Collar 14:R_Collar 15:Head 16:L_Shoulder 17:R_Shoulder")
print(" 18:L_Elbow 19:R_Elbow 20:L_Wrist 21:R_Wrist")
print()
def check_consistency(root, n_samples=10):
"""Randomly sample and check data consistency"""
print_header(f"Consistency Check (N={n_samples})")
all_paths_file = root / "all_npz_paths.txt"
all_paths = load_path_file(all_paths_file)
if not all_paths:
print("Error: all_npz_paths.txt is empty or does not exist")
return
samples = random.sample(all_paths, min(n_samples, len(all_paths)))
issues = []
for i, rel_path in enumerate(samples, 1):
# Path format: EgomotionData/Scene0/.../1.npz
npz_path = root.parent / rel_path
if not npz_path.exists():
npz_path = root / "/".join(rel_path.split("/")[1:])
if not npz_path.exists():
issues.append(f"File does not exist: {rel_path}")
continue
try:
data = np.load(npz_path, allow_pickle=True)
keys = set(data.keys())
expected_keys = set(FIELD_DESCRIPTIONS.keys())
missing = expected_keys - keys
extra = keys - expected_keys
# Check frame consistency
T = data["input_rgbd"].shape[0]
frame_checks = {
"gt_joints_relativeCam_2Dpos": data["gt_joints_relativeCam_2Dpos"].shape[0],
"gt_joints_relativePelvis_3Dpos": data["gt_joints_relativePelvis_3Dpos"].shape[0],
"gt_pelvis_camera_3Dpos": data["gt_pelvis_camera_3Dpos"].shape[0],
"pred_2d": data["pred_2d"].shape[0],
"pred_3d": data["pred_3d"].shape[0],
}
frame_mismatch = {k: v for k, v in frame_checks.items() if v != T}
status = "OK" if (not missing and not extra and not frame_mismatch) else "WARN"
print(f" [{i}/{len(samples)}] {status} - {rel_path} (T={T})")
if missing:
msg = f" Missing fields: {missing}"
print(f" {msg}")
issues.append(msg)
if extra:
print(f" Extra fields: {extra}")
if frame_mismatch:
msg = f" Frame count mismatch: {frame_mismatch}"
print(f" {msg}")
issues.append(msg)
data.close()
except Exception as e:
msg = f"Failed to read: {rel_path} - {e}"
print(f" [{i}/{len(samples)}] ERROR - {msg}")
issues.append(msg)
print()
if issues:
print(f"Found {len(issues)} issues:")
for issue in issues:
print(f" - {issue}")
else:
print("All samples passed consistency check.")
print()
def main():
parser = argparse.ArgumentParser(
description="EgomotionData Dataset Inspector",
formatter_class=argparse.RawDescriptionHelpFormatter,
epilog="""
Examples:
python inspect_dataset.py # Overview + sample check
python inspect_dataset.py --overview-only # Show only statistics overview
python inspect_dataset.py --schema # Show data schema documentation
python inspect_dataset.py --sample Scene0/xxx/1.npz # Check specific file
python inspect_dataset.py --check-consistency 20 # Check 20 random files
""",
)
parser.add_argument("--sample", type=str, default=None,
help="Specify npz file path to inspect")
parser.add_argument("--overview-only", action="store_true",
help="Show only dataset overview statistics")
parser.add_argument("--schema", action="store_true",
help="Show complete data schema documentation")
parser.add_argument("--check-consistency", type=int, default=0, metavar="N",
help="Randomly check consistency of N files")
parser.add_argument("--seed", type=int, default=42,
help="Random seed (default: 42)")
args = parser.parse_args()
random.seed(args.seed)
root = get_dataset_root()
if args.schema:
print_data_schema()
return
if args.sample:
sample_path = Path(args.sample)
if not sample_path.is_absolute():
sample_path = root / args.sample
if not sample_path.exists():
print(f"Error: file does not exist - {sample_path}")
sys.exit(1)
inspect_single_npz(sample_path)
return
# Default: show overview
dataset_overview(root)
if args.overview_only:
return
# Show data schema
print_data_schema()
# Show details of a random sample
all_paths = load_path_file(root / "all_npz_paths.txt")
if all_paths:
sample_rel = random.choice(all_paths)
sample_path = root.parent / sample_rel
if not sample_path.exists():
sample_path = root / "/".join(sample_rel.split("/")[1:])
if sample_path.exists():
print_header("Random Sample Details")
inspect_single_npz(sample_path)
# Consistency check
n_check = args.check_consistency if args.check_consistency > 0 else 5
check_consistency(root, n_check)
if __name__ == "__main__":
main()
|