MoCapDataset / analyze_dataset.py
gourav-wadhwa's picture
adding everything from the dataset
b2e15d5
#!/usr/bin/env python3
"""Dataset analysis utility for the Unitree G1 AMASS trajectories.
Run:
python analyze_dataset.py --root /home/ubuntu/MoCapDataset/AMASSDataset/UnitreeG1
The script outputs two files in the same folder:
├─ episode_stats.json – per-trajectory statistics
└─ aggregate_stats.json – averages / extrema over the whole dataset
Useful for diagnosing AMP training issues (range mismatch, outliers, etc.).
"""
from __future__ import annotations
import argparse
import json
import os
from pathlib import Path
from collections import defaultdict
from typing import Any, Dict, List
import torch
import numpy as np
import isaaclab.utils.math as math_utils
_ALLOWED_EXT = {".pt", ".pth", ".pkl", ".npz"}
# ----------------------------------------------------------------------------------
# helpers
# ----------------------------------------------------------------------------------
def _load_file(path: Path) -> Dict[str, torch.Tensor]:
"""Load torch / numpy trajectory file into a dict of torch tensors."""
if path.suffix in {".npz", ".pkl"}:
data = dict(np.load(path, allow_pickle=True))
for k, v in data.items():
if isinstance(v, np.ndarray) and v.dtype.kind in {"f", "c", "i", "u", "b"}:
data[k] = torch.from_numpy(v)
else:
data[k] = v # keep python objects (e.g. list[str]) as-is
return data # type: ignore[return-value]
# torch files
return torch.load(path, map_location="cpu") # type: ignore[return-value]
def _tensor_range(t: torch.Tensor):
return t.min().item(), t.max().item()
def _analyze_field(value: Any) -> Dict[str, Any]:
"""Analyze a field value and return metadata about it."""
if isinstance(value, torch.Tensor):
return {
"type": "tensor",
"dtype": str(value.dtype),
"shape": list(value.shape),
"ndim": value.ndim,
"size": value.numel(),
}
elif isinstance(value, np.ndarray):
return {
"type": "numpy_array",
"dtype": str(value.dtype),
"shape": list(value.shape),
"ndim": value.ndim,
"size": value.size,
}
elif isinstance(value, list):
return {
"type": "list",
"length": len(value),
"element_type": type(value[0]).__name__ if value else "unknown",
}
elif isinstance(value, (str, int, float, bool)):
return {
"type": type(value).__name__,
"value": value if isinstance(value, (int, float, bool)) else f"<string of length {len(value)}>",
}
else:
return {
"type": type(value).__name__,
"repr": str(value)[:100] + ("..." if len(str(value)) > 100 else ""),
}
def _determine_angular_unit(max_vel: float) -> str:
"""Determine if angular velocity is in deg/s or rad/s based on magnitude."""
return "deg/s" if max_vel > 20.0 else "rad/s"
def _analyze_angular_velocity(velocities: torch.Tensor, name: str = "") -> Dict[str, Any]:
"""Analyze angular velocity data to determine if it's in deg/s or rad/s.
Uses multiple heuristics:
1. Physical limits - robots rarely exceed 1000 deg/s or ~17 rad/s
2. Distribution of values - deg/s values tend to be larger
3. Common ranges for motion capture data
Args:
velocities: Tensor of angular velocities
name: Name of the joint/axis for reporting
Returns:
Dict with analysis results
"""
abs_max = float(torch.abs(velocities).max())
abs_mean = float(torch.abs(velocities).mean())
# Convert to both units for analysis
if abs_max > 20.0: # Assuming it might be deg/s
rad_max = abs_max * np.pi / 180
rad_mean = abs_mean * np.pi / 180
deg_max = abs_max
deg_mean = abs_mean
original_unit = "deg/s"
else: # Assuming it might be rad/s
rad_max = abs_max
rad_mean = abs_mean
deg_max = abs_max * 180 / np.pi
deg_mean = abs_mean * 180 / np.pi
original_unit = "rad/s"
# Scoring system for unit determination
deg_score = 0
rad_score = 0
# Physical limits check
if deg_max > 1000: # Unusually high for deg/s
rad_score += 2
if rad_max > 17: # Unusually high for rad/s (~1000 deg/s)
deg_score += 2
# Common ranges check
if 20 <= deg_max <= 720: # Common range for deg/s in motion capture
deg_score += 1
if 0.3 <= rad_max <= 12: # Common range for rad/s in motion capture
rad_score += 1
# Mean value check
if 5 <= deg_mean <= 180: # Common mean range for deg/s
deg_score += 1
if 0.1 <= rad_mean <= 3: # Common mean range for rad/s
rad_score += 1
# Determine most likely unit
likely_unit = "deg/s" if deg_score > rad_score else "rad/s"
confidence = abs(deg_score - rad_score) / (deg_score + rad_score) if (deg_score + rad_score) > 0 else 0
return {
"likely_unit": likely_unit,
"confidence": confidence,
"max_value": abs_max,
"mean_value": abs_mean,
"deg_score": deg_score,
"rad_score": rad_score,
"analysis": {
"deg/s": {"max": deg_max, "mean": deg_mean},
"rad/s": {"max": rad_max, "mean": rad_mean}
},
"original_unit": original_unit
}
# ----------------------------------------------------------------------------------
# main analysis
# ----------------------------------------------------------------------------------
def analyse_dataset(root: Path) -> None:
# discover files
files: List[Path] = []
for p, _, names in os.walk(root):
for n in names:
if n == "shape_optimized.pkl":
continue
if Path(n).suffix in _ALLOWED_EXT:
files.append(Path(p) / n)
files.sort()
if not files:
raise RuntimeError(f"No trajectory files found under {root}")
print(f"Found {len(files)} trajectory files. Analysing…")
# aggregate accumulators
agg = {
"num_episodes": len(files),
"total_frames": 0,
"lengths": [],
"base_lin_vel_b": {"min": torch.full((3,), torch.inf), "max": torch.full((3,), -torch.inf)},
"base_ang_vel_b": {"min": torch.full((3,), torch.inf), "max": torch.full((3,), -torch.inf)},
"base_height": {"min": torch.tensor(torch.inf), "max": torch.tensor(-torch.inf)},
"base_quat": {"min": torch.full((4,), torch.inf), "max": torch.full((4,), -torch.inf)},
"joint_pos": {}, # type: ignore[dict[str, Any]]
"joint_vel": {}, # type: ignore[dict[str, Any]]
"base_pos": {"min": torch.full((3,), torch.inf), "max": torch.full((3,), -torch.inf)},
"start_pos_list": [],
}
# field analysis accumulator
field_analysis = defaultdict(lambda: {
"count": 0,
"files_with_field": [],
"metadata": None,
"consistent_shape": True,
"shapes_seen": set(),
})
episode_stats: List[Dict[str, Any]] = []
for f_idx, path in enumerate(files):
data = _load_file(path)
# Analyze all fields in this file
for field_name, field_value in data.items():
field_info = field_analysis[field_name]
field_info["count"] += 1
field_info["files_with_field"].append(str(path.relative_to(root)))
# Analyze field metadata
metadata = _analyze_field(field_value)
if field_info["metadata"] is None:
field_info["metadata"] = metadata
# Track shape consistency for tensors/arrays
if metadata["type"] in ["tensor", "numpy_array"]:
shape_tuple = tuple(metadata["shape"])
field_info["shapes_seen"].add(shape_tuple)
if len(field_info["shapes_seen"]) > 1:
field_info["consistent_shape"] = False
# Continue with existing analysis for specific fields
if "qpos" not in data or "qvel" not in data:
print(f"[WARN] {path.name}: Missing qpos or qvel, skipping detailed analysis")
continue
qpos, qvel = data["qpos"].float(), data["qvel"].float()
n = qpos.shape[0]
agg["total_frames"] += n
agg["lengths"].append(n)
# root quantities
base_pos = qpos[:, :3]
base_quat = math_utils.quat_unique(qpos[:, 3:7]) # shape [n,4]
base_lin_vel = qvel[:, :3]
base_ang_vel = qvel[:, 3:6]
base_lin_vel_b = math_utils.quat_rotate_inverse(base_quat, base_lin_vel)
base_ang_vel_b = math_utils.quat_rotate_inverse(base_quat, base_ang_vel)
height = base_pos[:, 2]
# update global min/max
agg["base_height"]["min"] = torch.minimum(agg["base_height"]["min"], height.min())
agg["base_height"]["max"] = torch.maximum(agg["base_height"]["max"], height.max())
for k, tensor in zip(["base_lin_vel_b", "base_ang_vel_b"], [base_lin_vel_b, base_ang_vel_b]):
agg[k]["min"] = torch.minimum(agg[k]["min"], tensor.min(dim=0).values)
agg[k]["max"] = torch.maximum(agg[k]["max"], tensor.max(dim=0).values)
agg["base_quat"]["min"] = torch.minimum(agg["base_quat"]["min"], base_quat.min(dim=0).values)
agg["base_quat"]["max"] = torch.maximum(agg["base_quat"]["max"], base_quat.max(dim=0).values)
# update base_pos ranges
agg["base_pos"]["min"] = torch.minimum(agg["base_pos"]["min"], base_pos.min(dim=0).values)
agg["base_pos"]["max"] = torch.maximum(agg["base_pos"]["max"], base_pos.max(dim=0).values)
# save starting position
agg["start_pos_list"].append(base_pos[0].tolist())
# joint ranges per episode
joint_pos = qpos[:, 7:]
joint_vel = qvel[:, 6:] # skip floating base velocities (first 6)
num_joints = joint_pos.shape[1]
joint_names = data.get("joint_names", None)
if joint_names is not None and len(joint_names) == num_joints + 1:
# Likely the first entry is the floating base/root joint which is not in joint_pos
joint_names = joint_names[1:]
if joint_names is None or len(joint_names) != num_joints:
# Generate fallback names if missing or mismatched
if joint_names is not None and len(joint_names) != num_joints:
print(f"[WARN] {path.name}: joint_names length {len(joint_names)} != joint dim {num_joints}. Using generic names.")
joint_names = [f"joint_{i}" for i in range(num_joints)]
ep_joint_range = {}
ep_joint_vel_range = {}
for j in range(num_joints):
name = joint_names[j]
# Position ranges
j_min, j_max = _tensor_range(joint_pos[:, j])
ep_joint_range[name] = {"min": j_min, "max": j_max}
# Velocity ranges
v_min, v_max = _tensor_range(joint_vel[:, j])
ep_joint_vel_range[name] = {"min": v_min, "max": v_max}
# accumulate global position ranges
if name not in agg["joint_pos"]:
agg["joint_pos"][name] = {"min": j_min, "max": j_max}
else:
agg["joint_pos"][name]["min"] = min(agg["joint_pos"][name]["min"], j_min)
agg["joint_pos"][name]["max"] = max(agg["joint_pos"][name]["max"], j_max)
# accumulate global velocity ranges
if name not in agg["joint_vel"]:
agg["joint_vel"][name] = {"min": v_min, "max": v_max}
else:
agg["joint_vel"][name]["min"] = min(agg["joint_vel"][name]["min"], v_min)
agg["joint_vel"][name]["max"] = max(agg["joint_vel"][name]["max"], v_max)
# Analyze base angular velocities
base_ang_vel_analysis = {}
for i, axis in enumerate(['x', 'y', 'z']):
base_ang_vel_axis = base_ang_vel_b[:, i]
base_ang_vel_analysis[axis] = _analyze_angular_velocity(base_ang_vel_axis, f"base_ang_vel_{axis}")
# Analyze joint velocities
joint_vel_analysis = {}
for j in range(num_joints):
name = joint_names[j]
joint_vel_analysis[name] = _analyze_angular_velocity(joint_vel[:, j], name)
# store per-episode stats
episode_stats.append({
"file": str(path.relative_to(root)),
"length": n,
"base_height": _tensor_range(height),
"base_pos_start": base_pos[0].tolist(),
"base_pos_range": {
"min": base_pos.min(dim=0).values.tolist(),
"max": base_pos.max(dim=0).values.tolist(),
},
"base_lin_vel_b": {
"min": base_lin_vel_b.min(dim=0).values.tolist(),
"max": base_lin_vel_b.max(dim=0).values.tolist(),
},
"base_ang_vel_b": {
"min": base_ang_vel_b.min(dim=0).values.tolist(),
"max": base_ang_vel_b.max(dim=0).values.tolist(),
},
"base_quat": {k: v.tolist() for k, v in agg["base_quat"].items()},
"joint_pos_range": ep_joint_range,
"joint_vel_range": ep_joint_vel_range,
"base_ang_vel_analysis": base_ang_vel_analysis,
"joint_vel_analysis": joint_vel_analysis,
})
if (f_idx + 1) % 50 == 0:
print(f"Processed {f_idx+1}/{len(files)} files…")
# Process field analysis results
# print("\n" + "="*60)
# print("DATASET FIELD ANALYSIS")
# print("="*60)
# all_fields = sorted(field_analysis.keys())
# for field_name in all_fields:
# info = field_analysis[field_name]
# print(f"\nField: '{field_name}'")
# print(f" Present in: {info['count']}/{len(files)} files ({info['count']/len(files)*100:.1f}%)")
# if info['metadata']:
# meta = info['metadata']
# if meta['type'] in ['tensor', 'numpy_array']:
# if info['consistent_shape']:
# print(f" Type: {meta['type']} ({meta['dtype']})")
# print(f" Shape: {meta['shape']}")
# else:
# print(f" Type: {meta['type']} ({meta['dtype']}) - INCONSISTENT SHAPES")
# print(f" Shapes seen: {sorted(info['shapes_seen'])}")
# elif meta['type'] == 'list':
# print(f" Type: {meta['type']} (length: {meta['length']}, elements: {meta['element_type']})")
# else:
# print(f" Type: {meta['type']}")
# if 'value' in meta:
# print(f" Value: {meta['value']}")
# Create field summary for JSON output
field_summary = {}
for field_name, info in field_analysis.items():
field_summary[field_name] = {
"present_in_files": info["count"],
"present_in_percentage": round(info["count"] / len(files) * 100, 1),
"metadata": info["metadata"],
"consistent_shape": info["consistent_shape"],
}
if not info["consistent_shape"]:
field_summary[field_name]["shapes_seen"] = sorted([list(s) for s in info["shapes_seen"]])
# final aggregate statistics
agg_stats = {
"num_episodes": agg["num_episodes"],
"average_length": float(np.mean(agg["lengths"])),
"min_length": int(min(agg["lengths"])),
"max_length": int(max(agg["lengths"])),
"base_height": {k: v.item() if torch.is_tensor(v) else v.tolist() for k, v in agg["base_height"].items()},
"base_lin_vel_b": {k: v.tolist() for k, v in agg["base_lin_vel_b"].items()},
"base_ang_vel_b": {k: v.tolist() for k, v in agg["base_ang_vel_b"].items()},
"base_quat": {k: v.tolist() for k, v in agg["base_quat"].items()},
"joint_pos_global_range": agg["joint_pos"],
"joint_vel_global_range": agg["joint_vel"], # Added joint velocity global ranges
"base_pos": {k: v.tolist() for k, v in agg["base_pos"].items()},
"avg_start_pos": np.mean(agg["start_pos_list"], axis=0).tolist(),
"field_analysis": field_summary,
"angular_velocity_analysis": {
"base": base_ang_vel_analysis,
"joints": joint_vel_analysis,
}
}
# Determine angular-velocity unit (deg/s vs rad/s)
max_ang_vel = max(abs(x) for x in agg_stats["base_ang_vel_b"]["max"])
base_ang_unit = _determine_angular_unit(max_ang_vel)
agg_stats["base_ang_vel_unit"] = base_ang_unit
# Analyze joint velocity units
joint_vel_units = {}
max_joint_vels = {}
for joint_name, vel_range in agg["joint_vel"].items():
max_vel = max(abs(vel_range["min"]), abs(vel_range["max"]))
max_joint_vels[joint_name] = max_vel
joint_vel_units[joint_name] = _determine_angular_unit(max_vel)
# Check if all joints use the same unit
unique_units = set(joint_vel_units.values())
if len(unique_units) == 1:
joint_vel_unit = next(iter(unique_units))
print(f"\nAll joint velocities appear to be in {joint_vel_unit}")
else:
print("\nWARNING: Inconsistent joint velocity units detected:")
for unit in unique_units:
joints = [name for name, u in joint_vel_units.items() if u == unit]
print(f" {unit}: {', '.join(joints)}")
# Add joint velocity analysis to aggregate stats
agg_stats["joint_vel_units"] = joint_vel_units
agg_stats["joint_vel_max_magnitude"] = max_joint_vels
print(f"\nEstimated base angular velocity unit: {base_ang_unit} (max ω ≈ {max_ang_vel:.2f})")
print("Joint velocity analysis added to aggregate_stats.json")
# Print detailed analysis
print("\nAngular Velocity Analysis:")
print("\nBase Angular Velocity:")
for axis, analysis in base_ang_vel_analysis.items():
print(f" {axis}-axis: Likely {analysis['likely_unit']} (confidence: {analysis['confidence']:.2f})")
print(f" Max: {analysis['max_value']:.2f} {analysis['original_unit']}")
print(f" Mean: {analysis['mean_value']:.2f} {analysis['original_unit']}")
print("\nJoint Velocities:")
for joint, analysis in joint_vel_analysis.items():
print(f" {joint}: Likely {analysis['likely_unit']} (confidence: {analysis['confidence']:.2f})")
print(f" Max: {analysis['max_value']:.2f} {analysis['original_unit']}")
print(f" Mean: {analysis['mean_value']:.2f} {analysis['original_unit']}")
# write files
(root / "episode_stats.json").write_text(json.dumps(episode_stats, indent=2))
(root / "aggregate_stats.json").write_text(json.dumps(agg_stats, indent=2))
print(f"\nAnalysis complete. Results saved to episode_stats.json and aggregate_stats.json")
# print(f"Field analysis included for {len(all_fields)} unique fields found across all files.")
# ----------------------------------------------------------------------------------
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Analyse Unitree G1 AMASS dataset")
parser.add_argument("--root", type=str, required=True, help="Root folder of trajectories")
args = parser.parse_args()
analyse_dataset(Path(args.root).expanduser())