Isaac-GR00T / scripts /deployment /benchmark_inference.py
zhenyuzhao's picture
Add files using upload-large-folder tool
34df37c verified
#!/usr/bin/env python3
"""
Benchmark script for GR00T inference timing.
Measures component-wise timing for:
- Data Processing: VLAStepData preparation and collation
- Backbone (VLM): Eagle VLM forward pass
- Action Head (DiT): Flow-matching diffusion model
- E2E: Full end-to-end inference
Supports three inference modes:
1. PyTorch Eager: Standard PyTorch execution
2. torch.compile: PyTorch 2.0+ JIT compilation with max-autotune
3. TensorRT: Optimized DiT action head using TensorRT engine
Usage:
python scripts/deployment/benchmark_inference.py \
--model_path nvidia/GR00T-N1.6-3B \
--dataset_path /path/to/dataset \
--trt_engine_path ./groot_n1d6_onnx/dit_model_bf16.trt
"""
import argparse
import os
import time
import gr00t
from gr00t.data.dataset.lerobot_episode_loader import LeRobotEpisodeLoader
from gr00t.data.dataset.sharded_single_step_dataset import extract_step_data
from gr00t.data.embodiment_tags import EmbodimentTag
from gr00t.data.types import MessageType, VLAStepData
from gr00t.policy.gr00t_policy import Gr00tPolicy
import numpy as np
import torch
def set_seed(seed: int = 42):
"""Set random seed for reproducibility."""
import random
random.seed(seed)
np.random.seed(seed)
torch.manual_seed(seed)
if torch.cuda.is_available():
torch.cuda.manual_seed_all(seed)
def _rec_to_dtype(x, dtype):
"""Recursively convert all floating point tensors to the given dtype."""
if isinstance(x, torch.Tensor) and torch.is_floating_point(x):
return x.to(dtype=dtype)
elif isinstance(x, dict) or hasattr(x, "items"):
return {k: _rec_to_dtype(v, dtype) for k, v in x.items()}
elif isinstance(x, list):
return [_rec_to_dtype(v, dtype) for v in x]
else:
return x
def prepare_model_inputs(policy, observation, return_states=False):
"""
Prepare inputs for the model, mimicking what happens inside _get_action.
Returns collated_inputs that can be passed to model.get_action()
Args:
policy: The Gr00tPolicy instance
observation: Dict with "video", "state", "language" keys
return_states: If True, also return the states list (for action denormalization)
Returns:
collated_inputs if return_states=False, else (collated_inputs, states)
"""
unbatched_obs = []
batch_size = observation["video"][list(observation["video"].keys())[0]].shape[0]
for i in range(batch_size):
unbatched_value = {
"video": {k: v[i] for k, v in observation["video"].items()},
"state": {k: v[i] for k, v in observation["state"].items()},
"language": {k: v[i] for k, v in observation["language"].items()},
}
unbatched_obs.append(unbatched_value)
processed_inputs = []
states = []
for obs in unbatched_obs:
vla_step_data = VLAStepData(
images=obs["video"],
states=obs["state"],
actions={},
text=obs["language"][policy.language_key][0],
embodiment=policy.embodiment_tag,
)
states.append(vla_step_data.states)
messages = [{"type": MessageType.EPISODE_STEP.value, "content": vla_step_data}]
processed_inputs.append(policy.processor(messages))
collated_inputs = policy.collate_fn(processed_inputs)
collated_inputs = collated_inputs["inputs"]
collated_inputs = _rec_to_dtype(collated_inputs, dtype=torch.bfloat16)
if return_states:
return collated_inputs, states
return collated_inputs
def get_device_name():
"""Get short device name for table."""
if torch.cuda.is_available():
name = torch.cuda.get_device_name(0)
# Shorten common names
if "H100" in name:
return "H100"
elif "A100" in name:
return "A100"
elif "RTX 5090" in name:
return "RTX 5090"
elif "RTX 4090" in name:
return "RTX 4090"
elif "RTX 3090" in name:
return "RTX 3090"
elif "Orin" in name:
return "Jetson Orin"
else:
# Return first meaningful part
return name.split()[1] if len(name.split()) > 1 else name
return "CPU"
def compute_e2e_from_components(components):
"""Compute E2E timing as sum of components (more stable than separate measurement)."""
return components["data_processing"] + components["backbone"] + components["action_head"]
def benchmark_data_processing(policy, observation, num_iterations=20, warmup=10):
"""
Benchmark data processing separately with proper warmup.
Data processing is CPU-bound and needs more warmup iterations.
Args:
policy: The Gr00tPolicy instance
observation: Either a single observation dict OR a list of observation dicts (trajectory)
num_iterations: Number of benchmark iterations
warmup: Number of warmup iterations
If observation is a list (trajectory), cycles through observations during benchmarking.
"""
import gc
# Handle both single observation and trajectory (list of observations)
if isinstance(observation, list):
observations = observation
else:
observations = [observation]
num_obs = len(observations)
# Force GC before warmup to reduce variance
gc.collect()
# Warmup - helps with CPU caching and JIT for consistent benchmarks
# For trajectory mode, warmup benefit is reduced since each observation is different
if warmup > 0:
for i in range(warmup):
obs = observations[i % num_obs]
_ = prepare_model_inputs(policy, obs)
# Force GC after warmup
gc.collect()
# Benchmark
times = []
for i in range(num_iterations):
obs = observations[i % num_obs]
start = time.perf_counter()
_ = prepare_model_inputs(policy, obs)
end = time.perf_counter()
times.append(end - start)
return np.array(times) * 1000
def benchmark_components(policy, observation, num_iterations=20, warmup=3):
"""
Benchmark component-wise timing.
Returns dict with times for: data_processing, backbone, action_head
Args:
policy: The Gr00tPolicy instance
observation: Either a single observation dict OR a list of observation dicts (trajectory)
num_iterations: Number of benchmark iterations
warmup: Number of warmup iterations
If observation is a list (trajectory), cycles through observations during benchmarking.
"""
import gc
# Handle both single observation and trajectory (list of observations)
if isinstance(observation, list):
observations = observation
else:
observations = [observation]
num_obs = len(observations)
# Prepare inputs once for backbone/action_head warmup
collated_inputs = prepare_model_inputs(policy, observations[0])
# Warmup backbone + action head
for i in range(warmup):
obs = observations[i % num_obs]
collated_inputs = prepare_model_inputs(policy, obs)
with torch.inference_mode():
backbone_inputs, action_inputs = policy.model.prepare_input(collated_inputs)
backbone_outputs = policy.model.backbone(backbone_inputs)
_ = policy.model.action_head.get_action(backbone_outputs, action_inputs)
torch.cuda.synchronize()
# Force GC before timing
gc.collect()
# Benchmark backbone and action head (GPU-bound)
backbone_times = []
action_head_times = []
for i in range(num_iterations):
obs = observations[i % num_obs]
collated_inputs = prepare_model_inputs(policy, obs)
# Backbone timing
torch.cuda.synchronize()
start = time.perf_counter()
with torch.inference_mode():
backbone_inputs, action_inputs = policy.model.prepare_input(collated_inputs)
backbone_outputs = policy.model.backbone(backbone_inputs)
torch.cuda.synchronize()
end = time.perf_counter()
backbone_times.append(end - start)
# Action head timing
torch.cuda.synchronize()
start = time.perf_counter()
with torch.inference_mode():
_ = policy.model.action_head.get_action(backbone_outputs, action_inputs)
torch.cuda.synchronize()
end = time.perf_counter()
action_head_times.append(end - start)
# Benchmark data processing separately with proper warmup
data_processing_times = benchmark_data_processing(
policy, observation, num_iterations, warmup=10
)
return {
"data_processing": data_processing_times,
"backbone": np.array(backbone_times) * 1000,
"action_head": np.array(action_head_times) * 1000,
}
def print_markdown_table(results, device_name, denoising_steps):
"""Print results as a markdown table using median for robustness."""
print("\n" + "=" * 100)
print("MARKDOWN TABLE (copy/paste into README)")
print("=" * 100)
print(f"\nGR00T-N1.6-3B Inference Timing ({denoising_steps} denoising steps):\n")
# Component breakdown table (using median for robustness against outliers)
print("### Component-wise Breakdown\n")
print("| Device | Mode | Data Processing | Backbone | Action Head | E2E | Frequency |")
print("|--------|------|-----------------|----------|-------------|-----|-----------|")
for mode, data in results.items():
dp_median = np.median(data["data_processing"])
bb_median = np.median(data["backbone"])
ah_median = np.median(data["action_head"])
e2e_median = np.median(data["e2e"])
freq = 1000 / e2e_median
print(
f"| {device_name} | {mode} | {dp_median:.0f} ms | {bb_median:.0f} ms | {ah_median:.0f} ms | {e2e_median:.0f} ms | {freq:.1f} Hz |"
)
# Speedup table
if "PyTorch Eager" in results and len(results) > 1:
print("\n### Speedup vs PyTorch Eager\n")
print("| Device | Mode | E2E Speedup | Action Head Speedup |")
print("|--------|------|-------------|---------------------|")
baseline_e2e = np.median(results["PyTorch Eager"]["e2e"])
baseline_ah = np.median(results["PyTorch Eager"]["action_head"])
for mode, data in results.items():
e2e_median = np.median(data["e2e"])
ah_median = np.median(data["action_head"])
e2e_speedup = baseline_e2e / e2e_median
ah_speedup = baseline_ah / ah_median
print(f"| {device_name} | {mode} | {e2e_speedup:.2f}x | {ah_speedup:.2f}x |")
print("\n" + "=" * 100)
def main():
parser = argparse.ArgumentParser(description="Benchmark GR00T inference timing")
parser.add_argument("--model_path", type=str, default="nvidia/GR00T-N1.6-3B")
parser.add_argument(
"--dataset_path",
type=str,
default=None,
help="Path to dataset. Defaults to demo_data/gr1.PickNPlace",
)
parser.add_argument("--embodiment_tag", type=str, default="gr1")
parser.add_argument(
"--trt_engine_path",
type=str,
default=None,
help="Path to TensorRT engine. If not provided, TensorRT benchmark is skipped.",
)
parser.add_argument("--num_iterations", type=int, default=20)
parser.add_argument("--warmup", type=int, default=5)
parser.add_argument("--seed", type=int, default=42)
parser.add_argument(
"--skip_compile",
action="store_true",
help="Skip torch.compile benchmark (can take a while due to JIT compilation)",
)
parser.add_argument(
"--use_trajectory",
action="store_true",
help="Benchmark on full trajectory instead of single data point. "
"This cycles through all steps in an episode for more realistic benchmarking.",
)
args = parser.parse_args()
set_seed(args.seed)
device = "cuda" if torch.cuda.is_available() else "cpu"
device_name = get_device_name()
# Default dataset path
if args.dataset_path is None:
repo_path = os.path.dirname(os.path.dirname(gr00t.__file__))
args.dataset_path = os.path.join(repo_path, "demo_data/gr1.PickNPlace")
print("=" * 100)
print("GR00T INFERENCE BENCHMARK")
print("=" * 100)
print(
f"Device: {device_name} ({torch.cuda.get_device_name(0) if torch.cuda.is_available() else 'CPU'})"
)
print(f"Model: {args.model_path}")
print(f"Dataset: {args.dataset_path}")
print(f"Iterations: {args.num_iterations}")
print(f"Warmup: {args.warmup}")
print(f"Use Trajectory: {args.use_trajectory}")
print()
# Load dataset and prepare observation
print("Loading policy...")
policy = Gr00tPolicy(
model_path=args.model_path,
embodiment_tag=EmbodimentTag(args.embodiment_tag),
device=device,
strict=True,
)
denoising_steps = policy.model.action_head.num_inference_timesteps
action_horizon = policy.model.action_head.action_horizon
print(f"Action Horizon: {action_horizon}")
print(f"Denoising Steps: {denoising_steps}")
modality_config = policy.get_modality_config()
dataset = LeRobotEpisodeLoader(
dataset_path=args.dataset_path,
modality_configs=modality_config,
video_backend="torchcodec",
)
episode_data = dataset[0]
if args.use_trajectory:
# Load all steps from the episode for trajectory-based benchmarking
# episode_data is a pandas DataFrame, so len() gives the number of steps
trajectory_length = len(episode_data)
observations = []
for step_idx in range(trajectory_length):
try:
step_data = extract_step_data(
episode_data,
step_index=step_idx,
modality_configs=modality_config,
embodiment_tag=EmbodimentTag(args.embodiment_tag),
allow_padding=False,
)
obs = {
"video": {k: np.stack(step_data.images[k])[None] for k in step_data.images},
"state": {k: step_data.states[k][None] for k in step_data.states},
"language": {modality_config["language"].modality_keys[0]: [[step_data.text]]},
}
observations.append(obs)
except Exception:
# Stop if we can't extract more steps (e.g., due to video frame requirements)
break
print(f"Loaded trajectory with {len(observations)} steps")
observation = observations # Pass list to benchmark functions
else:
step_data = extract_step_data(
episode_data,
step_index=0,
modality_configs=modality_config,
embodiment_tag=EmbodimentTag(args.embodiment_tag),
allow_padding=False,
)
observation = {
"video": {k: np.stack(step_data.images[k])[None] for k in step_data.images},
"state": {k: step_data.states[k][None] for k in step_data.states},
"language": {modality_config["language"].modality_keys[0]: [[step_data.text]]},
}
results = {}
# ========================================
# 0. Benchmark Data Processing (shared across all modes)
# ========================================
# Data processing is the same regardless of inference mode (PyTorch/compile/TensorRT)
# so we benchmark it once with proper warmup to get consistent measurements
print("\n" + "-" * 50)
print("Benchmarking Data Processing (shared across all modes)...")
print("-" * 50)
shared_data_processing_times = benchmark_data_processing(
policy, observation, args.num_iterations, warmup=10
)
print(
f" Data Processing: {np.mean(shared_data_processing_times):.2f} ± {np.std(shared_data_processing_times):.2f} ms"
)
# ========================================
# 1. PyTorch Eager
# ========================================
print("\n" + "-" * 50)
print("Benchmarking PyTorch Eager...")
print("-" * 50)
times_components = benchmark_components(policy, observation, args.num_iterations, args.warmup)
components = {
"data_processing": shared_data_processing_times,
"backbone": times_components["backbone"],
"action_head": times_components["action_head"],
}
components["e2e"] = compute_e2e_from_components(components)
results["PyTorch Eager"] = components
e2e_median = np.median(components["e2e"])
print(f" E2E: {e2e_median:.0f} ms ({1000 / e2e_median:.1f} Hz)")
print(f" Data Processing: {np.median(components['data_processing']):.0f} ms")
print(f" Backbone: {np.median(components['backbone']):.0f} ms")
print(f" Action Head: {np.median(components['action_head']):.0f} ms")
# ========================================
# 2. torch.compile
# ========================================
if not args.skip_compile:
print("\n" + "-" * 50)
print("Benchmarking torch.compile (mode='max-autotune')...")
print("(This may take a while due to JIT compilation on first run)")
print("-" * 50)
policy_compiled = Gr00tPolicy(
model_path=args.model_path,
embodiment_tag=EmbodimentTag(args.embodiment_tag),
device=device,
strict=True,
)
policy_compiled.model.action_head.model.forward = torch.compile(
policy_compiled.model.action_head.model.forward, mode="max-autotune"
)
# Enable cuDNN benchmark for additional optimization
if torch.cuda.is_available():
torch.backends.cudnn.benchmark = True
# Extra warmup for torch.compile JIT
times_components = benchmark_components(
policy_compiled, observation, args.num_iterations, warmup=args.warmup + 2
)
components = {
"data_processing": shared_data_processing_times,
"backbone": times_components["backbone"],
"action_head": times_components["action_head"],
}
components["e2e"] = compute_e2e_from_components(components)
results["torch.compile"] = components
e2e_median = np.median(components["e2e"])
print(f" E2E: {e2e_median:.0f} ms ({1000 / e2e_median:.1f} Hz)")
print(f" Data Processing: {np.median(components['data_processing']):.0f} ms")
print(f" Backbone: {np.median(components['backbone']):.0f} ms")
print(f" Action Head: {np.median(components['action_head']):.0f} ms")
# ========================================
# 3. TensorRT (if available)
# ========================================
if args.trt_engine_path and os.path.exists(args.trt_engine_path):
print("\n" + "-" * 50)
print("Benchmarking TensorRT...")
print("-" * 50)
from standalone_inference_script import replace_dit_with_tensorrt
policy_trt = Gr00tPolicy(
model_path=args.model_path,
embodiment_tag=EmbodimentTag(args.embodiment_tag),
device=device,
strict=True,
)
replace_dit_with_tensorrt(policy_trt, args.trt_engine_path)
# TensorRT needs extra warmup for engine initialization and CUDA context setup
trt_warmup = max(args.warmup + 5, 10)
times_components = benchmark_components(
policy_trt, observation, args.num_iterations, warmup=trt_warmup
)
components = {
"data_processing": shared_data_processing_times,
"backbone": times_components["backbone"],
"action_head": times_components["action_head"],
}
components["e2e"] = compute_e2e_from_components(components)
results["TensorRT"] = components
e2e_median = np.median(components["e2e"])
print(f" E2E: {e2e_median:.0f} ms ({1000 / e2e_median:.1f} Hz)")
print(f" Data Processing: {np.median(components['data_processing']):.0f} ms")
print(f" Backbone: {np.median(components['backbone']):.0f} ms")
print(f" Action Head: {np.median(components['action_head']):.0f} ms")
elif args.trt_engine_path:
print(f"\nTensorRT engine not found: {args.trt_engine_path}")
print("To build the engine, run:")
print(
" python scripts/deployment/export_onnx_n1d6.py --model_path nvidia/GR00T-N1.6-3B --output_dir ./groot_n1d6_onnx"
)
print(
" python scripts/deployment/build_tensorrt_engine.py --onnx ./groot_n1d6_onnx/dit_model.onnx --engine <path>.trt --precision bf16"
)
# ========================================
# Print Summary Tables
# ========================================
print_markdown_table(results, device_name, denoising_steps)
# Detailed summary
print("\n" + "=" * 100)
print("DETAILED SUMMARY")
print("=" * 100)
print(f"\nHardware: {torch.cuda.get_device_name(0) if torch.cuda.is_available() else 'CPU'}")
print(f"Model: {args.model_path}")
print(f"Action Horizon: {action_horizon}")
print(f"Denoising Steps: {denoising_steps}")
for mode, data in results.items():
print(f"\n{mode}:")
e2e = data["e2e"]
print(
f" E2E: median={np.median(e2e):.1f} ms, mean={np.mean(e2e):.1f} ± {np.std(e2e):.1f} ms, "
f"min={np.min(e2e):.1f}, max={np.max(e2e):.1f} ({1000 / np.median(e2e):.1f} Hz)"
)
print(f" Data Processing: {np.median(data['data_processing']):.2f} ms (median)")
print(f" Backbone: {np.median(data['backbone']):.2f} ms (median)")
print(f" Action Head: {np.median(data['action_head']):.2f} ms (median)")
print("\n" + "=" * 100)
if __name__ == "__main__":
main()