zhenyuzhao's picture
Add files using upload-large-folder tool
34df37c verified
# GR00T Deployment & Inference Guide
Run inference with PyTorch or TensorRT acceleration for the GR00T policy.
---
## Prerequisites
- Model checkpoint (e.g., `nvidia/GR00T-N1.6-3B`)
- Dataset in LeRobot format
- CUDA-enabled GPU
### Installation
**PyTorch mode** (default installation):
```bash
uv sync
```
**TensorRT mode** (includes ONNX and TensorRT dependencies):
```bash
uv sync --extra tensorrt
```
---
## Quick Start: PyTorch Mode
```bash
python scripts/deployment/standalone_inference_script.py \
--model-path nvidia/GR00T-N1.6-3B \
--dataset-path /path/to/dataset \
--embodiment-tag GR1 \
--traj-ids 0 1 2 \
--inference-mode pytorch \
--action-horizon 8
```
---
## TensorRT Mode (2x Faster)
### Step 1: Export to ONNX
```bash
python scripts/deployment/export_onnx_n1d6.py \
--model-path nvidia/GR00T-N1.6-3B \
--dataset-path /path/to/dataset \
--embodiment-tag GR1 \
--output-dir ./groot_n1d6_onnx
```
**Output:** `./groot_n1d6_onnx/dit_model.onnx`
### Step 2: Build TensorRT Engine
```bash
python scripts/deployment/build_tensorrt_engine.py \
--onnx ./groot_n1d6_onnx/dit_model.onnx \
--engine ./groot_n1d6_onnx/dit_model_bf16.trt \
--precision bf16
```
**Output:** `./groot_n1d6_onnx/dit_model_bf16.trt`
> **Note:** Engine build takes ~5-10 minutes depending on GPU. The engine is GPU-specific and needs to be rebuilt for different GPU architectures.
### Step 3: Run with TensorRT
```bash
python scripts/deployment/standalone_inference_script.py \
--model-path nvidia/GR00T-N1.6-3B \
--dataset-path /path/to/dataset \
--embodiment-tag GR1 \
--traj-ids 0 1 2 \
--inference-mode tensorrt \
--trt-engine-path ./groot_n1d6_onnx/dit_model_bf16.trt \
--action-horizon 8
```
---
## Command-Line Arguments
### `standalone_inference_script.py`
| Argument | Default | Description |
|----------|---------|-------------|
| `--model-path` | (required) | Path to model checkpoint |
| `--dataset-path` | (required) | Path to LeRobot dataset |
| `--embodiment-tag` | `GR1` | Embodiment tag |
| `--traj-ids` | `[0]` | List of trajectory IDs to evaluate |
| `--steps` | `200` | Max steps per trajectory |
| `--action-horizon` | `16` | Action horizon for inference |
| `--inference-mode` | `pytorch` | `pytorch` or `tensorrt` |
| `--trt-engine-path` | `./groot_n1d6_onnx/dit_model_bf16.trt` | TensorRT engine path |
| `--denoising-steps` | `4` | Number of denoising steps |
| `--skip-timing-steps` | `1` | Steps to skip for timing (warmup) |
| `--seed` | `42` | Random seed for reproducibility |
| `--video-backend` | `torchcodec` | Video backend (`decord`, `torchvision_av`, `torchcodec`) |
### `export_onnx_n1d6.py`
| Argument | Default | Description |
|----------|---------|-------------|
| `--model-path` | (required) | Path to model checkpoint |
| `--dataset-path` | (required) | Path to dataset (for input shape capture) |
| `--embodiment-tag` | `GR1` | Embodiment tag |
| `--output-dir` | `./groot_n1d6_onnx` | Output directory for ONNX model |
| `--video-backend` | `torchcodec` | Video backend |
### `build_tensorrt_engine.py`
| Argument | Default | Description |
|----------|---------|-------------|
| `--onnx` | (required) | Path to ONNX model |
| `--engine` | (required) | Path to save TensorRT engine |
| `--precision` | `bf16` | Precision (`fp32`, `fp16`, `bf16`, `fp8`) |
| `--workspace` | `8192` | Workspace size in MB |
### `benchmark_inference.py`
| Argument | Default | Description |
|----------|---------|-------------|
| `--model-path` | `nvidia/GR00T-N1.6-3B` | Path to model checkpoint |
| `--dataset-path` | `demo_data/gr1.PickNPlace` | Path to dataset |
| `--embodiment-tag` | `GR1` | Embodiment tag |
| `--trt-engine-path` | (optional) | Path to TensorRT engine |
| `--num-iterations` | `20` | Number of benchmark iterations |
| `--warmup` | `5` | Number of warmup iterations |
| `--skip-compile` | `false` | Skip torch.compile benchmark |
| `--seed` | `42` | Random seed for reproducibility |
---
## Benchmarks
### Component-wise Breakdown
> **Note:** The backbone (Vision Encoder + Language Model) timing is the same across all modes (Eager, torch.compile, TensorRT). Only the **Action Head (DiT)** is optimized with torch.compile or TensorRT, which is why you see significant speedups in the Action Head column while the Backbone column remains constant.
GR00T-N1.6-3B inference timing (4 denoising steps):
| Device | Mode | Data Processing | Backbone | Action Head | E2E | Frequency |
|--------|------|-----------------|----------|-------------|-----|-----------|
| RTX 5090 | PyTorch Eager | 2 ms | 18 ms | 38 ms | 58 ms | 17.3 Hz |
| RTX 5090 | torch.compile | 2 ms | 18 ms | 16 ms | 37 ms | 27.3 Hz |
| RTX 5090 | TensorRT | 2 ms | 18 ms | 11 ms | 31 ms | 32.1 Hz |
| H100 | PyTorch Eager | 4 ms | 23 ms | 49 ms | 77 ms | 13.0 Hz |
| H100 | torch.compile | 4 ms | 23 ms | 11 ms | 38 ms | 26.3 Hz |
| H100 | TensorRT | 4 ms | 22 ms | 10 ms | 36 ms | 27.9 Hz |
| RTX 4090 | PyTorch Eager | 2 ms | 25 ms | 55 ms | 82 ms | 12.2 Hz |
| RTX 4090 | torch.compile | 2 ms | 25 ms | 17 ms | 44 ms | 22.8 Hz |
| RTX 4090 | TensorRT | 2 ms | 24 ms | 16 ms | 43 ms | 23.3 Hz |
| Thor | PyTorch Eager | 5 ms | 38 ms | 74 ms | 117 ms | 8.6 Hz |
| Thor | torch.compile | 5 ms | 39 ms | 61 ms | 105 ms | 9.5 Hz |
| Thor | TensorRT | 5 ms | 38 ms | 49 ms | 92 ms | 10.9 Hz |
| Orin | PyTorch Eager | 6 ms | 93 ms | 202 ms | 300 ms | 3.3 Hz |
| Orin | torch.compile | 6 ms | 93 ms | 101 ms | 199 ms | 5.0 Hz |
| Orin | TensorRT | 6 ms | 95 ms | 72 ms | 173 ms | 5.8 Hz |
### Speedup vs PyTorch Eager
| Device | Mode | E2E Speedup | Action Head Speedup |
|--------|------|-------------|---------------------|
| RTX 5090 | PyTorch Eager | 1.00x | 1.00x |
| RTX 5090 | torch.compile | 1.58x | 2.32x |
| RTX 5090 | TensorRT | 1.86x | 3.59x |
| H100 | PyTorch Eager | 1.00x | 1.00x |
| H100 | torch.compile | 2.02x | 4.60x |
| H100 | TensorRT | 2.14x | 4.80x |
| RTX 4090 | PyTorch Eager | 1.00x | 1.00x |
| RTX 4090 | torch.compile | 1.87x | 3.26x |
| RTX 4090 | TensorRT | 1.92x | 3.48x |
| Thor | PyTorch Eager | 1.00x | 1.00x |
| Thor | torch.compile | 1.11x | 1.20x |
| Thor | TensorRT | 1.27x | 1.49x |
| Orin | PyTorch Eager | 1.00x | 1.00x |
| Orin | torch.compile | 1.50x | 2.00x |
| Orin | TensorRT | 1.73x | 2.80x |
> Run `python scripts/deployment/benchmark_inference.py` to generate benchmarks for your hardware.
> See `GR00T_inference_timing.ipynb` for detailed analysis and visualizations.
> Experiments on Thor and Orin used different dependency stacks. Thor with CUDA 13, PyTorch 2.9, using supporting packages sourced from the [Jetson AI Lab cu130 index](https://pypi.jetson-ai-lab.io/sbsa/cu130); and Orin with CUDA 12.6, PyTorch 2.8, using supporting packages sourced from the [Jetson AI Lab cu126 index](https://pypi.jetson-ai-lab.io/jp6/cu126).
---
## Troubleshooting
### Engine Build Fails
- Ensure you have enough GPU memory (8GB+ recommended)
- Try reducing workspace size: `--workspace 4096`
- Ensure TensorRT version matches your CUDA version
### ONNX Export Issues
- If export fails, ensure the model loads correctly in PyTorch first
- Check that the dataset path is valid and contains at least one trajectory
---
## Architecture
```
┌─────────────────────────────────────────────────────────────┐
│ GR00T Policy │
│ ┌───────────────┐ ┌───────────────┐ ┌─────────────────┐ │
│ │ Vision Encoder│ │Language Model │ │ Action Head │ │
│ │(Cosmos-Reason)│──│(Cosmos-Reason)│──│ (DiT) │ │
│ └───────────────┘ └───────────────┘ └─────────────────┘ │
│ ▲ │
│ │ │
│ ┌─────────┴─────────┐ │
│ │ TensorRT Engine │ │
│ │ (dit_model.trt) │ │
│ └───────────────────┘ │
└─────────────────────────────────────────────────────────────┘
```
The TensorRT optimization targets the **DiT (Diffusion Transformer)** component of the action head, which is the main computational bottleneck during inference.
---
## Files
| File | Description |
|------|-------------|
| `standalone_inference_script.py` | Main inference script (PyTorch + TensorRT) |
| `export_onnx_n1d6.py` | Export DiT model to ONNX format |
| `build_tensorrt_engine.py` | Build TensorRT engine from ONNX |
| `benchmark_inference.py` | Benchmark data processing, backbone, action head, and E2E timing |
| `GR00T_inference_timing.ipynb` | Inference timing analysis notebook with visualizations |