RDT-1B LIBERO Checkpoint
RDT-1B full fine-tuned on LIBERO-90 dataset (checkpoint-65000). Base model for all LIBERO tasks.
Model Information
- Base Model: RDT-1B (Residual Diffusion Transformer)
- Training Framework: DeepSpeed ZeRO Stage 2
- Precision: BF16
Checkpoint Contents
This checkpoint includes:
For Inference
ema/model.safetensors- EMA model weights (recommended for inference)config.json- Model configuration
For Training
pytorch_model/- DeepSpeed distributed training checkpointbf16_zero_pp_rank_*_optim_states.pt- Optimizer states (ZeRO Stage 2)mp_rank_00_model_states.pt- Model states
scheduler.bin- Learning rate scheduler staterandom_states_*.pkl- Random number generator stateszero_to_fp32.py- Utility to convert DeepSpeed checkpoint to FP32
Usage
For Inference
from transformers import AutoModel
import torch
# Load the EMA model for inference
model = AutoModel.from_pretrained(
"TJ-chen/RDT-1B-LIBERO-Base",
subfolder="ema",
trust_remote_code=True
)
model.eval()
For Continued Training
Download the complete checkpoint and use DeepSpeed to resume training:
# The checkpoint can be loaded with DeepSpeed ZeRO Stage 2
# Make sure your training script is configured with the same DeepSpeed settings
Citation
If you use this model, please cite:
@article{rdt2024,
title={Residual Diffusion Transformer for Robotic Manipulation},
author={Your Name},
journal={arXiv preprint},
year={2024}
}
License
Apache 2.0
- Downloads last month
- 11