| license: mit | |
| pipeline_tag: robotics | |
| tags: | |
| - robotics | |
| - imitation-learning | |
| - diffusion-policy | |
| # SeedPolicy: Horizon Scaling via Self-Evolving Diffusion Policy for Robot Manipulation | |
| This repository contains the pre-trained model checkpoints for the tasks highlighted in the paper **SeedPolicy: Horizon Scaling via Self-Evolving Diffusion Policy for Robot Manipulation**. | |
| ## ๐ Associated Paper & Links | |
| * **Hugging Face Paper Page:** [https://huggingface.co/papers/2603.05117](https://huggingface.co/papers/2603.05117) | |
| * **GitHub Repository:** [https://github.com/Youqiang-Gui/SeedPolicy](https://github.com/Youqiang-Gui/SeedPolicy) | |
| ## ๐ก Overview | |
| SeedPolicy introduces **Self-Evolving Gated Attention (SEGA)**, a temporal module that maintains a time-evolving latent state via gated attention. This enables efficient recurrent updates that compress long-horizon observations into a fixed-size representation while filtering irrelevant temporal information. Integrating SEGA into Diffusion Policy (DP) resolves temporal modeling bottlenecks and enables scalable horizon extension for long-horizon robotic manipulation tasks. | |
| ## ๐ ๏ธ Usage | |
| Detailed installation and data generation instructions are available in the [official GitHub repository](https://github.com/Youqiang-Gui/SeedPolicy). | |
| ### 1. Train Policy | |
| ```bash | |
| bash train.sh ${task_name} ${task_config} ${expert_data_num} ${seed} ${action_dim} ${gpu_id} ${config_name} | |
| # Example: | |
| # bash train.sh beat_block_hammer demo_clean 50 0 14 0 train_diffusion_transformer_hybrid_workspace | |
| ``` | |
| ### 2. Eval Policy | |
| ```bash | |
| bash eval.sh ${task_name} ${task_config} ${ckpt_setting} ${expert_data_num} ${seed} ${gpu_id} ${config_name} ${timestamp} | |
| # Example 1: Standard Evaluation | |
| # bash eval.sh beat_block_hammer demo_clean demo_clean 50 0 0 train_diffusion_transformer_hybrid_workspace "'20260106-143723'" | |
| # Example 2: Generalization Evaluation | |
| # To evaluate a policy trained on the `demo_clean` setting and tested on the `demo_randomized` setting, run: | |
| # bash eval.sh beat_block_hammer demo_randomized demo_clean 50 0 0 train_diffusion_transformer_hybrid_workspace "'20260106-143723'" | |
| ``` | |
| The evaluation results, including videos, will be saved in the `eval_result` directory under the project root. | |
| ## ๐บ Acknowledgements | |
| Our code is generally built upon: [Diffusion Policy](https://github.com/real-stanford/diffusion_policy) and [RoboTwin 2.0](https://github.com/RoboTwin-Platform/RoboTwin). Specifically, the implementation of our state update code references [CUT3R](https://github.com/CUT3R/CUT3R) and [TTT3R](https://github.com/Inception3D/TTT3R). |