Add robotics pipeline tag and improve model card

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +39 -5
README.md CHANGED
@@ -1,16 +1,50 @@
1
  ---
2
  license: mit
 
 
 
 
 
3
  ---
4
- This repository contains the pre-trained model checkpoints for the three typical tasks highlighted in our paper: **SeedPolicy: Horizon Scaling via Self-Evolving Diffusion Policy for Robot Manipulation**.
 
 
 
5
 
6
  ## πŸ“„ Associated Paper & Links
7
 
8
- * **Hugging Face Paper Page:** https://huggingface.co/papers/2603.05117
 
9
 
10
- * **GitHub Repository:** [https://github.com/Youqiang-Gui/SeedPolicy]
11
 
 
12
 
 
13
 
14
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- *(For detailed installation, usage instructions, and data generation, please refer to our main GitHub repository.)*
 
 
1
  ---
2
  license: mit
3
+ pipeline_tag: robotics
4
+ tags:
5
+ - robotics
6
+ - imitation-learning
7
+ - diffusion-policy
8
  ---
9
+
10
+ # SeedPolicy: Horizon Scaling via Self-Evolving Diffusion Policy for Robot Manipulation
11
+
12
+ This repository contains the pre-trained model checkpoints for the tasks highlighted in the paper **SeedPolicy: Horizon Scaling via Self-Evolving Diffusion Policy for Robot Manipulation**.
13
 
14
  ## πŸ“„ Associated Paper & Links
15
 
16
+ * **Hugging Face Paper Page:** [https://huggingface.co/papers/2603.05117](https://huggingface.co/papers/2603.05117)
17
+ * **GitHub Repository:** [https://github.com/Youqiang-Gui/SeedPolicy](https://github.com/Youqiang-Gui/SeedPolicy)
18
 
19
+ ## πŸ’‘ Overview
20
 
21
+ SeedPolicy introduces **Self-Evolving Gated Attention (SEGA)**, a temporal module that maintains a time-evolving latent state via gated attention. This enables efficient recurrent updates that compress long-horizon observations into a fixed-size representation while filtering irrelevant temporal information. Integrating SEGA into Diffusion Policy (DP) resolves temporal modeling bottlenecks and enables scalable horizon extension for long-horizon robotic manipulation tasks.
22
 
23
+ ## πŸ› οΈ Usage
24
 
25
+ Detailed installation and data generation instructions are available in the [official GitHub repository](https://github.com/Youqiang-Gui/SeedPolicy).
26
+
27
+ ### 1. Train Policy
28
+ ```bash
29
+ bash train.sh ${task_name} ${task_config} ${expert_data_num} ${seed} ${action_dim} ${gpu_id} ${config_name}
30
+
31
+ # Example:
32
+ # bash train.sh beat_block_hammer demo_clean 50 0 14 0 train_diffusion_transformer_hybrid_workspace
33
+ ```
34
+
35
+ ### 2. Eval Policy
36
+ ```bash
37
+ bash eval.sh ${task_name} ${task_config} ${ckpt_setting} ${expert_data_num} ${seed} ${gpu_id} ${config_name} ${timestamp}
38
+
39
+ # Example 1: Standard Evaluation
40
+ # bash eval.sh beat_block_hammer demo_clean demo_clean 50 0 0 train_diffusion_transformer_hybrid_workspace "'20260106-143723'"
41
+
42
+ # Example 2: Generalization Evaluation
43
+ # To evaluate a policy trained on the `demo_clean` setting and tested on the `demo_randomized` setting, run:
44
+ # bash eval.sh beat_block_hammer demo_randomized demo_clean 50 0 0 train_diffusion_transformer_hybrid_workspace "'20260106-143723'"
45
+ ```
46
+
47
+ The evaluation results, including videos, will be saved in the `eval_result` directory under the project root.
48
 
49
+ ## 😺 Acknowledgements
50
+ Our code is generally built upon: [Diffusion Policy](https://github.com/real-stanford/diffusion_policy) and [RoboTwin 2.0](https://github.com/RoboTwin-Platform/RoboTwin). Specifically, the implementation of our state update code references [CUT3R](https://github.com/CUT3R/CUT3R) and [TTT3R](https://github.com/Inception3D/TTT3R).