bryandong commited on
Commit
9f99dae
·
verified ·
1 Parent(s): 185ee96

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - diffusion-policy
5
+ - imitation-learning
6
+ - behavioral-cloning
7
+ - robotics
8
+ - push-task
9
+ library_name: pytorch
10
+ pipeline_tag: robotics
11
+ ---
12
+
13
+ # Push-F Diffusion Policy
14
+
15
+ A visuomotor diffusion policy trained to push an F-shaped block into a target orientation, adapted from the [Diffusion Policy](https://diffusion-policy.cs.columbia.edu/) codebase (Chi et al., 2023).
16
+
17
+ ## Model Description
18
+
19
+ - **Architecture:** Diffusion UNet with ResNet18 image encoder
20
+ - **Parameters:** 278M
21
+ - **Observations:** 96x96 RGB image + 2D agent position
22
+ - **Actions:** 2D target position for the agent
23
+ - **Training data:** 101 human demonstrations (~29,800 timesteps)
24
+ - **Training:** 250 epochs on NVIDIA H100, ~3.5 hours
25
+ - **Framework:** PyTorch 2.0.1
26
+
27
+ ## Performance
28
+
29
+ Evaluated on 50 held-out environment seeds:
30
+
31
+ | Time Limit | Mean Score | Perfect Seeds (1.0) |
32
+ |:----------:|:----------:|:-------------------:|
33
+ | 30s | 0.837 | 19/50 |
34
+ | 45s | 0.945 | 38/50 |
35
+ | 60s | 0.961 | 45/50 |
36
+ | 90s | 1.000 | 50/50 |
37
+
38
+ ## Usage
39
+
40
+ ```bash
41
+ git clone https://github.com/bryandong24/reu_adaptation.git
42
+ cd reu_adaptation
43
+
44
+ # Set up environment
45
+ mamba env create -f conda_environment.yaml
46
+ conda activate robodiff
47
+ pip install torch==2.0.1+cu118 torchvision==0.15.2+cu118 --extra-index-url https://download.pytorch.org/whl/cu118
48
+ pip install -e .
49
+
50
+ # Download checkpoint and evaluate
51
+ python eval.py --checkpoint epoch=0250-test_mean_score=0.880.ckpt -o eval_output
52
+ ```
53
+
54
+ ## Training Details
55
+
56
+ - **Loss:** MSE denoising loss (DDPM)
57
+ - **Optimizer:** AdamW (lr=1e-4, weight_decay=1e-6)
58
+ - **LR Schedule:** Cosine with 500-step warmup
59
+ - **Batch size:** 64
60
+ - **Horizon:** 16 steps (n_obs=2, n_action=8)
61
+ - **Diffusion steps:** 100 (training), 100 (inference)
62
+ - **EMA:** Enabled
63
+
64
+ ## Citation
65
+
66
+ Based on:
67
+ ```bibtex
68
+ @inproceedings{chi2023diffusionpolicy,
69
+ title={Diffusion Policy: Visuomotor Policy Learning via Action Diffusion},
70
+ author={Chi, Cheng and Feng, Siyuan and Du, Yilun and Xu, Zhenjia and Cousineau, Eric and Burchfiel, Benjamin and Song, Shuran},
71
+ booktitle={Proceedings of Robotics: Science and Systems (RSS)},
72
+ year={2023}
73
+ }
74
+ ```
75
+
76
+ ## Links
77
+
78
+ - [Full Report](https://github.com/bryandong24/reu_adaptation/blob/main/REPORT.md)
79
+ - [Source Code](https://github.com/bryandong24/reu_adaptation)
80
+ - [W&B Training Logs](https://wandb.ai/bryandong24-stanford-university/diffusion_policy_pushf)