quincyu commited on
Commit
921cb8d
·
verified ·
1 Parent(s): 38c0de1

Upload policy weights, train config and readme

Browse files
Files changed (4) hide show
  1. README.md +5 -5
  2. config.json +44 -27
  3. model.safetensors +2 -2
  4. train_config.json +60 -38
README.md CHANGED
@@ -2,20 +2,20 @@
2
  datasets: quincyu/trace_robotwin
3
  library_name: lerobot
4
  license: apache-2.0
5
- model_name: tg_act_266m
6
  pipeline_tag: robotics
7
  tags:
8
- - tg_act_266m
9
- - robotics
10
  - lerobot
 
11
  ---
12
 
13
- # Model Card for tg_act_266m
14
 
15
  <!-- Provide a quick summary of what the model is/does. -->
16
 
17
 
18
- _Model type not recognized please update this template._
19
 
20
 
21
  This policy has been trained and pushed to the Hub using [LeRobot](https://github.com/huggingface/lerobot).
 
2
  datasets: quincyu/trace_robotwin
3
  library_name: lerobot
4
  license: apache-2.0
5
+ model_name: diffusion
6
  pipeline_tag: robotics
7
  tags:
8
+ - diffusion
 
9
  - lerobot
10
+ - robotics
11
  ---
12
 
13
+ # Model Card for diffusion
14
 
15
  <!-- Provide a quick summary of what the model is/does. -->
16
 
17
 
18
+ [Diffusion Policy](https://huggingface.co/papers/2303.04137) treats visuomotor control as a generative diffusion process, producing smooth, multi-step action trajectories that excel at contact-rich manipulation.
19
 
20
 
21
  This policy has been trained and pushed to the Hub using [LeRobot](https://github.com/huggingface/lerobot).
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "type": "tg_act_266m",
3
- "n_obs_steps": 1,
4
  "input_features": {
5
  "observation.state": {
6
  "type": "STATE",
@@ -62,33 +62,50 @@
62
  "tags": null,
63
  "license": null,
64
  "pretrained_path": null,
65
- "chunk_size": 100,
66
- "n_action_steps": 100,
67
- "ee_pixel_image_resolution": 256,
68
  "normalization_mapping": {
69
  "VISUAL": "MEAN_STD",
70
- "STATE": "MEAN_STD",
71
  "ACTION": "MIN_MAX"
72
  },
73
- "vision_backbone": "resnet50",
74
- "pretrained_backbone_weights": "ResNet50_Weights.IMAGENET1K_V2",
75
- "replace_final_stride_with_dilation": false,
76
- "pre_norm": false,
77
- "dim_model": 960,
78
- "n_heads": 16,
79
- "dim_feedforward": 6144,
80
- "feedforward_activation": "relu",
81
- "n_encoder_layers": 7,
82
- "n_decoder_layers": 2,
83
- "use_vae": true,
84
- "latent_dim": 64,
85
- "n_vae_encoder_layers": 6,
86
- "temporal_ensemble_coeff": null,
87
- "dropout": 0.1,
88
- "kl_weight": 10.0,
89
- "optimizer_lr": 1e-05,
90
- "optimizer_weight_decay": 0.0001,
91
- "optimizer_lr_backbone": 1e-05,
92
- "compile_model": false,
93
- "compile_mode": "max-autotune"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
94
  }
 
1
  {
2
+ "type": "diffusion",
3
+ "n_obs_steps": 2,
4
  "input_features": {
5
  "observation.state": {
6
  "type": "STATE",
 
62
  "tags": null,
63
  "license": null,
64
  "pretrained_path": null,
65
+ "horizon": 16,
66
+ "n_action_steps": 8,
 
67
  "normalization_mapping": {
68
  "VISUAL": "MEAN_STD",
69
+ "STATE": "MIN_MAX",
70
  "ACTION": "MIN_MAX"
71
  },
72
+ "drop_n_last_frames": 7,
73
+ "vision_backbone": "resnet18",
74
+ "crop_shape": [
75
+ 224,
76
+ 224
77
+ ],
78
+ "crop_is_random": true,
79
+ "pretrained_backbone_weights": null,
80
+ "use_group_norm": true,
81
+ "spatial_softmax_num_keypoints": 32,
82
+ "use_separate_rgb_encoder_per_camera": false,
83
+ "down_dims": [
84
+ 512,
85
+ 1024,
86
+ 2048
87
+ ],
88
+ "kernel_size": 5,
89
+ "n_groups": 8,
90
+ "diffusion_step_embed_dim": 128,
91
+ "use_film_scale_modulation": true,
92
+ "noise_scheduler_type": "DDPM",
93
+ "num_train_timesteps": 100,
94
+ "beta_schedule": "squaredcos_cap_v2",
95
+ "beta_start": 0.0001,
96
+ "beta_end": 0.02,
97
+ "prediction_type": "epsilon",
98
+ "clip_sample": true,
99
+ "clip_sample_range": 1.0,
100
+ "num_inference_steps": null,
101
+ "do_mask_loss_for_padding": false,
102
+ "optimizer_lr": 0.0001,
103
+ "optimizer_betas": [
104
+ 0.95,
105
+ 0.999
106
+ ],
107
+ "optimizer_eps": 1e-08,
108
+ "optimizer_weight_decay": 1e-06,
109
+ "scheduler_name": "cosine",
110
+ "scheduler_warmup_steps": 500
111
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55436b7f64d5a7bd442085344e0e8aa6096120fedd37b830601a3186a39dee56
3
- size 1064876344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14ff88f12b7d60ea57b8c324b56ab5981af8bf99748f6ca8d71806d8fe15b01e
3
+ size 1083146504
train_config.json CHANGED
@@ -81,8 +81,8 @@
81
  },
82
  "env": null,
83
  "policy": {
84
- "type": "tg_act_266m",
85
- "n_obs_steps": 1,
86
  "input_features": {
87
  "observation.state": {
88
  "type": "STATE",
@@ -144,43 +144,61 @@
144
  "tags": null,
145
  "license": null,
146
  "pretrained_path": null,
147
- "chunk_size": 100,
148
- "n_action_steps": 100,
149
- "ee_pixel_image_resolution": 256,
150
  "normalization_mapping": {
151
  "VISUAL": "MEAN_STD",
152
- "STATE": "MEAN_STD",
153
  "ACTION": "MIN_MAX"
154
  },
155
- "vision_backbone": "resnet50",
156
- "pretrained_backbone_weights": "ResNet50_Weights.IMAGENET1K_V2",
157
- "replace_final_stride_with_dilation": false,
158
- "pre_norm": false,
159
- "dim_model": 960,
160
- "n_heads": 16,
161
- "dim_feedforward": 6144,
162
- "feedforward_activation": "relu",
163
- "n_encoder_layers": 7,
164
- "n_decoder_layers": 2,
165
- "use_vae": true,
166
- "latent_dim": 64,
167
- "n_vae_encoder_layers": 6,
168
- "temporal_ensemble_coeff": null,
169
- "dropout": 0.1,
170
- "kl_weight": 10.0,
171
- "optimizer_lr": 1e-05,
172
- "optimizer_weight_decay": 0.0001,
173
- "optimizer_lr_backbone": 1e-05,
174
- "compile_model": false,
175
- "compile_mode": "max-autotune"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
176
  },
177
- "output_dir": "outputs/train/2026-02-12/19-23-15_tg_act_266m",
178
- "job_name": "tg_act_266m",
179
  "resume": false,
180
  "seed": 0,
181
- "num_workers": 4,
182
- "batch_size": 128,
183
- "steps": 100000,
 
184
  "eval_freq": 20000,
185
  "log_freq": 200,
186
  "tolerance_s": 0.0001,
@@ -188,17 +206,21 @@
188
  "save_freq": 25000,
189
  "use_policy_training_preset": true,
190
  "optimizer": {
191
- "type": "adamw",
192
- "lr": 1e-05,
193
- "weight_decay": 0.0001,
194
  "grad_clip_norm": 10.0,
195
  "betas": [
196
- 0.9,
197
  0.999
198
  ],
199
  "eps": 1e-08
200
  },
201
- "scheduler": null,
 
 
 
 
202
  "eval": {
203
  "n_episodes": 50,
204
  "batch_size": 50,
@@ -210,7 +232,7 @@
210
  "project": "lerobot",
211
  "entity": null,
212
  "notes": null,
213
- "run_id": "r5hwbuha",
214
  "mode": null
215
  },
216
  "peft": null,
 
81
  },
82
  "env": null,
83
  "policy": {
84
+ "type": "diffusion",
85
+ "n_obs_steps": 2,
86
  "input_features": {
87
  "observation.state": {
88
  "type": "STATE",
 
144
  "tags": null,
145
  "license": null,
146
  "pretrained_path": null,
147
+ "horizon": 16,
148
+ "n_action_steps": 8,
 
149
  "normalization_mapping": {
150
  "VISUAL": "MEAN_STD",
151
+ "STATE": "MIN_MAX",
152
  "ACTION": "MIN_MAX"
153
  },
154
+ "drop_n_last_frames": 7,
155
+ "vision_backbone": "resnet18",
156
+ "crop_shape": [
157
+ 224,
158
+ 224
159
+ ],
160
+ "crop_is_random": true,
161
+ "pretrained_backbone_weights": null,
162
+ "use_group_norm": true,
163
+ "spatial_softmax_num_keypoints": 32,
164
+ "use_separate_rgb_encoder_per_camera": false,
165
+ "down_dims": [
166
+ 512,
167
+ 1024,
168
+ 2048
169
+ ],
170
+ "kernel_size": 5,
171
+ "n_groups": 8,
172
+ "diffusion_step_embed_dim": 128,
173
+ "use_film_scale_modulation": true,
174
+ "noise_scheduler_type": "DDPM",
175
+ "num_train_timesteps": 100,
176
+ "beta_schedule": "squaredcos_cap_v2",
177
+ "beta_start": 0.0001,
178
+ "beta_end": 0.02,
179
+ "prediction_type": "epsilon",
180
+ "clip_sample": true,
181
+ "clip_sample_range": 1.0,
182
+ "num_inference_steps": null,
183
+ "do_mask_loss_for_padding": false,
184
+ "optimizer_lr": 0.0001,
185
+ "optimizer_betas": [
186
+ 0.95,
187
+ 0.999
188
+ ],
189
+ "optimizer_eps": 1e-08,
190
+ "optimizer_weight_decay": 1e-06,
191
+ "scheduler_name": "cosine",
192
+ "scheduler_warmup_steps": 500
193
  },
194
+ "output_dir": "outputs/train/2026-02-13/16-30-33_diffusion",
195
+ "job_name": "diffusion",
196
  "resume": false,
197
  "seed": 0,
198
+ "num_workers": 8,
199
+ "video_backend": null,
200
+ "batch_size": 256,
201
+ "steps": 50000,
202
  "eval_freq": 20000,
203
  "log_freq": 200,
204
  "tolerance_s": 0.0001,
 
206
  "save_freq": 25000,
207
  "use_policy_training_preset": true,
208
  "optimizer": {
209
+ "type": "adam",
210
+ "lr": 0.0001,
211
+ "weight_decay": 1e-06,
212
  "grad_clip_norm": 10.0,
213
  "betas": [
214
+ 0.95,
215
  0.999
216
  ],
217
  "eps": 1e-08
218
  },
219
+ "scheduler": {
220
+ "type": "diffuser",
221
+ "num_warmup_steps": 500,
222
+ "name": "cosine"
223
+ },
224
  "eval": {
225
  "n_episodes": 50,
226
  "batch_size": 50,
 
232
  "project": "lerobot",
233
  "entity": null,
234
  "notes": null,
235
+ "run_id": "2oralbh2",
236
  "mode": null
237
  },
238
  "peft": null,