Upload policy weights, train config and readme
Browse files- README.md +4 -4
- config.json +8 -7
- model.safetensors +2 -2
- train_config.json +34 -16
README.md
CHANGED
|
@@ -5,9 +5,9 @@ license: apache-2.0
|
|
| 5 |
model_name: diffusion
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
| 8 |
-
- robotics
|
| 9 |
-
- diffusion
|
| 10 |
- lerobot
|
|
|
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for diffusion
|
|
@@ -31,7 +31,7 @@ Below is the short version on how to train and run inference/eval:
|
|
| 31 |
### Train from scratch
|
| 32 |
|
| 33 |
```bash
|
| 34 |
-
|
| 35 |
--dataset.repo_id=${HF_USER}/<dataset> \
|
| 36 |
--policy.type=act \
|
| 37 |
--output_dir=outputs/train/<desired_policy_repo_id> \
|
|
@@ -46,7 +46,7 @@ _Writes checkpoints to `outputs/train/<desired_policy_repo_id>/checkpoints/`._
|
|
| 46 |
### Evaluate the policy/run inference
|
| 47 |
|
| 48 |
```bash
|
| 49 |
-
|
| 50 |
--robot.type=so100_follower \
|
| 51 |
--dataset.repo_id=<hf_user>/eval_<dataset> \
|
| 52 |
--policy.path=<hf_user>/<desired_policy_repo_id> \
|
|
|
|
| 5 |
model_name: diffusion
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
|
|
|
|
|
|
| 8 |
- lerobot
|
| 9 |
+
- diffusion
|
| 10 |
+
- robotics
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for diffusion
|
|
|
|
| 31 |
### Train from scratch
|
| 32 |
|
| 33 |
```bash
|
| 34 |
+
lerobot-train \
|
| 35 |
--dataset.repo_id=${HF_USER}/<dataset> \
|
| 36 |
--policy.type=act \
|
| 37 |
--output_dir=outputs/train/<desired_policy_repo_id> \
|
|
|
|
| 46 |
### Evaluate the policy/run inference
|
| 47 |
|
| 48 |
```bash
|
| 49 |
+
lerobot-record \
|
| 50 |
--robot.type=so100_follower \
|
| 51 |
--dataset.repo_id=<hf_user>/eval_<dataset> \
|
| 52 |
--policy.path=<hf_user>/<desired_policy_repo_id> \
|
config.json
CHANGED
|
@@ -1,11 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"type": "diffusion",
|
| 3 |
"n_obs_steps": 2,
|
| 4 |
-
"normalization_mapping": {
|
| 5 |
-
"VISUAL": "MEAN_STD",
|
| 6 |
-
"STATE": "MIN_MAX",
|
| 7 |
-
"ACTION": "MIN_MAX"
|
| 8 |
-
},
|
| 9 |
"input_features": {
|
| 10 |
"observation.images.cam0_img": {
|
| 11 |
"type": "VISUAL",
|
|
@@ -45,13 +40,19 @@
|
|
| 45 |
"private": null,
|
| 46 |
"tags": null,
|
| 47 |
"license": null,
|
|
|
|
| 48 |
"horizon": 16,
|
| 49 |
"n_action_steps": 8,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
"drop_n_last_frames": 7,
|
| 51 |
"vision_backbone": "resnet18",
|
| 52 |
"crop_shape": [
|
| 53 |
-
|
| 54 |
-
|
| 55 |
],
|
| 56 |
"crop_is_random": true,
|
| 57 |
"pretrained_backbone_weights": null,
|
|
|
|
| 1 |
{
|
| 2 |
"type": "diffusion",
|
| 3 |
"n_obs_steps": 2,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
"input_features": {
|
| 5 |
"observation.images.cam0_img": {
|
| 6 |
"type": "VISUAL",
|
|
|
|
| 40 |
"private": null,
|
| 41 |
"tags": null,
|
| 42 |
"license": null,
|
| 43 |
+
"pretrained_path": null,
|
| 44 |
"horizon": 16,
|
| 45 |
"n_action_steps": 8,
|
| 46 |
+
"normalization_mapping": {
|
| 47 |
+
"VISUAL": "MEAN_STD",
|
| 48 |
+
"STATE": "MIN_MAX",
|
| 49 |
+
"ACTION": "MIN_MAX"
|
| 50 |
+
},
|
| 51 |
"drop_n_last_frames": 7,
|
| 52 |
"vision_backbone": "resnet18",
|
| 53 |
"crop_shape": [
|
| 54 |
+
84,
|
| 55 |
+
84
|
| 56 |
],
|
| 57 |
"crop_is_random": true,
|
| 58 |
"pretrained_backbone_weights": null,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c39b3ba90e434037ae94e00c0c6aced105738fe65ca3f9071ac9c0cfeb049750
|
| 3 |
+
size 356686904
|
train_config.json
CHANGED
|
@@ -4,7 +4,7 @@
|
|
| 4 |
"root": null,
|
| 5 |
"episodes": null,
|
| 6 |
"image_transforms": {
|
| 7 |
-
"enable":
|
| 8 |
"max_num_transforms": 3,
|
| 9 |
"random_order": false,
|
| 10 |
"tfs": {
|
|
@@ -57,22 +57,32 @@
|
|
| 57 |
1.5
|
| 58 |
]
|
| 59 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
}
|
| 61 |
}
|
| 62 |
},
|
| 63 |
"revision": null,
|
| 64 |
"use_imagenet_stats": true,
|
| 65 |
-
"video_backend": "torchcodec"
|
|
|
|
| 66 |
},
|
| 67 |
"env": null,
|
| 68 |
"policy": {
|
| 69 |
"type": "diffusion",
|
| 70 |
"n_obs_steps": 2,
|
| 71 |
-
"normalization_mapping": {
|
| 72 |
-
"VISUAL": "MEAN_STD",
|
| 73 |
-
"STATE": "MIN_MAX",
|
| 74 |
-
"ACTION": "MIN_MAX"
|
| 75 |
-
},
|
| 76 |
"input_features": {
|
| 77 |
"observation.images.cam0_img": {
|
| 78 |
"type": "VISUAL",
|
|
@@ -112,13 +122,19 @@
|
|
| 112 |
"private": null,
|
| 113 |
"tags": null,
|
| 114 |
"license": null,
|
|
|
|
| 115 |
"horizon": 16,
|
| 116 |
"n_action_steps": 8,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 117 |
"drop_n_last_frames": 7,
|
| 118 |
"vision_backbone": "resnet18",
|
| 119 |
"crop_shape": [
|
| 120 |
-
|
| 121 |
-
|
| 122 |
],
|
| 123 |
"crop_is_random": true,
|
| 124 |
"pretrained_backbone_weights": null,
|
|
@@ -154,13 +170,13 @@
|
|
| 154 |
"scheduler_name": "cosine",
|
| 155 |
"scheduler_warmup_steps": 500
|
| 156 |
},
|
| 157 |
-
"output_dir": "/scratch/
|
| 158 |
"job_name": "diffusion",
|
| 159 |
"resume": false,
|
| 160 |
"seed": 1000,
|
| 161 |
-
"num_workers":
|
| 162 |
-
"batch_size":
|
| 163 |
-
"steps":
|
| 164 |
"eval_freq": 20000,
|
| 165 |
"log_freq": 200,
|
| 166 |
"save_checkpoint": true,
|
|
@@ -190,10 +206,12 @@
|
|
| 190 |
"wandb": {
|
| 191 |
"enable": false,
|
| 192 |
"disable_artifact": false,
|
| 193 |
-
"project": "
|
| 194 |
-
"entity": "
|
| 195 |
"notes": null,
|
| 196 |
"run_id": null,
|
| 197 |
"mode": null
|
| 198 |
-
}
|
|
|
|
|
|
|
| 199 |
}
|
|
|
|
| 4 |
"root": null,
|
| 5 |
"episodes": null,
|
| 6 |
"image_transforms": {
|
| 7 |
+
"enable": false,
|
| 8 |
"max_num_transforms": 3,
|
| 9 |
"random_order": false,
|
| 10 |
"tfs": {
|
|
|
|
| 57 |
1.5
|
| 58 |
]
|
| 59 |
}
|
| 60 |
+
},
|
| 61 |
+
"affine": {
|
| 62 |
+
"weight": 1.0,
|
| 63 |
+
"type": "RandomAffine",
|
| 64 |
+
"kwargs": {
|
| 65 |
+
"degrees": [
|
| 66 |
+
-5.0,
|
| 67 |
+
5.0
|
| 68 |
+
],
|
| 69 |
+
"translate": [
|
| 70 |
+
0.05,
|
| 71 |
+
0.05
|
| 72 |
+
]
|
| 73 |
+
}
|
| 74 |
}
|
| 75 |
}
|
| 76 |
},
|
| 77 |
"revision": null,
|
| 78 |
"use_imagenet_stats": true,
|
| 79 |
+
"video_backend": "torchcodec",
|
| 80 |
+
"streaming": false
|
| 81 |
},
|
| 82 |
"env": null,
|
| 83 |
"policy": {
|
| 84 |
"type": "diffusion",
|
| 85 |
"n_obs_steps": 2,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 86 |
"input_features": {
|
| 87 |
"observation.images.cam0_img": {
|
| 88 |
"type": "VISUAL",
|
|
|
|
| 122 |
"private": null,
|
| 123 |
"tags": null,
|
| 124 |
"license": null,
|
| 125 |
+
"pretrained_path": null,
|
| 126 |
"horizon": 16,
|
| 127 |
"n_action_steps": 8,
|
| 128 |
+
"normalization_mapping": {
|
| 129 |
+
"VISUAL": "MEAN_STD",
|
| 130 |
+
"STATE": "MIN_MAX",
|
| 131 |
+
"ACTION": "MIN_MAX"
|
| 132 |
+
},
|
| 133 |
"drop_n_last_frames": 7,
|
| 134 |
"vision_backbone": "resnet18",
|
| 135 |
"crop_shape": [
|
| 136 |
+
84,
|
| 137 |
+
84
|
| 138 |
],
|
| 139 |
"crop_is_random": true,
|
| 140 |
"pretrained_backbone_weights": null,
|
|
|
|
| 170 |
"scheduler_name": "cosine",
|
| 171 |
"scheduler_warmup_steps": 500
|
| 172 |
},
|
| 173 |
+
"output_dir": "/home/shahha35/scratch/runs/lerobot/10490694/checkpoints",
|
| 174 |
"job_name": "diffusion",
|
| 175 |
"resume": false,
|
| 176 |
"seed": 1000,
|
| 177 |
+
"num_workers": 8,
|
| 178 |
+
"batch_size": 32,
|
| 179 |
+
"steps": 100000,
|
| 180 |
"eval_freq": 20000,
|
| 181 |
"log_freq": 200,
|
| 182 |
"save_checkpoint": true,
|
|
|
|
| 206 |
"wandb": {
|
| 207 |
"enable": false,
|
| 208 |
"disable_artifact": false,
|
| 209 |
+
"project": "lerobot_dvrk_DP",
|
| 210 |
+
"entity": "hj-shah",
|
| 211 |
"notes": null,
|
| 212 |
"run_id": null,
|
| 213 |
"mode": null
|
| 214 |
+
},
|
| 215 |
+
"checkpoint_path": null,
|
| 216 |
+
"rename_map": {}
|
| 217 |
}
|