Upload policy weights, train config and readme
Browse files- README.md +3 -3
- config.json +5 -6
- model.safetensors +2 -2
- train_config.json +7 -9
README.md
CHANGED
|
@@ -5,9 +5,9 @@ license: apache-2.0
|
|
| 5 |
model_name: act
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
| 8 |
-
- lerobot
|
| 9 |
- robotics
|
| 10 |
- act
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for act
|
|
@@ -31,7 +31,7 @@ Below is the short version on how to train and run inference/eval:
|
|
| 31 |
### Train from scratch
|
| 32 |
|
| 33 |
```bash
|
| 34 |
-
|
| 35 |
--dataset.repo_id=${HF_USER}/<dataset> \
|
| 36 |
--policy.type=act \
|
| 37 |
--output_dir=outputs/train/<desired_policy_repo_id> \
|
|
@@ -46,7 +46,7 @@ _Writes checkpoints to `outputs/train/<desired_policy_repo_id>/checkpoints/`._
|
|
| 46 |
### Evaluate the policy/run inference
|
| 47 |
|
| 48 |
```bash
|
| 49 |
-
|
| 50 |
--robot.type=so100_follower \
|
| 51 |
--dataset.repo_id=<hf_user>/eval_<dataset> \
|
| 52 |
--policy.path=<hf_user>/<desired_policy_repo_id> \
|
|
|
|
| 5 |
model_name: act
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
|
|
|
| 8 |
- robotics
|
| 9 |
- act
|
| 10 |
+
- lerobot
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for act
|
|
|
|
| 31 |
### Train from scratch
|
| 32 |
|
| 33 |
```bash
|
| 34 |
+
python -m lerobot.scripts.train \
|
| 35 |
--dataset.repo_id=${HF_USER}/<dataset> \
|
| 36 |
--policy.type=act \
|
| 37 |
--output_dir=outputs/train/<desired_policy_repo_id> \
|
|
|
|
| 46 |
### Evaluate the policy/run inference
|
| 47 |
|
| 48 |
```bash
|
| 49 |
+
python -m lerobot.record \
|
| 50 |
--robot.type=so100_follower \
|
| 51 |
--dataset.repo_id=<hf_user>/eval_<dataset> \
|
| 52 |
--policy.path=<hf_user>/<desired_policy_repo_id> \
|
config.json
CHANGED
|
@@ -1,6 +1,11 @@
|
|
| 1 |
{
|
| 2 |
"type": "act",
|
| 3 |
"n_obs_steps": 1,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
"input_features": {
|
| 5 |
"observation.state": {
|
| 6 |
"type": "STATE",
|
|
@@ -48,14 +53,8 @@
|
|
| 48 |
"private": null,
|
| 49 |
"tags": null,
|
| 50 |
"license": null,
|
| 51 |
-
"pretrained_path": null,
|
| 52 |
"chunk_size": 100,
|
| 53 |
"n_action_steps": 100,
|
| 54 |
-
"normalization_mapping": {
|
| 55 |
-
"VISUAL": "MEAN_STD",
|
| 56 |
-
"STATE": "MEAN_STD",
|
| 57 |
-
"ACTION": "MEAN_STD"
|
| 58 |
-
},
|
| 59 |
"vision_backbone": "resnet18",
|
| 60 |
"pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
|
| 61 |
"replace_final_stride_with_dilation": false,
|
|
|
|
| 1 |
{
|
| 2 |
"type": "act",
|
| 3 |
"n_obs_steps": 1,
|
| 4 |
+
"normalization_mapping": {
|
| 5 |
+
"VISUAL": "MEAN_STD",
|
| 6 |
+
"STATE": "MEAN_STD",
|
| 7 |
+
"ACTION": "MEAN_STD"
|
| 8 |
+
},
|
| 9 |
"input_features": {
|
| 10 |
"observation.state": {
|
| 11 |
"type": "STATE",
|
|
|
|
| 53 |
"private": null,
|
| 54 |
"tags": null,
|
| 55 |
"license": null,
|
|
|
|
| 56 |
"chunk_size": 100,
|
| 57 |
"n_action_steps": 100,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 58 |
"vision_backbone": "resnet18",
|
| 59 |
"pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
|
| 60 |
"replace_final_stride_with_dilation": false,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81e69e4c9195380f237f8a21875e9bba184938713f374cc040f24f7935a78882
|
| 3 |
+
size 206701336
|
train_config.json
CHANGED
|
@@ -62,13 +62,17 @@
|
|
| 62 |
},
|
| 63 |
"revision": null,
|
| 64 |
"use_imagenet_stats": true,
|
| 65 |
-
"video_backend": "torchcodec"
|
| 66 |
-
"streaming": false
|
| 67 |
},
|
| 68 |
"env": null,
|
| 69 |
"policy": {
|
| 70 |
"type": "act",
|
| 71 |
"n_obs_steps": 1,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 72 |
"input_features": {
|
| 73 |
"observation.state": {
|
| 74 |
"type": "STATE",
|
|
@@ -116,14 +120,8 @@
|
|
| 116 |
"private": null,
|
| 117 |
"tags": null,
|
| 118 |
"license": null,
|
| 119 |
-
"pretrained_path": null,
|
| 120 |
"chunk_size": 100,
|
| 121 |
"n_action_steps": 100,
|
| 122 |
-
"normalization_mapping": {
|
| 123 |
-
"VISUAL": "MEAN_STD",
|
| 124 |
-
"STATE": "MEAN_STD",
|
| 125 |
-
"ACTION": "MEAN_STD"
|
| 126 |
-
},
|
| 127 |
"vision_backbone": "resnet18",
|
| 128 |
"pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
|
| 129 |
"replace_final_stride_with_dilation": false,
|
|
@@ -154,7 +152,7 @@
|
|
| 154 |
"eval_freq": 20000,
|
| 155 |
"log_freq": 50,
|
| 156 |
"save_checkpoint": true,
|
| 157 |
-
"save_freq":
|
| 158 |
"use_policy_training_preset": true,
|
| 159 |
"optimizer": {
|
| 160 |
"type": "adamw",
|
|
|
|
| 62 |
},
|
| 63 |
"revision": null,
|
| 64 |
"use_imagenet_stats": true,
|
| 65 |
+
"video_backend": "torchcodec"
|
|
|
|
| 66 |
},
|
| 67 |
"env": null,
|
| 68 |
"policy": {
|
| 69 |
"type": "act",
|
| 70 |
"n_obs_steps": 1,
|
| 71 |
+
"normalization_mapping": {
|
| 72 |
+
"VISUAL": "MEAN_STD",
|
| 73 |
+
"STATE": "MEAN_STD",
|
| 74 |
+
"ACTION": "MEAN_STD"
|
| 75 |
+
},
|
| 76 |
"input_features": {
|
| 77 |
"observation.state": {
|
| 78 |
"type": "STATE",
|
|
|
|
| 120 |
"private": null,
|
| 121 |
"tags": null,
|
| 122 |
"license": null,
|
|
|
|
| 123 |
"chunk_size": 100,
|
| 124 |
"n_action_steps": 100,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 125 |
"vision_backbone": "resnet18",
|
| 126 |
"pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
|
| 127 |
"replace_final_stride_with_dilation": false,
|
|
|
|
| 152 |
"eval_freq": 20000,
|
| 153 |
"log_freq": 50,
|
| 154 |
"save_checkpoint": true,
|
| 155 |
+
"save_freq": 2000,
|
| 156 |
"use_policy_training_preset": true,
|
| 157 |
"optimizer": {
|
| 158 |
"type": "adamw",
|