Upload policy weights, train config and readme
Browse files- README.md +2 -2
- config.json +12 -4
- model.safetensors +2 -2
- train_config.json +16 -8
README.md
CHANGED
|
@@ -1,13 +1,13 @@
|
|
| 1 |
---
|
| 2 |
-
datasets: longzhiying/
|
| 3 |
library_name: lerobot
|
| 4 |
license: apache-2.0
|
| 5 |
model_name: reward_classifier
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
| 8 |
-
- reward_classifier
|
| 9 |
- lerobot
|
| 10 |
- robotics
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for reward_classifier
|
|
|
|
| 1 |
---
|
| 2 |
+
datasets: longzhiying/rokae_assembly_reward
|
| 3 |
library_name: lerobot
|
| 4 |
license: apache-2.0
|
| 5 |
model_name: reward_classifier
|
| 6 |
pipeline_tag: robotics
|
| 7 |
tags:
|
|
|
|
| 8 |
- lerobot
|
| 9 |
- robotics
|
| 10 |
+
- reward_classifier
|
| 11 |
---
|
| 12 |
|
| 13 |
# Model Card for reward_classifier
|
config.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"type": "reward_classifier",
|
| 3 |
"n_obs_steps": 1,
|
| 4 |
"input_features": {
|
| 5 |
-
"observation.images.
|
| 6 |
"type": "VISUAL",
|
| 7 |
"shape": [
|
| 8 |
3,
|
|
@@ -10,7 +10,15 @@
|
|
| 10 |
128
|
| 11 |
]
|
| 12 |
},
|
| 13 |
-
"observation.images.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
"type": "VISUAL",
|
| 15 |
"shape": [
|
| 16 |
3,
|
|
@@ -23,7 +31,7 @@
|
|
| 23 |
"action": {
|
| 24 |
"type": "ACTION",
|
| 25 |
"shape": [
|
| 26 |
-
|
| 27 |
]
|
| 28 |
}
|
| 29 |
},
|
|
@@ -43,7 +51,7 @@
|
|
| 43 |
"dropout_rate": 0.1,
|
| 44 |
"model_name": "helper2424/resnet10",
|
| 45 |
"model_type": "cnn",
|
| 46 |
-
"num_cameras":
|
| 47 |
"learning_rate": 0.0001,
|
| 48 |
"weight_decay": 0.01,
|
| 49 |
"grad_clip_norm": 1.0,
|
|
|
|
| 2 |
"type": "reward_classifier",
|
| 3 |
"n_obs_steps": 1,
|
| 4 |
"input_features": {
|
| 5 |
+
"observation.images.side": {
|
| 6 |
"type": "VISUAL",
|
| 7 |
"shape": [
|
| 8 |
3,
|
|
|
|
| 10 |
128
|
| 11 |
]
|
| 12 |
},
|
| 13 |
+
"observation.images.wrist1": {
|
| 14 |
+
"type": "VISUAL",
|
| 15 |
+
"shape": [
|
| 16 |
+
3,
|
| 17 |
+
128,
|
| 18 |
+
128
|
| 19 |
+
]
|
| 20 |
+
},
|
| 21 |
+
"observation.images.wrist2": {
|
| 22 |
"type": "VISUAL",
|
| 23 |
"shape": [
|
| 24 |
3,
|
|
|
|
| 31 |
"action": {
|
| 32 |
"type": "ACTION",
|
| 33 |
"shape": [
|
| 34 |
+
3
|
| 35 |
]
|
| 36 |
}
|
| 37 |
},
|
|
|
|
| 51 |
"dropout_rate": 0.1,
|
| 52 |
"model_name": "helper2424/resnet10",
|
| 53 |
"model_type": "cnn",
|
| 54 |
+
"num_cameras": 3,
|
| 55 |
"learning_rate": 0.0001,
|
| 56 |
"weight_decay": 0.01,
|
| 57 |
"grad_clip_norm": 1.0,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f05c42db8b55bb76f19be5cbf3ba456ecb9a73d3d1fab249753a230503bf1eb7
|
| 3 |
+
size 33810260
|
train_config.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"dataset": {
|
| 3 |
-
"repo_id": "longzhiying/
|
| 4 |
"root": null,
|
| 5 |
"episodes": null,
|
| 6 |
"image_transforms": {
|
|
@@ -84,7 +84,7 @@
|
|
| 84 |
"type": "reward_classifier",
|
| 85 |
"n_obs_steps": 1,
|
| 86 |
"input_features": {
|
| 87 |
-
"observation.images.
|
| 88 |
"type": "VISUAL",
|
| 89 |
"shape": [
|
| 90 |
3,
|
|
@@ -92,7 +92,15 @@
|
|
| 92 |
128
|
| 93 |
]
|
| 94 |
},
|
| 95 |
-
"observation.images.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 96 |
"type": "VISUAL",
|
| 97 |
"shape": [
|
| 98 |
3,
|
|
@@ -105,7 +113,7 @@
|
|
| 105 |
"action": {
|
| 106 |
"type": "ACTION",
|
| 107 |
"shape": [
|
| 108 |
-
|
| 109 |
]
|
| 110 |
}
|
| 111 |
},
|
|
@@ -125,7 +133,7 @@
|
|
| 125 |
"dropout_rate": 0.1,
|
| 126 |
"model_name": "helper2424/resnet10",
|
| 127 |
"model_type": "cnn",
|
| 128 |
-
"num_cameras":
|
| 129 |
"learning_rate": 0.0001,
|
| 130 |
"weight_decay": 0.01,
|
| 131 |
"grad_clip_norm": 1.0,
|
|
@@ -133,12 +141,12 @@
|
|
| 133 |
"VISUAL": "MEAN_STD"
|
| 134 |
}
|
| 135 |
},
|
| 136 |
-
"output_dir": "outputs/train/
|
| 137 |
"job_name": "reward-classifier",
|
| 138 |
"resume": false,
|
| 139 |
"seed": 2,
|
| 140 |
"num_workers": 4,
|
| 141 |
-
"batch_size":
|
| 142 |
"steps": 5000,
|
| 143 |
"eval_freq": 1000,
|
| 144 |
"log_freq": 10,
|
|
@@ -168,7 +176,7 @@
|
|
| 168 |
"project": "reward-classifier",
|
| 169 |
"entity": null,
|
| 170 |
"notes": null,
|
| 171 |
-
"run_id": "
|
| 172 |
"mode": null
|
| 173 |
},
|
| 174 |
"checkpoint_path": null,
|
|
|
|
| 1 |
{
|
| 2 |
"dataset": {
|
| 3 |
+
"repo_id": "longzhiying/rokae_assembly_reward",
|
| 4 |
"root": null,
|
| 5 |
"episodes": null,
|
| 6 |
"image_transforms": {
|
|
|
|
| 84 |
"type": "reward_classifier",
|
| 85 |
"n_obs_steps": 1,
|
| 86 |
"input_features": {
|
| 87 |
+
"observation.images.side": {
|
| 88 |
"type": "VISUAL",
|
| 89 |
"shape": [
|
| 90 |
3,
|
|
|
|
| 92 |
128
|
| 93 |
]
|
| 94 |
},
|
| 95 |
+
"observation.images.wrist1": {
|
| 96 |
+
"type": "VISUAL",
|
| 97 |
+
"shape": [
|
| 98 |
+
3,
|
| 99 |
+
128,
|
| 100 |
+
128
|
| 101 |
+
]
|
| 102 |
+
},
|
| 103 |
+
"observation.images.wrist2": {
|
| 104 |
"type": "VISUAL",
|
| 105 |
"shape": [
|
| 106 |
3,
|
|
|
|
| 113 |
"action": {
|
| 114 |
"type": "ACTION",
|
| 115 |
"shape": [
|
| 116 |
+
3
|
| 117 |
]
|
| 118 |
}
|
| 119 |
},
|
|
|
|
| 133 |
"dropout_rate": 0.1,
|
| 134 |
"model_name": "helper2424/resnet10",
|
| 135 |
"model_type": "cnn",
|
| 136 |
+
"num_cameras": 3,
|
| 137 |
"learning_rate": 0.0001,
|
| 138 |
"weight_decay": 0.01,
|
| 139 |
"grad_clip_norm": 1.0,
|
|
|
|
| 141 |
"VISUAL": "MEAN_STD"
|
| 142 |
}
|
| 143 |
},
|
| 144 |
+
"output_dir": "outputs/train/2026-02-03/14-35-22_reward-classifier",
|
| 145 |
"job_name": "reward-classifier",
|
| 146 |
"resume": false,
|
| 147 |
"seed": 2,
|
| 148 |
"num_workers": 4,
|
| 149 |
+
"batch_size": 64,
|
| 150 |
"steps": 5000,
|
| 151 |
"eval_freq": 1000,
|
| 152 |
"log_freq": 10,
|
|
|
|
| 176 |
"project": "reward-classifier",
|
| 177 |
"entity": null,
|
| 178 |
"notes": null,
|
| 179 |
+
"run_id": "apcbu561",
|
| 180 |
"mode": null
|
| 181 |
},
|
| 182 |
"checkpoint_path": null,
|