b10902046 commited on
Commit
e942228
·
verified ·
1 Parent(s): 7ee960f

Upload policy weights, train config and readme

Browse files
Files changed (4) hide show
  1. README.md +5 -5
  2. config.json +19 -43
  3. model.safetensors +2 -2
  4. train_config.json +30 -50
README.md CHANGED
@@ -1,21 +1,21 @@
1
  ---
2
- datasets: b10902046/il_gym
3
  library_name: lerobot
4
  license: apache-2.0
5
- model_name: act
6
  pipeline_tag: robotics
7
  tags:
8
- - act
9
  - robotics
10
  - lerobot
 
11
  ---
12
 
13
- # Model Card for act
14
 
15
  <!-- Provide a quick summary of what the model is/does. -->
16
 
17
 
18
- [Action Chunking with Transformers (ACT)](https://huggingface.co/papers/2304.13705) is an imitation-learning method that predicts short action chunks instead of single steps. It learns from teleoperated data and often achieves high success rates.
19
 
20
 
21
  This policy has been trained and pushed to the Hub using [LeRobot](https://github.com/huggingface/lerobot).
 
1
  ---
2
+ datasets: b10902046/so101_test
3
  library_name: lerobot
4
  license: apache-2.0
5
+ model_name: behavior_cloning
6
  pipeline_tag: robotics
7
  tags:
 
8
  - robotics
9
  - lerobot
10
+ - behavior_cloning
11
  ---
12
 
13
+ # Model Card for behavior_cloning
14
 
15
  <!-- Provide a quick summary of what the model is/does. -->
16
 
17
 
18
+ _Model type not recognized please update this template._
19
 
20
 
21
  This policy has been trained and pushed to the Hub using [LeRobot](https://github.com/huggingface/lerobot).
config.json CHANGED
@@ -1,32 +1,15 @@
1
  {
2
- "type": "act",
3
  "n_obs_steps": 1,
4
  "normalization_mapping": {
5
- "VISUAL": "MEAN_STD",
6
- "STATE": "MEAN_STD",
7
- "ACTION": "MEAN_STD"
8
  },
9
  "input_features": {
10
  "observation.state": {
11
  "type": "STATE",
12
  "shape": [
13
- 18
14
- ]
15
- },
16
- "observation.images.front": {
17
- "type": "VISUAL",
18
- "shape": [
19
- 3,
20
- 128,
21
- 128
22
- ]
23
- },
24
- "observation.images.wrist": {
25
- "type": "VISUAL",
26
- "shape": [
27
- 3,
28
- 128,
29
- 128
30
  ]
31
  }
32
  },
@@ -34,7 +17,7 @@
34
  "action": {
35
  "type": "ACTION",
36
  "shape": [
37
- 4
38
  ]
39
  }
40
  },
@@ -45,25 +28,18 @@
45
  "private": null,
46
  "tags": null,
47
  "license": null,
48
- "chunk_size": 100,
49
- "n_action_steps": 100,
50
- "vision_backbone": "resnet18",
51
- "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
52
- "replace_final_stride_with_dilation": false,
53
- "pre_norm": false,
54
- "dim_model": 512,
55
- "n_heads": 8,
56
- "dim_feedforward": 3200,
57
- "feedforward_activation": "relu",
58
- "n_encoder_layers": 4,
59
- "n_decoder_layers": 1,
60
- "use_vae": true,
61
- "latent_dim": 32,
62
- "n_vae_encoder_layers": 4,
63
- "temporal_ensemble_coeff": null,
64
- "dropout": 0.1,
65
- "kl_weight": 10.0,
66
- "optimizer_lr": 1e-05,
67
- "optimizer_weight_decay": 0.0001,
68
- "optimizer_lr_backbone": 1e-05
69
  }
 
1
  {
2
+ "type": "behavior_cloning",
3
  "n_obs_steps": 1,
4
  "normalization_mapping": {
5
+ "STATE": "MIN_MAX",
6
+ "ACTION": "MIN_MAX"
 
7
  },
8
  "input_features": {
9
  "observation.state": {
10
  "type": "STATE",
11
  "shape": [
12
+ 6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ]
14
  }
15
  },
 
17
  "action": {
18
  "type": "ACTION",
19
  "shape": [
20
+ 6
21
  ]
22
  }
23
  },
 
28
  "private": null,
29
  "tags": null,
30
  "license": null,
31
+ "hidden_dims": [
32
+ 256,
33
+ 256
34
+ ],
35
+ "activation": "relu",
36
+ "optimizer_lr": 0.0001,
37
+ "optimizer_betas": [
38
+ 0.9,
39
+ 0.999
40
+ ],
41
+ "optimizer_eps": 1e-08,
42
+ "optimizer_weight_decay": 1e-06,
43
+ "scheduler_name": "cosine",
44
+ "scheduler_warmup_steps": 500
 
 
 
 
 
 
 
45
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7642c24bdf86fe5b40d246d0af0ff7720c5a1bf0aea9b2bc55174c66ecc07eb
3
- size 206742080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67bca0eb7179b50fcfaefdb050bea6feeb46711940e2b6a42e26d6b7a2c62150
3
+ size 277736
train_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "dataset": {
3
- "repo_id": "b10902046/il_gym",
4
  "root": null,
5
  "episodes": null,
6
  "image_transforms": {
@@ -66,34 +66,17 @@
66
  },
67
  "env": null,
68
  "policy": {
69
- "type": "act",
70
  "n_obs_steps": 1,
71
  "normalization_mapping": {
72
- "VISUAL": "MEAN_STD",
73
- "STATE": "MEAN_STD",
74
- "ACTION": "MEAN_STD"
75
  },
76
  "input_features": {
77
  "observation.state": {
78
  "type": "STATE",
79
  "shape": [
80
- 18
81
- ]
82
- },
83
- "observation.images.front": {
84
- "type": "VISUAL",
85
- "shape": [
86
- 3,
87
- 128,
88
- 128
89
- ]
90
- },
91
- "observation.images.wrist": {
92
- "type": "VISUAL",
93
- "shape": [
94
- 3,
95
- 128,
96
- 128
97
  ]
98
  }
99
  },
@@ -101,7 +84,7 @@
101
  "action": {
102
  "type": "ACTION",
103
  "shape": [
104
- 4
105
  ]
106
  }
107
  },
@@ -112,30 +95,23 @@
112
  "private": null,
113
  "tags": null,
114
  "license": null,
115
- "chunk_size": 100,
116
- "n_action_steps": 100,
117
- "vision_backbone": "resnet18",
118
- "pretrained_backbone_weights": "ResNet18_Weights.IMAGENET1K_V1",
119
- "replace_final_stride_with_dilation": false,
120
- "pre_norm": false,
121
- "dim_model": 512,
122
- "n_heads": 8,
123
- "dim_feedforward": 3200,
124
- "feedforward_activation": "relu",
125
- "n_encoder_layers": 4,
126
- "n_decoder_layers": 1,
127
- "use_vae": true,
128
- "latent_dim": 32,
129
- "n_vae_encoder_layers": 4,
130
- "temporal_ensemble_coeff": null,
131
- "dropout": 0.1,
132
- "kl_weight": 10.0,
133
- "optimizer_lr": 1e-05,
134
- "optimizer_weight_decay": 0.0001,
135
- "optimizer_lr_backbone": 1e-05
136
  },
137
- "output_dir": "outputs/train/il_sim_test",
138
- "job_name": "il_sim_test",
139
  "resume": false,
140
  "seed": 1000,
141
  "num_workers": 4,
@@ -147,9 +123,9 @@
147
  "save_freq": 20000,
148
  "use_policy_training_preset": true,
149
  "optimizer": {
150
- "type": "adamw",
151
- "lr": 1e-05,
152
- "weight_decay": 0.0001,
153
  "grad_clip_norm": 10.0,
154
  "betas": [
155
  0.9,
@@ -157,7 +133,11 @@
157
  ],
158
  "eps": 1e-08
159
  },
160
- "scheduler": null,
 
 
 
 
161
  "eval": {
162
  "n_episodes": 50,
163
  "batch_size": 50,
 
1
  {
2
  "dataset": {
3
+ "repo_id": "b10902046/so101_test",
4
  "root": null,
5
  "episodes": null,
6
  "image_transforms": {
 
66
  },
67
  "env": null,
68
  "policy": {
69
+ "type": "behavior_cloning",
70
  "n_obs_steps": 1,
71
  "normalization_mapping": {
72
+ "STATE": "MIN_MAX",
73
+ "ACTION": "MIN_MAX"
 
74
  },
75
  "input_features": {
76
  "observation.state": {
77
  "type": "STATE",
78
  "shape": [
79
+ 6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
  ]
81
  }
82
  },
 
84
  "action": {
85
  "type": "ACTION",
86
  "shape": [
87
+ 6
88
  ]
89
  }
90
  },
 
95
  "private": null,
96
  "tags": null,
97
  "license": null,
98
+ "hidden_dims": [
99
+ 256,
100
+ 256
101
+ ],
102
+ "activation": "relu",
103
+ "optimizer_lr": 0.0001,
104
+ "optimizer_betas": [
105
+ 0.9,
106
+ 0.999
107
+ ],
108
+ "optimizer_eps": 1e-08,
109
+ "optimizer_weight_decay": 1e-06,
110
+ "scheduler_name": "cosine",
111
+ "scheduler_warmup_steps": 500
 
 
 
 
 
 
 
112
  },
113
+ "output_dir": "outputs/train/act_so101_test",
114
+ "job_name": "act_so101_test",
115
  "resume": false,
116
  "seed": 1000,
117
  "num_workers": 4,
 
123
  "save_freq": 20000,
124
  "use_policy_training_preset": true,
125
  "optimizer": {
126
+ "type": "adam",
127
+ "lr": 0.0001,
128
+ "weight_decay": 1e-06,
129
  "grad_clip_norm": 10.0,
130
  "betas": [
131
  0.9,
 
133
  ],
134
  "eps": 1e-08
135
  },
136
+ "scheduler": {
137
+ "type": "diffuser",
138
+ "num_warmup_steps": 500,
139
+ "name": "cosine"
140
+ },
141
  "eval": {
142
  "n_episodes": 50,
143
  "batch_size": 50,