chuanmew commited on
Commit
38cab28
·
verified ·
1 Parent(s): 181a9cc

test-1-checkpoint-1500

Browse files
test-1-checkpoint-1500/config.json ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_horizon": 50,
3
+ "batch_size": 8,
4
+ "buffer_size": 64,
5
+ "eval_mix": "qiwu_test",
6
+ "frozen_keys": [
7
+ "*hf_model*"
8
+ ],
9
+ "lora": false,
10
+ "lora_config": {
11
+ "bias": "none",
12
+ "lora_alpha": 16,
13
+ "lora_dropout": 0.05,
14
+ "r": 8
15
+ },
16
+ "model": {
17
+ "heads": {
18
+ "action": {
19
+ "args": [],
20
+ "kwargs": {
21
+ "action_dim": 7,
22
+ "action_horizon": 50,
23
+ "readout_key": "readout_action",
24
+ "token_embedding_size": 768
25
+ },
26
+ "module": "model.octo.components.action_heads",
27
+ "name": "L1ActionHead"
28
+ }
29
+ },
30
+ "max_horizon": 10,
31
+ "observation_tokenizers": {
32
+ "primary": {
33
+ "args": [],
34
+ "kwargs": {
35
+ "encoder": {
36
+ "args": [],
37
+ "kwargs": {
38
+ "in_features": 6
39
+ },
40
+ "module": "octo.model.components.vit_encoders",
41
+ "name": "SmallStem16"
42
+ },
43
+ "obs_stack_keys": [
44
+ "image_primary"
45
+ ],
46
+ "task_stack_keys": [
47
+ "image_primary"
48
+ ]
49
+ },
50
+ "module": "octo.model.components.tokenizers",
51
+ "name": "ImageTokenizer"
52
+ },
53
+ "state": {
54
+ "args": [],
55
+ "kwargs": {
56
+ "bin_type": "normal",
57
+ "high": 2.0,
58
+ "low": -2.0,
59
+ "n_bins": 256,
60
+ "obs_keys": [
61
+ "state"
62
+ ]
63
+ },
64
+ "module": "model.octo.components.tokenizers",
65
+ "name": "LowdimObsTokenizer"
66
+ }
67
+ },
68
+ "readouts": {
69
+ "action": 1
70
+ },
71
+ "repeat_task_tokens": true,
72
+ "task_tokenizers": {
73
+ "language": {
74
+ "args": [],
75
+ "kwargs": {
76
+ "encoder": "t5-base",
77
+ "finetune_encoder": false
78
+ },
79
+ "module": "octo.model.components.tokenizers",
80
+ "name": "LanguageTokenizer"
81
+ }
82
+ },
83
+ "token_embedding_size": 768,
84
+ "transformer_kwargs": {
85
+ "add_position_embedding": false,
86
+ "attention_dropout_rate": 0.0,
87
+ "dropout_rate": 0.0,
88
+ "mlp_dim": 3072,
89
+ "num_attention_heads": 12,
90
+ "num_layers": 12
91
+ },
92
+ "use_correct_attention": true
93
+ },
94
+ "no_state": false,
95
+ "obs_token_nums": {
96
+ "primary": 256,
97
+ "state": 7
98
+ },
99
+ "output_dir": "qiwu_checkpoints4",
100
+ "pretrained": {
101
+ "kwargs": {
102
+ "subpath": "oxe-g2-checkpoint-300000"
103
+ },
104
+ "model": "hf://chuanmew/octo_torch"
105
+ },
106
+ "run_name": "qiwu",
107
+ "sampler_num_samples": 2560000,
108
+ "seed": 42,
109
+ "subsample_length": 99999,
110
+ "text_processor": {
111
+ "args": [],
112
+ "kwargs": {
113
+ "encode_with_model": false,
114
+ "tokenizer_kwargs": {
115
+ "max_length": 16,
116
+ "padding": "max_length",
117
+ "return_tensors": "np",
118
+ "truncation": true
119
+ },
120
+ "tokenizer_name": "t5-base"
121
+ },
122
+ "module": "octo.components.text_processing",
123
+ "name": "HFTokenizer"
124
+ },
125
+ "train_mix": "qiwu_train",
126
+ "training_arguments": {
127
+ "bf16": true,
128
+ "dataloader_pin_memory": true,
129
+ "ddp_find_unused_parameters": true,
130
+ "eval_steps": 100,
131
+ "gradient_accumulation_steps": 1,
132
+ "learning_rate": 3e-05,
133
+ "logging_nan_inf_filter": false,
134
+ "logging_steps": 100,
135
+ "lr_scheduler_kwargs": {
136
+ "decay_type": "cosine",
137
+ "min_lr_ratio": 0.05,
138
+ "num_decay_steps": 0
139
+ },
140
+ "lr_scheduler_type": "warmup_stable_decay",
141
+ "max_grad_norm": 1.0,
142
+ "max_steps": 300000,
143
+ "optim": "adamw_torch_fused",
144
+ "per_device_eval_batch_size": 1,
145
+ "per_device_train_batch_size": 1,
146
+ "report_to": "wandb",
147
+ "save_safetensors": true,
148
+ "save_steps": 500,
149
+ "torch_compile": true,
150
+ "warmup_steps": 100,
151
+ "weight_decay": 0.1
152
+ },
153
+ "training_keys": [
154
+ "*action*",
155
+ "*observation_tokenizers*",
156
+ "*transformer.transformer*",
157
+ "*task_projections*",
158
+ "*obs_projections*",
159
+ "*readout_embeddings*",
160
+ "*task_pos_embeddings*",
161
+ "*obs_pos_embeddings*"
162
+ ]
163
+ }
test-1-checkpoint-1500/dataset_statistics.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "action": {
4
+ "mean": [
5
+ 0.005713334307074547,
6
+ -8.683370651851874e-06,
7
+ -0.0027405666187405586,
8
+ -0.01428219024091959,
9
+ -0.00569964898750186,
10
+ 0.03178633004426956,
11
+ 0.31847861409187317
12
+ ],
13
+ "std": [
14
+ 0.017133653163909912,
15
+ 0.012163525447249413,
16
+ 0.028012460097670555,
17
+ 0.056618135422468185,
18
+ 0.059768617153167725,
19
+ 0.08718060702085495,
20
+ 0.45297929644584656
21
+ ],
22
+ "min": [
23
+ -0.045306507498025894,
24
+ -0.058269500732421875,
25
+ -0.0945921391248703,
26
+ -0.39226841926574707,
27
+ -0.31778502464294434,
28
+ -0.46861958503723145,
29
+ 0.0
30
+ ],
31
+ "max": [
32
+ 0.10087767243385315,
33
+ 0.06844352185726166,
34
+ 0.13764026761054993,
35
+ 0.39099645614624023,
36
+ 0.4911177158355713,
37
+ 0.7427346706390381,
38
+ 1.0
39
+ ],
40
+ "mask": [
41
+ true,
42
+ true,
43
+ true,
44
+ true,
45
+ true,
46
+ true,
47
+ false
48
+ ]
49
+ },
50
+ "state": {
51
+ "mean": [
52
+ -1.2245632410049438,
53
+ -1.050716519355774,
54
+ -1.258064866065979,
55
+ -0.5181024670600891,
56
+ -0.9791191816329956,
57
+ 0.11057784408330917,
58
+ 0.26232674717903137
59
+ ],
60
+ "std": [
61
+ 0.23113861680030823,
62
+ 0.32922741770744324,
63
+ 0.2693513333797455,
64
+ 0.32120275497436523,
65
+ 0.2501857280731201,
66
+ 0.4673791527748108,
67
+ 0.42628398537635803
68
+ ],
69
+ "min": [
70
+ -1.8634008169174194,
71
+ -1.9834270477294922,
72
+ -1.991769790649414,
73
+ -1.7809165716171265,
74
+ -1.5987040996551514,
75
+ -1.2469679117202759,
76
+ 0.0
77
+ ],
78
+ "max": [
79
+ -0.6335545182228088,
80
+ -0.20619319379329681,
81
+ -0.40194931626319885,
82
+ 0.5385388135910034,
83
+ 0.03621558099985123,
84
+ 1.5524179935455322,
85
+ 1.0
86
+ ],
87
+ "mask": [
88
+ true,
89
+ true,
90
+ true,
91
+ true,
92
+ true,
93
+ true,
94
+ false
95
+ ]
96
+ }
97
+ }
98
+ ]
test-1-checkpoint-1500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:226a8bda34b980c3c87f83da6748f0ae0fa8cb5bcd2627d30b9179bbfd8a32a7
3
+ size 412295788
test-1-checkpoint-1500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e128857b241bc01a1db11900bc9169b249cff6c24549645d4251a94614e59226
3
+ size 386143930
test-1-checkpoint-1500/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:916ccac3d8b32b81123571dde2d9dbe2f1f07ebd90ef54a304aaf7c937ef1a95
3
+ size 15984
test-1-checkpoint-1500/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:333aef0819b926923bec9f240bca2466cd68a800ab2e9ab4c6e2335f27143f83
3
+ size 15984
test-1-checkpoint-1500/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81f770f6f8cab2724d297e3f5e935dc10562e83c9b4a2bafe5cacde08d67d220
3
+ size 15984
test-1-checkpoint-1500/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b2a15f877a0d8e205b82d2ee3a0daaa5b69a41a8787ae96d50447e7fd8616a4
3
+ size 15984
test-1-checkpoint-1500/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50dce3f47a9c4d3baf720ac97339bde0244e272c039c1780d2fce53bdc309080
3
+ size 15984
test-1-checkpoint-1500/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d92fd2e6d58ec26cc909989125659fa97e9aa17e622cefad055005be36d6b81a
3
+ size 15984
test-1-checkpoint-1500/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c306d4a4f506e6752ae90755d2e4fcb4604eea922888c2e19758160d22b62474
3
+ size 15984
test-1-checkpoint-1500/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab9712425612fc7ead06367303ef37eacf7dcc969665997a0cac0f28a009a04a
3
+ size 15984
test-1-checkpoint-1500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed4336dbf0ba1587dcd2e69dc77b38f8ff64571073dd364cce7a4f08ec4c85e5
3
+ size 1064
test-1-checkpoint-1500/trainer_state.json ADDED
@@ -0,0 +1,259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.0375,
6
+ "eval_steps": 100,
7
+ "global_step": 1500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0025,
14
+ "grad_norm": 4.0,
15
+ "learning_rate": 2.9715000000000003e-05,
16
+ "loss": 4.4555,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.0025,
21
+ "eval_loss": 4.083842754364014,
22
+ "eval_runtime": 15.1972,
23
+ "eval_samples_per_second": 1330.307,
24
+ "eval_steps_per_second": 166.346,
25
+ "step": 100
26
+ },
27
+ {
28
+ "epoch": 0.005,
29
+ "grad_norm": 6.375,
30
+ "learning_rate": 3e-05,
31
+ "loss": 3.7452,
32
+ "step": 200
33
+ },
34
+ {
35
+ "epoch": 0.005,
36
+ "eval_loss": 3.76338529586792,
37
+ "eval_runtime": 14.6937,
38
+ "eval_samples_per_second": 1375.896,
39
+ "eval_steps_per_second": 172.047,
40
+ "step": 200
41
+ },
42
+ {
43
+ "epoch": 0.0075,
44
+ "grad_norm": 10.875,
45
+ "learning_rate": 3e-05,
46
+ "loss": 3.5303,
47
+ "step": 300
48
+ },
49
+ {
50
+ "epoch": 0.0075,
51
+ "eval_loss": 3.6259424686431885,
52
+ "eval_runtime": 14.8394,
53
+ "eval_samples_per_second": 1362.388,
54
+ "eval_steps_per_second": 170.357,
55
+ "step": 300
56
+ },
57
+ {
58
+ "epoch": 0.01,
59
+ "grad_norm": 5.3125,
60
+ "learning_rate": 3e-05,
61
+ "loss": 3.3359,
62
+ "step": 400
63
+ },
64
+ {
65
+ "epoch": 0.01,
66
+ "eval_loss": 3.4924354553222656,
67
+ "eval_runtime": 14.869,
68
+ "eval_samples_per_second": 1359.679,
69
+ "eval_steps_per_second": 170.019,
70
+ "step": 400
71
+ },
72
+ {
73
+ "epoch": 0.0125,
74
+ "grad_norm": 7.5625,
75
+ "learning_rate": 3e-05,
76
+ "loss": 3.2483,
77
+ "step": 500
78
+ },
79
+ {
80
+ "epoch": 0.0125,
81
+ "eval_loss": 3.427826404571533,
82
+ "eval_runtime": 14.83,
83
+ "eval_samples_per_second": 1363.252,
84
+ "eval_steps_per_second": 170.466,
85
+ "step": 500
86
+ },
87
+ {
88
+ "epoch": 0.015,
89
+ "grad_norm": 6.75,
90
+ "learning_rate": 3e-05,
91
+ "loss": 3.174,
92
+ "step": 600
93
+ },
94
+ {
95
+ "epoch": 0.015,
96
+ "eval_loss": 3.4111289978027344,
97
+ "eval_runtime": 14.9413,
98
+ "eval_samples_per_second": 1353.099,
99
+ "eval_steps_per_second": 169.196,
100
+ "step": 600
101
+ },
102
+ {
103
+ "epoch": 0.0175,
104
+ "grad_norm": 9.125,
105
+ "learning_rate": 3e-05,
106
+ "loss": 3.0925,
107
+ "step": 700
108
+ },
109
+ {
110
+ "epoch": 0.0175,
111
+ "eval_loss": 3.364833354949951,
112
+ "eval_runtime": 14.9006,
113
+ "eval_samples_per_second": 1356.79,
114
+ "eval_steps_per_second": 169.657,
115
+ "step": 700
116
+ },
117
+ {
118
+ "epoch": 0.02,
119
+ "grad_norm": 6.4375,
120
+ "learning_rate": 3e-05,
121
+ "loss": 3.0564,
122
+ "step": 800
123
+ },
124
+ {
125
+ "epoch": 0.02,
126
+ "eval_loss": 3.3411409854888916,
127
+ "eval_runtime": 14.9667,
128
+ "eval_samples_per_second": 1350.796,
129
+ "eval_steps_per_second": 168.908,
130
+ "step": 800
131
+ },
132
+ {
133
+ "epoch": 0.0225,
134
+ "grad_norm": 6.34375,
135
+ "learning_rate": 3e-05,
136
+ "loss": 2.9535,
137
+ "step": 900
138
+ },
139
+ {
140
+ "epoch": 0.0225,
141
+ "eval_loss": 3.31355881690979,
142
+ "eval_runtime": 14.9497,
143
+ "eval_samples_per_second": 1352.334,
144
+ "eval_steps_per_second": 169.1,
145
+ "step": 900
146
+ },
147
+ {
148
+ "epoch": 0.025,
149
+ "grad_norm": 6.375,
150
+ "learning_rate": 3e-05,
151
+ "loss": 2.9478,
152
+ "step": 1000
153
+ },
154
+ {
155
+ "epoch": 0.025,
156
+ "eval_loss": 3.3087844848632812,
157
+ "eval_runtime": 14.8619,
158
+ "eval_samples_per_second": 1360.323,
159
+ "eval_steps_per_second": 170.099,
160
+ "step": 1000
161
+ },
162
+ {
163
+ "epoch": 0.0275,
164
+ "grad_norm": 13.0,
165
+ "learning_rate": 3e-05,
166
+ "loss": 2.876,
167
+ "step": 1100
168
+ },
169
+ {
170
+ "epoch": 0.0275,
171
+ "eval_loss": 3.295400381088257,
172
+ "eval_runtime": 15.1916,
173
+ "eval_samples_per_second": 1330.805,
174
+ "eval_steps_per_second": 166.408,
175
+ "step": 1100
176
+ },
177
+ {
178
+ "epoch": 0.03,
179
+ "grad_norm": 6.75,
180
+ "learning_rate": 3e-05,
181
+ "loss": 2.8775,
182
+ "step": 1200
183
+ },
184
+ {
185
+ "epoch": 0.03,
186
+ "eval_loss": 3.292489767074585,
187
+ "eval_runtime": 14.7952,
188
+ "eval_samples_per_second": 1366.461,
189
+ "eval_steps_per_second": 170.867,
190
+ "step": 1200
191
+ },
192
+ {
193
+ "epoch": 0.0325,
194
+ "grad_norm": 14.25,
195
+ "learning_rate": 3e-05,
196
+ "loss": 2.8152,
197
+ "step": 1300
198
+ },
199
+ {
200
+ "epoch": 0.0325,
201
+ "eval_loss": 3.2735142707824707,
202
+ "eval_runtime": 14.9427,
203
+ "eval_samples_per_second": 1352.969,
204
+ "eval_steps_per_second": 169.18,
205
+ "step": 1300
206
+ },
207
+ {
208
+ "epoch": 0.035,
209
+ "grad_norm": 8.25,
210
+ "learning_rate": 3e-05,
211
+ "loss": 2.7434,
212
+ "step": 1400
213
+ },
214
+ {
215
+ "epoch": 0.035,
216
+ "eval_loss": 3.2721683979034424,
217
+ "eval_runtime": 14.9214,
218
+ "eval_samples_per_second": 1354.903,
219
+ "eval_steps_per_second": 169.422,
220
+ "step": 1400
221
+ },
222
+ {
223
+ "epoch": 0.0375,
224
+ "grad_norm": 8.625,
225
+ "learning_rate": 3e-05,
226
+ "loss": 2.7628,
227
+ "step": 1500
228
+ },
229
+ {
230
+ "epoch": 0.0375,
231
+ "eval_loss": 3.264232873916626,
232
+ "eval_runtime": 14.791,
233
+ "eval_samples_per_second": 1366.845,
234
+ "eval_steps_per_second": 170.915,
235
+ "step": 1500
236
+ }
237
+ ],
238
+ "logging_steps": 100,
239
+ "max_steps": 300000,
240
+ "num_input_tokens_seen": 0,
241
+ "num_train_epochs": 8,
242
+ "save_steps": 500,
243
+ "stateful_callbacks": {
244
+ "TrainerControl": {
245
+ "args": {
246
+ "should_epoch_stop": false,
247
+ "should_evaluate": false,
248
+ "should_log": false,
249
+ "should_save": true,
250
+ "should_training_stop": false
251
+ },
252
+ "attributes": {}
253
+ }
254
+ },
255
+ "total_flos": 0.0,
256
+ "train_batch_size": 1,
257
+ "trial_name": null,
258
+ "trial_params": null
259
+ }
test-1-checkpoint-1500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dbd37b641e7bea235d5fc9c06dc239b106449ff9247ac5d78f46fac43e0b9ad
3
+ size 5432