nnh-pbbb commited on
Commit
eecddb8
·
verified ·
1 Parent(s): c60354b

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "pi05",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 32
9
+ ]
10
+ },
11
+ "observation.state_eepose": {
12
+ "type": "STATE",
13
+ "shape": [
14
+ 14
15
+ ]
16
+ },
17
+ "observation.images.cam_high": {
18
+ "type": "VISUAL",
19
+ "shape": [
20
+ 3,
21
+ 360,
22
+ 480
23
+ ]
24
+ },
25
+ "observation.images.cam_left_wrist": {
26
+ "type": "VISUAL",
27
+ "shape": [
28
+ 3,
29
+ 360,
30
+ 480
31
+ ]
32
+ },
33
+ "observation.images.cam_right_wrist": {
34
+ "type": "VISUAL",
35
+ "shape": [
36
+ 3,
37
+ 360,
38
+ 480
39
+ ]
40
+ }
41
+ },
42
+ "output_features": {
43
+ "action": {
44
+ "type": "ACTION",
45
+ "shape": [
46
+ 32
47
+ ]
48
+ },
49
+ "action_eepose": {
50
+ "type": "ACTION",
51
+ "shape": [
52
+ 14
53
+ ]
54
+ }
55
+ },
56
+ "device": "cpu",
57
+ "use_amp": false,
58
+ "push_to_hub": false,
59
+ "repo_id": null,
60
+ "private": null,
61
+ "tags": null,
62
+ "license": null,
63
+ "pretrained_path": "checkpoints/pi05_base",
64
+ "paligemma_variant": "gemma_2b",
65
+ "action_expert_variant": "gemma_300m",
66
+ "dtype": "bfloat16",
67
+ "chunk_size": 50,
68
+ "n_action_steps": 50,
69
+ "max_state_dim": 32,
70
+ "max_action_dim": 32,
71
+ "state_cond": true,
72
+ "num_inference_steps": 10,
73
+ "time_sampling_beta_alpha": 1.5,
74
+ "time_sampling_beta_beta": 1.0,
75
+ "time_sampling_scale": 0.999,
76
+ "time_sampling_offset": 0.001,
77
+ "min_period": 0.004,
78
+ "max_period": 4.0,
79
+ "image_resolution": [
80
+ 224,
81
+ 224
82
+ ],
83
+ "empty_cameras": 0,
84
+ "tokenizer_max_length": 200,
85
+ "normalization_mapping": {
86
+ "VISUAL": "IDENTITY",
87
+ "STATE": "MEAN_STD",
88
+ "ACTION": "MEAN_STD"
89
+ },
90
+ "gradient_checkpointing": false,
91
+ "compile_model": false,
92
+ "compile_mode": "max-autotune",
93
+ "fuse_qkv": false,
94
+ "fuse_gate_up": false,
95
+ "optimizer_lr": 2.5e-05,
96
+ "optimizer_betas": [
97
+ 0.9,
98
+ 0.95
99
+ ],
100
+ "optimizer_eps": 1e-08,
101
+ "optimizer_weight_decay": 0.01,
102
+ "optimizer_grad_clip_norm": 1.0,
103
+ "scheduler_warmup_steps": 1000,
104
+ "scheduler_decay_steps": 30000,
105
+ "scheduler_decay_lr": 2.5e-06,
106
+ "vlm_config": {},
107
+ "action_expert_config": {}
108
+ }
lora_adapters/README.md ADDED
@@ -0,0 +1,203 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ tags:
4
+ - lora
5
+ ---
6
+
7
+ # Model Card for Model ID
8
+
9
+ <!-- Provide a quick summary of what the model is/does. -->
10
+
11
+
12
+
13
+ ## Model Details
14
+
15
+ ### Model Description
16
+
17
+ <!-- Provide a longer summary of what this model is. -->
18
+
19
+
20
+
21
+ - **Developed by:** [More Information Needed]
22
+ - **Funded by [optional]:** [More Information Needed]
23
+ - **Shared by [optional]:** [More Information Needed]
24
+ - **Model type:** [More Information Needed]
25
+ - **Language(s) (NLP):** [More Information Needed]
26
+ - **License:** [More Information Needed]
27
+ - **Finetuned from model [optional]:** [More Information Needed]
28
+
29
+ ### Model Sources [optional]
30
+
31
+ <!-- Provide the basic links for the model. -->
32
+
33
+ - **Repository:** [More Information Needed]
34
+ - **Paper [optional]:** [More Information Needed]
35
+ - **Demo [optional]:** [More Information Needed]
36
+
37
+ ## Uses
38
+
39
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
40
+
41
+ ### Direct Use
42
+
43
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
44
+
45
+ [More Information Needed]
46
+
47
+ ### Downstream Use [optional]
48
+
49
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Out-of-Scope Use
54
+
55
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
56
+
57
+ [More Information Needed]
58
+
59
+ ## Bias, Risks, and Limitations
60
+
61
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ### Recommendations
66
+
67
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
68
+
69
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
70
+
71
+ ## How to Get Started with the Model
72
+
73
+ Use the code below to get started with the model.
74
+
75
+ [More Information Needed]
76
+
77
+ ## Training Details
78
+
79
+ ### Training Data
80
+
81
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
82
+
83
+ [More Information Needed]
84
+
85
+ ### Training Procedure
86
+
87
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
88
+
89
+ #### Preprocessing [optional]
90
+
91
+ [More Information Needed]
92
+
93
+
94
+ #### Training Hyperparameters
95
+
96
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
97
+
98
+ #### Speeds, Sizes, Times [optional]
99
+
100
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
101
+
102
+ [More Information Needed]
103
+
104
+ ## Evaluation
105
+
106
+ <!-- This section describes the evaluation protocols and provides the results. -->
107
+
108
+ ### Testing Data, Factors & Metrics
109
+
110
+ #### Testing Data
111
+
112
+ <!-- This should link to a Dataset Card if possible. -->
113
+
114
+ [More Information Needed]
115
+
116
+ #### Factors
117
+
118
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Metrics
123
+
124
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
125
+
126
+ [More Information Needed]
127
+
128
+ ### Results
129
+
130
+ [More Information Needed]
131
+
132
+ #### Summary
133
+
134
+
135
+
136
+ ## Model Examination [optional]
137
+
138
+ <!-- Relevant interpretability work for the model goes here -->
139
+
140
+ [More Information Needed]
141
+
142
+ ## Environmental Impact
143
+
144
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
145
+
146
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
147
+
148
+ - **Hardware Type:** [More Information Needed]
149
+ - **Hours used:** [More Information Needed]
150
+ - **Cloud Provider:** [More Information Needed]
151
+ - **Compute Region:** [More Information Needed]
152
+ - **Carbon Emitted:** [More Information Needed]
153
+
154
+ ## Technical Specifications [optional]
155
+
156
+ ### Model Architecture and Objective
157
+
158
+ [More Information Needed]
159
+
160
+ ### Compute Infrastructure
161
+
162
+ [More Information Needed]
163
+
164
+ #### Hardware
165
+
166
+ [More Information Needed]
167
+
168
+ #### Software
169
+
170
+ [More Information Needed]
171
+
172
+ ## Citation [optional]
173
+
174
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
175
+
176
+ **BibTeX:**
177
+
178
+ [More Information Needed]
179
+
180
+ **APA:**
181
+
182
+ [More Information Needed]
183
+
184
+ ## Glossary [optional]
185
+
186
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
187
+
188
+ [More Information Needed]
189
+
190
+ ## More Information [optional]
191
+
192
+ [More Information Needed]
193
+
194
+ ## Model Card Authors [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Contact
199
+
200
+ [More Information Needed]
201
+ ### Framework versions
202
+
203
+ - PEFT 0.18.0
lora_adapters/adapter_config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": null,
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 16,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.0,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": [
25
+ "action_expert.model.layers.16.input_layernorm.dense",
26
+ "vlm.model.language_model.layers.13.input_layernorm",
27
+ "action_expert.model.layers.17.post_attention_layernorm.dense",
28
+ "action_expert.model.layers.6.input_layernorm.dense",
29
+ "action_expert.model.layers.5.post_attention_layernorm.dense",
30
+ "action_expert.model.layers.13.input_layernorm.dense",
31
+ "vlm.model.language_model.layers.7.input_layernorm",
32
+ "vlm.model.language_model.layers.0.post_attention_layernorm",
33
+ "action_expert.model.layers.3.post_attention_layernorm.dense",
34
+ "vlm.model.language_model.layers.6.input_layernorm",
35
+ "action_expert.model.layers.7.post_attention_layernorm.dense",
36
+ "action_expert.model.layers.10.post_attention_layernorm.dense",
37
+ "action_expert.model.layers.5.input_layernorm.dense",
38
+ "vlm.model.language_model.layers.14.post_attention_layernorm",
39
+ "suffix_embedder.time_mlp_in",
40
+ "vlm.model.language_model.layers.11.post_attention_layernorm",
41
+ "vlm.model.language_model.layers.15.post_attention_layernorm",
42
+ "vlm.model.language_model.layers.17.input_layernorm",
43
+ "action_expert.model.layers.8.post_attention_layernorm.dense",
44
+ "action_expert.model.layers.1.post_attention_layernorm.dense",
45
+ "action_expert.model.layers.14.post_attention_layernorm.dense",
46
+ "action_out_proj",
47
+ "action_expert.model.layers.14.input_layernorm.dense",
48
+ "vlm.model.language_model.layers.2.post_attention_layernorm",
49
+ "vlm.model.language_model.layers.1.post_attention_layernorm",
50
+ "action_expert.model.layers.4.input_layernorm.dense",
51
+ "action_expert.model.layers.15.input_layernorm.dense",
52
+ "vlm.model.vision_tower.vision_model.embeddings.position_embedding",
53
+ "vlm.model.language_model.layers.15.input_layernorm",
54
+ "action_expert.model.layers.4.post_attention_layernorm.dense",
55
+ "vlm.model.language_model.layers.17.post_attention_layernorm",
56
+ "vlm.model.language_model.layers.0.input_layernorm",
57
+ "action_expert.model.layers.0.post_attention_layernorm.dense",
58
+ "vlm.model.language_model.layers.6.post_attention_layernorm",
59
+ "vlm.model.language_model.layers.4.input_layernorm",
60
+ "action_expert.model.layers.0.input_layernorm.dense",
61
+ "vlm.model.language_model.layers.3.post_attention_layernorm",
62
+ "vlm.model.language_model.layers.5.post_attention_layernorm",
63
+ "action_expert.model.layers.11.post_attention_layernorm.dense",
64
+ "vlm.model.language_model.layers.8.input_layernorm",
65
+ "vlm.model.language_model.layers.12.post_attention_layernorm",
66
+ "suffix_embedder.state_proj",
67
+ "action_expert.model.layers.12.input_layernorm.dense",
68
+ "vlm.model.language_model.layers.9.post_attention_layernorm",
69
+ "action_expert.model.layers.15.post_attention_layernorm.dense",
70
+ "action_expert.model.layers.8.input_layernorm.dense",
71
+ "vlm.model.language_model.layers.12.input_layernorm",
72
+ "action_expert.model.layers.16.post_attention_layernorm.dense",
73
+ "vlm.model.language_model.layers.4.post_attention_layernorm",
74
+ "suffix_embedder.state_mlp_out",
75
+ "vlm.model.language_model.layers.8.post_attention_layernorm",
76
+ "action_expert.model.layers.17.input_layernorm.dense",
77
+ "vlm.model.language_model.layers.10.input_layernorm",
78
+ "action_expert.model.layers.2.input_layernorm.dense",
79
+ "vlm.model.language_model.layers.11.input_layernorm",
80
+ "action_expert.model.layers.13.post_attention_layernorm.dense",
81
+ "vlm.model.language_model.layers.16.input_layernorm",
82
+ "action_expert.model.layers.1.input_layernorm.dense",
83
+ "action_expert.model.layers.9.input_layernorm.dense",
84
+ "vlm.model.language_model.layers.13.post_attention_layernorm",
85
+ "action_expert.model.layers.7.input_layernorm.dense",
86
+ "action_expert.model.layers.6.post_attention_layernorm.dense",
87
+ "vlm.model.language_model.layers.2.input_layernorm",
88
+ "vlm.model.language_model.layers.16.post_attention_layernorm",
89
+ "vlm.model.language_model.layers.14.input_layernorm",
90
+ "vlm.model.language_model.layers.10.post_attention_layernorm",
91
+ "suffix_embedder.action_in_proj",
92
+ "vlm.model.vision_tower.vision_model.embeddings.patch_embedding",
93
+ "vlm.model.language_model.layers.7.post_attention_layernorm",
94
+ "vlm.model.language_model.layers.9.input_layernorm",
95
+ "action_expert.model.layers.12.post_attention_layernorm.dense",
96
+ "action_expert.model.layers.9.post_attention_layernorm.dense",
97
+ "vlm.model.language_model.layers.5.input_layernorm",
98
+ "suffix_embedder.time_mlp_out",
99
+ "action_expert.model.layers.10.input_layernorm.dense",
100
+ "action_expert.model.layers.3.input_layernorm.dense",
101
+ "action_expert.model.layers.11.input_layernorm.dense",
102
+ "suffix_embedder.state_mlp_in",
103
+ "action_expert.model.layers.2.post_attention_layernorm.dense",
104
+ "vlm.model.language_model.layers.3.input_layernorm",
105
+ "vlm.model.language_model.layers.1.input_layernorm"
106
+ ],
107
+ "peft_type": "LORA",
108
+ "peft_version": "0.18.0",
109
+ "qalora_group_size": 16,
110
+ "r": 16,
111
+ "rank_pattern": {},
112
+ "revision": null,
113
+ "target_modules": [
114
+ "v_proj",
115
+ "q_proj",
116
+ "fc1",
117
+ "o_proj",
118
+ "down_proj",
119
+ "out_proj",
120
+ "gate_proj",
121
+ "up_proj",
122
+ "k_proj",
123
+ "fc2"
124
+ ],
125
+ "target_parameters": null,
126
+ "task_type": "FEATURE_EXTRACTION",
127
+ "trainable_token_indices": null,
128
+ "use_dora": false,
129
+ "use_qalora": false,
130
+ "use_rslora": false
131
+ }
lora_adapters/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85261c03d43be91d17611d44d2257533f66c839cb1fe56fe934dbdf87d61b051
3
+ size 598447496
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8620fa70729c19e25f64c09597297f0e9266ebb8735b295712e0f86403462ce
3
+ size 7481487344
train_config.json ADDED
@@ -0,0 +1,274 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "data/pickup0102_310",
4
+ "root": "data/pickup0102_310",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "pi05",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 32
91
+ ]
92
+ },
93
+ "observation.state_eepose": {
94
+ "type": "STATE",
95
+ "shape": [
96
+ 14
97
+ ]
98
+ },
99
+ "observation.images.cam_high": {
100
+ "type": "VISUAL",
101
+ "shape": [
102
+ 3,
103
+ 360,
104
+ 480
105
+ ]
106
+ },
107
+ "observation.images.cam_left_wrist": {
108
+ "type": "VISUAL",
109
+ "shape": [
110
+ 3,
111
+ 360,
112
+ 480
113
+ ]
114
+ },
115
+ "observation.images.cam_right_wrist": {
116
+ "type": "VISUAL",
117
+ "shape": [
118
+ 3,
119
+ 360,
120
+ 480
121
+ ]
122
+ }
123
+ },
124
+ "output_features": {
125
+ "action": {
126
+ "type": "ACTION",
127
+ "shape": [
128
+ 32
129
+ ]
130
+ },
131
+ "action_eepose": {
132
+ "type": "ACTION",
133
+ "shape": [
134
+ 14
135
+ ]
136
+ }
137
+ },
138
+ "device": "cuda",
139
+ "use_amp": false,
140
+ "push_to_hub": false,
141
+ "repo_id": null,
142
+ "private": null,
143
+ "tags": null,
144
+ "license": null,
145
+ "pretrained_path": "checkpoints/pi05_base",
146
+ "paligemma_variant": "gemma_2b",
147
+ "action_expert_variant": "gemma_300m",
148
+ "dtype": "bfloat16",
149
+ "chunk_size": 50,
150
+ "n_action_steps": 50,
151
+ "max_state_dim": 32,
152
+ "max_action_dim": 32,
153
+ "state_cond": true,
154
+ "num_inference_steps": 10,
155
+ "time_sampling_beta_alpha": 1.5,
156
+ "time_sampling_beta_beta": 1.0,
157
+ "time_sampling_scale": 0.999,
158
+ "time_sampling_offset": 0.001,
159
+ "min_period": 0.004,
160
+ "max_period": 4.0,
161
+ "image_resolution": [
162
+ 224,
163
+ 224
164
+ ],
165
+ "empty_cameras": 0,
166
+ "tokenizer_max_length": 200,
167
+ "normalization_mapping": {
168
+ "VISUAL": "IDENTITY",
169
+ "STATE": "MEAN_STD",
170
+ "ACTION": "MEAN_STD"
171
+ },
172
+ "gradient_checkpointing": false,
173
+ "compile_model": false,
174
+ "compile_mode": "max-autotune",
175
+ "fuse_qkv": false,
176
+ "fuse_gate_up": false,
177
+ "optimizer_lr": 2.5e-05,
178
+ "optimizer_betas": [
179
+ 0.9,
180
+ 0.95
181
+ ],
182
+ "optimizer_eps": 1e-08,
183
+ "optimizer_weight_decay": 0.01,
184
+ "optimizer_grad_clip_norm": 1.0,
185
+ "scheduler_warmup_steps": 1000,
186
+ "scheduler_decay_steps": 30000,
187
+ "scheduler_decay_lr": 2.5e-06,
188
+ "vlm_config": {},
189
+ "action_expert_config": {}
190
+ },
191
+ "output_dir": "outputs/train/pi05_async_lora",
192
+ "job_name": "pi05_async_lora",
193
+ "resume": false,
194
+ "seed": 1000,
195
+ "num_workers": 1,
196
+ "batch_size": 2,
197
+ "steps": 50000,
198
+ "eval_freq": 20000,
199
+ "log_freq": 50,
200
+ "save_checkpoint": true,
201
+ "save_freq": 10000,
202
+ "use_policy_training_preset": false,
203
+ "optimizer": {
204
+ "type": "adamw",
205
+ "lr": 0.0001,
206
+ "weight_decay": 1e-10,
207
+ "grad_clip_norm": 10.0,
208
+ "betas": [
209
+ 0.9,
210
+ 0.95
211
+ ],
212
+ "eps": 1e-08
213
+ },
214
+ "scheduler": {
215
+ "type": "cosine_decay_with_warmup",
216
+ "num_warmup_steps": 1000,
217
+ "num_decay_steps": 50000,
218
+ "peak_lr": 5e-05,
219
+ "decay_lr": 2.5e-06
220
+ },
221
+ "eval": {
222
+ "n_episodes": 50,
223
+ "batch_size": 50,
224
+ "use_async_envs": false
225
+ },
226
+ "wandb": {
227
+ "enable": false,
228
+ "disable_artifact": true,
229
+ "project": "vlash",
230
+ "entity": null,
231
+ "notes": null,
232
+ "run_id": null,
233
+ "mode": null
234
+ },
235
+ "checkpoint_path": null,
236
+ "rename_map": {},
237
+ "max_delay_steps": 8,
238
+ "grad_accum_steps": 8,
239
+ "shared_observation": false,
240
+ "lora": {
241
+ "enable": true,
242
+ "backend": "peft",
243
+ "r": 16,
244
+ "alpha": 16,
245
+ "dropout": 0.0,
246
+ "extra_trainable_modules": [
247
+ "action_in_proj",
248
+ "action_out_proj",
249
+ "time_mlp_in",
250
+ "time_mlp_out",
251
+ "state_proj",
252
+ "state_mlp_in",
253
+ "state_mlp_out",
254
+ "embeddings",
255
+ "input_layernorm",
256
+ "post_attention_layernorm"
257
+ ],
258
+ "target_modules": [
259
+ "q_proj",
260
+ "k_proj",
261
+ "v_proj",
262
+ "o_proj",
263
+ "gate_proj",
264
+ "up_proj",
265
+ "down_proj",
266
+ "out_proj",
267
+ "fc1",
268
+ "fc2"
269
+ ],
270
+ "use_qlora": false,
271
+ "qlora_quant_type": "nf4",
272
+ "qlora_compute_dtype": "bfloat16"
273
+ }
274
+ }