float-lab commited on
Commit
1a8f688
·
verified ·
1 Parent(s): b30d02a

Upload folder using huggingface_hub

Browse files
Files changed (46) hide show
  1. README.md +207 -0
  2. adapter_config.json +38 -0
  3. adapter_model.safetensors +3 -0
  4. checkpoint-16000/README.md +207 -0
  5. checkpoint-16000/adapter_config.json +38 -0
  6. checkpoint-16000/adapter_model.safetensors +3 -0
  7. checkpoint-16000/experiment_cfg/metadata.json +195 -0
  8. checkpoint-16000/optimizer.pt +3 -0
  9. checkpoint-16000/rng_state.pth +3 -0
  10. checkpoint-16000/scheduler.pt +3 -0
  11. checkpoint-16000/trainer_state.json +0 -0
  12. checkpoint-17000/README.md +207 -0
  13. checkpoint-17000/adapter_config.json +38 -0
  14. checkpoint-17000/adapter_model.safetensors +3 -0
  15. checkpoint-17000/experiment_cfg/metadata.json +195 -0
  16. checkpoint-17000/optimizer.pt +3 -0
  17. checkpoint-17000/rng_state.pth +3 -0
  18. checkpoint-17000/scheduler.pt +3 -0
  19. checkpoint-17000/trainer_state.json +0 -0
  20. checkpoint-18000/README.md +207 -0
  21. checkpoint-18000/adapter_config.json +38 -0
  22. checkpoint-18000/adapter_model.safetensors +3 -0
  23. checkpoint-18000/experiment_cfg/metadata.json +195 -0
  24. checkpoint-18000/optimizer.pt +3 -0
  25. checkpoint-18000/rng_state.pth +3 -0
  26. checkpoint-18000/scheduler.pt +3 -0
  27. checkpoint-18000/trainer_state.json +0 -0
  28. checkpoint-19000/README.md +207 -0
  29. checkpoint-19000/adapter_config.json +38 -0
  30. checkpoint-19000/adapter_model.safetensors +3 -0
  31. checkpoint-19000/experiment_cfg/metadata.json +195 -0
  32. checkpoint-19000/optimizer.pt +3 -0
  33. checkpoint-19000/rng_state.pth +3 -0
  34. checkpoint-19000/scheduler.pt +3 -0
  35. checkpoint-19000/trainer_state.json +0 -0
  36. checkpoint-20000/README.md +207 -0
  37. checkpoint-20000/adapter_config.json +38 -0
  38. checkpoint-20000/adapter_model.safetensors +3 -0
  39. checkpoint-20000/experiment_cfg/metadata.json +195 -0
  40. checkpoint-20000/optimizer.pt +3 -0
  41. checkpoint-20000/rng_state.pth +3 -0
  42. checkpoint-20000/scheduler.pt +3 -0
  43. checkpoint-20000/trainer_state.json +0 -0
  44. experiment_cfg/metadata.json +195 -0
  45. trainer_state.json +0 -0
  46. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: nvidia/GR00T-N1.5-3B
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:nvidia/GR00T-N1.5-3B
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9663e276fa0529678792d5555af4c47d0be9d53a0014252cf288dd30c04d326b
3
+ size 52447352
checkpoint-16000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
checkpoint-16000/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
checkpoint-16000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94a57030077d1941a292f8e523df3fe565336ec924de84a61624577debf9aa58
3
+ size 52447352
checkpoint-16000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
checkpoint-16000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5f5280c12e0c35a7672d20218ecbc307ee91f0f2de223873701b4191c9891d8
3
+ size 104959499
checkpoint-16000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:813e31ccb8aafd66b6617c8d7585a4a3316582f1e4439c83678fddfd1964f9d2
3
+ size 14645
checkpoint-16000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26ef9df87f2ffcefbc6cc4a9b2f0b70732d553308b9e75c19dd7db38ad99b70e
3
+ size 1465
checkpoint-16000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-17000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
checkpoint-17000/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
checkpoint-17000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc6dd8d14f8c32ecc0622448ebfce77e9582cd064f06b66e466e5449426bb2a5
3
+ size 52447352
checkpoint-17000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
checkpoint-17000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be8e3621d8564346fce4c2314d8a0377942ae37aa7be82eb5a10134024be37e9
3
+ size 104959499
checkpoint-17000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b89b5b29346f539ef71e45db62e97edff528bb2f7444317de1199c85afae5b1f
3
+ size 14645
checkpoint-17000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f864ab1a29b091bc24e7ba13ded684437754dc733b2cd98dc882b1136479cc2c
3
+ size 1465
checkpoint-17000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-18000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
checkpoint-18000/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
checkpoint-18000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de9d201f39e517236d44d235ad414b534f93dc8d47dffc987a63c98ccebe07e9
3
+ size 52447352
checkpoint-18000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
checkpoint-18000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41d29a41bf4bdf9558baf6cc1999e48230abab59af5858fa96f3b11a23c1361e
3
+ size 104959499
checkpoint-18000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21686d86c820ad4f3405f6c59e396abf34d163d5f6d2c3c6f7e9b16f79059a07
3
+ size 14645
checkpoint-18000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b9ba779aac19f5f9fdf2399891476dbe504841d1284157ffc832a9d08c38c2
3
+ size 1465
checkpoint-18000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-19000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
checkpoint-19000/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
checkpoint-19000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6fd18c4016639011f74e335230219b5cd618793528f95fd4016af39059621c8
3
+ size 52447352
checkpoint-19000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
checkpoint-19000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b48c98f8775475bdd868d63086ccbff4142bd7121a19204a97ba43f44f749d3
3
+ size 104959499
checkpoint-19000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1738cd276bbf597ca5e41f7356b052cf308e9fb4a4a4ee9562a47430f8cdf5
3
+ size 14645
checkpoint-19000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58c2e92e32366392965f19e0f822f29f69b9ab1169b597e749c3b36ee341a779
3
+ size 1465
checkpoint-19000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: /home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.1
checkpoint-20000/adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/ajaya-rao/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_v",
30
+ "to_q"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
checkpoint-20000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9663e276fa0529678792d5555af4c47d0be9d53a0014252cf288dd30c04d326b
3
+ size 52447352
checkpoint-20000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
checkpoint-20000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c530a2fa3efa9da46359ca6fbaa8dd174e432c9a729b22ac05797c95bf9cfb9
3
+ size 104959499
checkpoint-20000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:594a61966820c0259a6307ea88d7d59ba09b471b05586b52973cae1c8a7e4f62
3
+ size 14645
checkpoint-20000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4db75d62db34e58c117f52804bbcf73f17f0787bfd8c708a85e80b7b061df5df
3
+ size 1465
checkpoint-20000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.7119408845901489,
8
+ 0.6336884498596191,
9
+ 1.6356228590011597,
10
+ 1.612607479095459,
11
+ 0.05677114799618721
12
+ ],
13
+ "min": [
14
+ -0.6996660828590393,
15
+ -1.807470679283142,
16
+ -0.421947717666626,
17
+ -1.0464303493499756,
18
+ -2.122013568878174
19
+ ],
20
+ "mean": [
21
+ 0.055539075285196304,
22
+ -0.846017599105835,
23
+ 1.0085504055023193,
24
+ 0.7250792384147644,
25
+ -1.54018235206604
26
+ ],
27
+ "std": [
28
+ 0.342788428068161,
29
+ 0.7538483142852783,
30
+ 0.4811265468597412,
31
+ 0.5023943781852722,
32
+ 0.13101021945476532
33
+ ],
34
+ "q01": [
35
+ -0.6413605809211731,
36
+ -1.778317928314209,
37
+ -0.2680672299861908,
38
+ -0.5278182625770569,
39
+ -1.7804966950416565
40
+ ],
41
+ "q99": [
42
+ 0.5124747157096863,
43
+ 0.55697101354599,
44
+ 1.5308417570590973,
45
+ 1.4034902346134186,
46
+ -1.2842555046081543
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 1.11394202709198
52
+ ],
53
+ "min": [
54
+ 0.0
55
+ ],
56
+ "mean": [
57
+ 0.2416374385356903
58
+ ],
59
+ "std": [
60
+ 0.2475089430809021
61
+ ],
62
+ "q01": [
63
+ 0.006137421354651451
64
+ ],
65
+ "q99": [
66
+ 0.9217026126384735
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 0.741093635559082,
74
+ 0.5953298807144165,
75
+ 1.6279510259628296,
76
+ 1.6003326177597046,
77
+ 0.06290856748819351
78
+ ],
79
+ "min": [
80
+ -0.741093635559082,
81
+ -1.8780510425567627,
82
+ -0.5554366707801819,
83
+ -1.1032015085220337,
84
+ -2.12508225440979
85
+ ],
86
+ "mean": [
87
+ 0.06122366711497307,
88
+ -0.9115849137306213,
89
+ 0.9640885591506958,
90
+ 0.7144355177879333,
91
+ -1.5393561124801636
92
+ ],
93
+ "std": [
94
+ 0.35077351331710815,
95
+ 0.7666987180709839,
96
+ 0.5521473288536072,
97
+ 0.5145319104194641,
98
+ 0.13183332979679108
99
+ ],
100
+ "q01": [
101
+ -0.6347781735658645,
102
+ -1.874982237815857,
103
+ -0.4076475277543068,
104
+ -0.5703352582454682,
105
+ -1.779852271080017
106
+ ],
107
+ "q99": [
108
+ 0.5339556932449341,
109
+ 0.5104953962564469,
110
+ 1.6049357652664185,
111
+ 1.412895840406418,
112
+ -1.2717811787128446
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 1.1246825456619263
118
+ ],
119
+ "min": [
120
+ -0.026084041222929955
121
+ ],
122
+ "mean": [
123
+ 0.22775891423225403
124
+ ],
125
+ "std": [
126
+ 0.2665541470050812
127
+ ],
128
+ "q01": [
129
+ -0.010740487836301327
130
+ ],
131
+ "q99": [
132
+ 0.9495204681158068
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "gripper_cam": {
140
+ "resolution": [
141
+ 1920,
142
+ 1080
143
+ ],
144
+ "channels": 3,
145
+ "fps": 60.0
146
+ },
147
+ "front_cam": {
148
+ "resolution": [
149
+ 1920,
150
+ 1080
151
+ ],
152
+ "channels": 3,
153
+ "fps": 60.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:641e0c419dc91be65448b1a64a395b7f0b68fee5ecdd831e72f938e46220c752
3
+ size 5777