KelinLiER commited on
Commit
bf99713
·
verified ·
1 Parent(s): e66af4d

Upload folder using huggingface_hub

Browse files
checkpoints/GR00T/StackCube/500/checkpoint-100000/config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 32,
3
+ "action_head_cfg": {
4
+ "action_dim": 32,
5
+ "action_horizon": 16,
6
+ "add_pos_embed": true,
7
+ "backbone_embedding_dim": 2048,
8
+ "diffusion_model_cfg": {
9
+ "attention_head_dim": 48,
10
+ "cross_attention_dim": 2048,
11
+ "dropout": 0.2,
12
+ "final_dropout": true,
13
+ "interleave_self_attention": true,
14
+ "norm_type": "ada_norm",
15
+ "num_attention_heads": 32,
16
+ "num_layers": 16,
17
+ "output_dim": 1024,
18
+ "positional_embeddings": null
19
+ },
20
+ "hidden_size": 1024,
21
+ "input_embedding_dim": 1536,
22
+ "max_action_dim": 32,
23
+ "max_state_dim": 64,
24
+ "model_dtype": "float32",
25
+ "noise_beta_alpha": 1.5,
26
+ "noise_beta_beta": 1.0,
27
+ "noise_s": 0.999,
28
+ "num_inference_timesteps": 4,
29
+ "num_target_vision_tokens": 32,
30
+ "num_timestep_buckets": 1000,
31
+ "tune_diffusion_model": true,
32
+ "tune_projector": true,
33
+ "use_vlln": true,
34
+ "vl_self_attention_cfg": {
35
+ "attention_head_dim": 64,
36
+ "dropout": 0.2,
37
+ "final_dropout": true,
38
+ "num_attention_heads": 32,
39
+ "num_layers": 4,
40
+ "positional_embeddings": null
41
+ }
42
+ },
43
+ "action_horizon": 16,
44
+ "architectures": [
45
+ "GR00T_N1_5"
46
+ ],
47
+ "attn_implementation": null,
48
+ "backbone_cfg": {
49
+ "eagle_path": "NVEagle/eagle_er-qwen3_1_7B-Siglip2_400M_stage1_5_128gpu_er_v7_1mlp_nops",
50
+ "load_bf16": false,
51
+ "project_to_dim": null,
52
+ "reproject_vision": false,
53
+ "select_layer": 12,
54
+ "tune_llm": false,
55
+ "tune_visual": true,
56
+ "use_flash_attention": true
57
+ },
58
+ "compute_dtype": "bfloat16",
59
+ "hidden_size": 2048,
60
+ "model_dtype": "float32",
61
+ "model_type": "gr00t_n1_5",
62
+ "torch_dtype": "bfloat16",
63
+ "transformers_version": "4.51.3"
64
+ }
checkpoints/GR00T/StackCube/500/checkpoint-100000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.4799862205982208,
8
+ 1.4723963737487793,
9
+ 0.21917130053043365,
10
+ -0.3866899311542511,
11
+ 0.25145837664604187,
12
+ 2.973768711090088,
13
+ 0.3961699306964874
14
+ ],
15
+ "min": [
16
+ -0.493145227432251,
17
+ 0.06495869904756546,
18
+ -0.2660270929336548,
19
+ -2.6494877338409424,
20
+ -0.2084779441356659,
21
+ 1.8309757709503174,
22
+ -1.9064351320266724
23
+ ],
24
+ "mean": [
25
+ -0.0018909514183178544,
26
+ 0.6288999319076538,
27
+ -0.0021672516595572233,
28
+ -1.8816386461257935,
29
+ 0.0022188471630215645,
30
+ 2.5068047046661377,
31
+ -0.7831947207450867
32
+ ],
33
+ "std": [
34
+ 0.19031475484371185,
35
+ 0.21112516522407532,
36
+ 0.07426253706216812,
37
+ 0.3778972923755646,
38
+ 0.0734574943780899,
39
+ 0.17731702327728271,
40
+ 0.4678885340690613
41
+ ],
42
+ "q01": [
43
+ -0.4116199314594269,
44
+ 0.2655360996723175,
45
+ -0.1563243567943573,
46
+ -2.543651580810547,
47
+ -0.14918068051338196,
48
+ 2.031217088699341,
49
+ -1.8015106916427612
50
+ ],
51
+ "q99": [
52
+ 0.3976084291934967,
53
+ 1.2021321368217506,
54
+ 0.14960874617099762,
55
+ -0.8438157194852827,
56
+ 0.15277107059955597,
57
+ 2.8299596309661865,
58
+ 0.17928330600261688
59
+ ]
60
+ },
61
+ "gripper": {
62
+ "max": [
63
+ 0.03999999910593033
64
+ ],
65
+ "min": [
66
+ 0.01827898807823658
67
+ ],
68
+ "mean": [
69
+ 0.03566020354628563
70
+ ],
71
+ "std": [
72
+ 0.006922334898263216
73
+ ],
74
+ "q01": [
75
+ 0.01830417472869158
76
+ ],
77
+ "q99": [
78
+ 0.03999999910593033
79
+ ]
80
+ }
81
+ },
82
+ "action": {
83
+ "single_arm": {
84
+ "max": [
85
+ 0.48105430603027344,
86
+ 1.474416732788086,
87
+ 0.22043050825595856,
88
+ -0.38181185722351074,
89
+ 0.25391659140586853,
90
+ 2.9778687953948975,
91
+ 0.4008225202560425
92
+ ],
93
+ "min": [
94
+ -0.4943549633026123,
95
+ 0.06074262410402298,
96
+ -0.26876476407051086,
97
+ -2.6502506732940674,
98
+ -0.21164120733737946,
99
+ 1.8290787935256958,
100
+ -1.9113181829452515
101
+ ],
102
+ "mean": [
103
+ -0.0019036835292354226,
104
+ 0.6303254961967468,
105
+ -0.0021859537810087204,
106
+ -1.8811068534851074,
107
+ 0.0022235822398215532,
108
+ 2.5078065395355225,
109
+ -0.7831820249557495
110
+ ],
111
+ "std": [
112
+ 0.19093813002109528,
113
+ 0.21112795174121857,
114
+ 0.07449864596128464,
115
+ 0.3792603313922882,
116
+ 0.07370107620954514,
117
+ 0.1776680201292038,
118
+ 0.46956679224967957
119
+ ],
120
+ "q01": [
121
+ -0.4116838872432709,
122
+ 0.26554471254348755,
123
+ -0.15631593763828278,
124
+ -2.54366135597229,
125
+ -0.1491798758506775,
126
+ 2.0278784155845644,
127
+ -1.8017024052143096
128
+ ],
129
+ "q99": [
130
+ 0.3975871205329895,
131
+ 1.2065871524810792,
132
+ 0.14970773458480835,
133
+ -0.836275672912596,
134
+ 0.15277215838432312,
135
+ 2.8299450874328613,
136
+ 0.1792844533920288
137
+ ]
138
+ },
139
+ "gripper": {
140
+ "max": [
141
+ 1.0
142
+ ],
143
+ "min": [
144
+ -1.0
145
+ ],
146
+ "mean": [
147
+ 0.7288543581962585
148
+ ],
149
+ "std": [
150
+ 0.6843376755714417
151
+ ],
152
+ "q01": [
153
+ -1.0
154
+ ],
155
+ "q99": [
156
+ 1.0
157
+ ]
158
+ }
159
+ }
160
+ },
161
+ "modalities": {
162
+ "video": {
163
+ "front": {
164
+ "resolution": [
165
+ 512,
166
+ 512
167
+ ],
168
+ "channels": 3,
169
+ "fps": 10.0
170
+ }
171
+ },
172
+ "state": {
173
+ "single_arm": {
174
+ "absolute": true,
175
+ "rotation_type": null,
176
+ "shape": [
177
+ 7
178
+ ],
179
+ "continuous": true
180
+ },
181
+ "gripper": {
182
+ "absolute": true,
183
+ "rotation_type": null,
184
+ "shape": [
185
+ 1
186
+ ],
187
+ "continuous": true
188
+ }
189
+ },
190
+ "action": {
191
+ "single_arm": {
192
+ "absolute": true,
193
+ "rotation_type": null,
194
+ "shape": [
195
+ 7
196
+ ],
197
+ "continuous": true
198
+ },
199
+ "gripper": {
200
+ "absolute": true,
201
+ "rotation_type": null,
202
+ "shape": [
203
+ 1
204
+ ],
205
+ "continuous": true
206
+ }
207
+ }
208
+ },
209
+ "embodiment_tag": "new_embodiment"
210
+ }
211
+ }
checkpoints/GR00T/StackCube/500/checkpoint-100000/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5a7764c6cdf0cc4adb4c74149f70f466928719855daf9785d8153f58d899bb0
3
+ size 4999367032
checkpoints/GR00T/StackCube/500/checkpoint-100000/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49cdd1cb25bb791c19aae4ce16474a32b698841a9d35c8b011f6beaefe9c05cc
3
+ size 2586705312
checkpoints/GR00T/StackCube/500/checkpoint-100000/model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/GR00T/StackCube/500/checkpoint-100000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c47ace2343a2b185c6d5d038edab534dbf150d39b7eb79296980c1d5bc68c00b
3
+ size 8550720062
checkpoints/GR00T/StackCube/500/checkpoint-100000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ffe0f44325869347a2cd47cdbf0ffe25f11678064c40b5d873d69f3b1523887
3
+ size 14244
checkpoints/GR00T/StackCube/500/checkpoint-100000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d6c5b93324fa72d02f994789c5634c26492fde161591c5c12fb60aea61a68f2
3
+ size 1064
checkpoints/GR00T/StackCube/500/checkpoint-100000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff