KelinLiER commited on
Commit
b1bfb1a
·
verified ·
1 Parent(s): 0ae728a

Upload folder using huggingface_hub

Browse files
checkpoints/GR00T/StackCube/1000/checkpoint-100000/config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 32,
3
+ "action_head_cfg": {
4
+ "action_dim": 32,
5
+ "action_horizon": 16,
6
+ "add_pos_embed": true,
7
+ "backbone_embedding_dim": 2048,
8
+ "diffusion_model_cfg": {
9
+ "attention_head_dim": 48,
10
+ "cross_attention_dim": 2048,
11
+ "dropout": 0.2,
12
+ "final_dropout": true,
13
+ "interleave_self_attention": true,
14
+ "norm_type": "ada_norm",
15
+ "num_attention_heads": 32,
16
+ "num_layers": 16,
17
+ "output_dim": 1024,
18
+ "positional_embeddings": null
19
+ },
20
+ "hidden_size": 1024,
21
+ "input_embedding_dim": 1536,
22
+ "max_action_dim": 32,
23
+ "max_state_dim": 64,
24
+ "model_dtype": "float32",
25
+ "noise_beta_alpha": 1.5,
26
+ "noise_beta_beta": 1.0,
27
+ "noise_s": 0.999,
28
+ "num_inference_timesteps": 4,
29
+ "num_target_vision_tokens": 32,
30
+ "num_timestep_buckets": 1000,
31
+ "tune_diffusion_model": true,
32
+ "tune_projector": true,
33
+ "use_vlln": true,
34
+ "vl_self_attention_cfg": {
35
+ "attention_head_dim": 64,
36
+ "dropout": 0.2,
37
+ "final_dropout": true,
38
+ "num_attention_heads": 32,
39
+ "num_layers": 4,
40
+ "positional_embeddings": null
41
+ }
42
+ },
43
+ "action_horizon": 16,
44
+ "architectures": [
45
+ "GR00T_N1_5"
46
+ ],
47
+ "attn_implementation": null,
48
+ "backbone_cfg": {
49
+ "eagle_path": "NVEagle/eagle_er-qwen3_1_7B-Siglip2_400M_stage1_5_128gpu_er_v7_1mlp_nops",
50
+ "load_bf16": false,
51
+ "project_to_dim": null,
52
+ "reproject_vision": false,
53
+ "select_layer": 12,
54
+ "tune_llm": false,
55
+ "tune_visual": true,
56
+ "use_flash_attention": true
57
+ },
58
+ "compute_dtype": "bfloat16",
59
+ "hidden_size": 2048,
60
+ "model_dtype": "float32",
61
+ "model_type": "gr00t_n1_5",
62
+ "torch_dtype": "bfloat16",
63
+ "transformers_version": "4.51.3"
64
+ }
checkpoints/GR00T/StackCube/1000/checkpoint-100000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.5193183422088623,
8
+ 1.4723963737487793,
9
+ 0.22858981788158417,
10
+ -0.3866899311542511,
11
+ 0.3576759397983551,
12
+ 2.980123996734619,
13
+ 0.9351816177368164
14
+ ],
15
+ "min": [
16
+ -0.5494305491447449,
17
+ 0.05584031343460083,
18
+ -0.36008816957473755,
19
+ -2.657780647277832,
20
+ -0.2084779441356659,
21
+ 1.8309757709503174,
22
+ -1.9884469509124756
23
+ ],
24
+ "mean": [
25
+ 0.00399363087490201,
26
+ 0.628786027431488,
27
+ -6.783907883800566e-05,
28
+ -1.882095456123352,
29
+ 0.00020766883972100914,
30
+ 2.5070674419403076,
31
+ -0.7744830250740051
32
+ ],
33
+ "std": [
34
+ 0.19156166911125183,
35
+ 0.20839014649391174,
36
+ 0.07672057300806046,
37
+ 0.3725069463253021,
38
+ 0.07596796005964279,
39
+ 0.17491254210472107,
40
+ 0.47586438059806824
41
+ ],
42
+ "q01": [
43
+ -0.41479846835136414,
44
+ 0.2655360996723175,
45
+ -0.15876399144530295,
46
+ -2.5404844546318053,
47
+ -0.15167126059532166,
48
+ 2.0420314860343933,
49
+ -1.7789210772514343
50
+ ],
51
+ "q99": [
52
+ 0.4131187805533415,
53
+ 1.1966094970703125,
54
+ 0.16153444349765778,
55
+ -0.8495927453041077,
56
+ 0.15550431609153748,
57
+ 2.8279815435409548,
58
+ 0.22226445376873016
59
+ ]
60
+ },
61
+ "gripper": {
62
+ "max": [
63
+ 0.03999999910593033
64
+ ],
65
+ "min": [
66
+ 0.01826813817024231
67
+ ],
68
+ "mean": [
69
+ 0.035652436316013336
70
+ ],
71
+ "std": [
72
+ 0.006893928628414869
73
+ ],
74
+ "q01": [
75
+ 0.01830432191491127
76
+ ],
77
+ "q99": [
78
+ 0.03999999910593033
79
+ ]
80
+ }
81
+ },
82
+ "action": {
83
+ "single_arm": {
84
+ "max": [
85
+ 0.5213524699211121,
86
+ 1.474416732788086,
87
+ 0.23018206655979156,
88
+ -0.38181185722351074,
89
+ 0.36067888140678406,
90
+ 2.9835205078125,
91
+ 0.9366947412490845
92
+ ],
93
+ "min": [
94
+ -0.5499738454818726,
95
+ 0.05203799903392792,
96
+ -0.36180999875068665,
97
+ -2.658363103866577,
98
+ -0.21164120733737946,
99
+ 1.8290787935256958,
100
+ -1.9936093091964722
101
+ ],
102
+ "mean": [
103
+ 0.004018161445856094,
104
+ 0.6302363276481628,
105
+ -7.031633867882192e-05,
106
+ -1.8815363645553589,
107
+ 0.0002096072566928342,
108
+ 2.508046865463257,
109
+ -0.7744160890579224
110
+ ],
111
+ "std": [
112
+ 0.19219373166561127,
113
+ 0.20838859677314758,
114
+ 0.0769633799791336,
115
+ 0.37379103899002075,
116
+ 0.07622040808200836,
117
+ 0.1752668023109436,
118
+ 0.47757381200790405
119
+ ],
120
+ "q01": [
121
+ -0.41855645179748535,
122
+ 0.26554471254348755,
123
+ -0.15897227823734283,
124
+ -2.5431983709335326,
125
+ -0.15167154371738434,
126
+ 2.0375001430511475,
127
+ -1.7934351456165314
128
+ ],
129
+ "q99": [
130
+ 0.4136534631252289,
131
+ 1.1965739727020264,
132
+ 0.16145287454128265,
133
+ -0.849585235118866,
134
+ 0.1555059850215912,
135
+ 2.8299450874328613,
136
+ 0.22226347029209137
137
+ ]
138
+ },
139
+ "gripper": {
140
+ "max": [
141
+ 1.0
142
+ ],
143
+ "min": [
144
+ -1.0
145
+ ],
146
+ "mean": [
147
+ 0.730601966381073
148
+ ],
149
+ "std": [
150
+ 0.6836643218994141
151
+ ],
152
+ "q01": [
153
+ -1.0
154
+ ],
155
+ "q99": [
156
+ 1.0
157
+ ]
158
+ }
159
+ }
160
+ },
161
+ "modalities": {
162
+ "video": {
163
+ "front": {
164
+ "resolution": [
165
+ 512,
166
+ 512
167
+ ],
168
+ "channels": 3,
169
+ "fps": 10.0
170
+ }
171
+ },
172
+ "state": {
173
+ "single_arm": {
174
+ "absolute": true,
175
+ "rotation_type": null,
176
+ "shape": [
177
+ 7
178
+ ],
179
+ "continuous": true
180
+ },
181
+ "gripper": {
182
+ "absolute": true,
183
+ "rotation_type": null,
184
+ "shape": [
185
+ 1
186
+ ],
187
+ "continuous": true
188
+ }
189
+ },
190
+ "action": {
191
+ "single_arm": {
192
+ "absolute": true,
193
+ "rotation_type": null,
194
+ "shape": [
195
+ 7
196
+ ],
197
+ "continuous": true
198
+ },
199
+ "gripper": {
200
+ "absolute": true,
201
+ "rotation_type": null,
202
+ "shape": [
203
+ 1
204
+ ],
205
+ "continuous": true
206
+ }
207
+ }
208
+ },
209
+ "embodiment_tag": "new_embodiment"
210
+ }
211
+ }
checkpoints/GR00T/StackCube/1000/checkpoint-100000/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58f7a59cc635e25b8abcca3133c74a0a90ba677ff85d165a523ba784a4eeb586
3
+ size 4999367032
checkpoints/GR00T/StackCube/1000/checkpoint-100000/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b7d01d8b307036cef61013d89c5749f6d04845c1c324be914402be30062b598
3
+ size 2586705312
checkpoints/GR00T/StackCube/1000/checkpoint-100000/model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/GR00T/StackCube/1000/checkpoint-100000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e51a2a8f2720325fe1c3a1f8c44f8be0e1110253d0f637ebfef0b0367eb75bbd
3
+ size 8550720062
checkpoints/GR00T/StackCube/1000/checkpoint-100000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed470da38063aec327a8f278a96d4b3282471c2a3fdfb3c99cfa4d892d65a59
3
+ size 14244
checkpoints/GR00T/StackCube/1000/checkpoint-100000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d6c5b93324fa72d02f994789c5634c26492fde161591c5c12fb60aea61a68f2
3
+ size 1064
checkpoints/GR00T/StackCube/1000/checkpoint-100000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff