KelinLiER commited on
Commit
59a7b8e
·
verified ·
1 Parent(s): 9b782de

Upload folder using huggingface_hub

Browse files
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "action_dim": 32,
3
+ "action_head_cfg": {
4
+ "action_dim": 32,
5
+ "action_horizon": 16,
6
+ "add_pos_embed": true,
7
+ "backbone_embedding_dim": 2048,
8
+ "diffusion_model_cfg": {
9
+ "attention_head_dim": 48,
10
+ "cross_attention_dim": 2048,
11
+ "dropout": 0.2,
12
+ "final_dropout": true,
13
+ "interleave_self_attention": true,
14
+ "norm_type": "ada_norm",
15
+ "num_attention_heads": 32,
16
+ "num_layers": 16,
17
+ "output_dim": 1024,
18
+ "positional_embeddings": null
19
+ },
20
+ "hidden_size": 1024,
21
+ "input_embedding_dim": 1536,
22
+ "max_action_dim": 32,
23
+ "max_state_dim": 64,
24
+ "model_dtype": "float32",
25
+ "noise_beta_alpha": 1.5,
26
+ "noise_beta_beta": 1.0,
27
+ "noise_s": 0.999,
28
+ "num_inference_timesteps": 4,
29
+ "num_target_vision_tokens": 32,
30
+ "num_timestep_buckets": 1000,
31
+ "tune_diffusion_model": true,
32
+ "tune_projector": true,
33
+ "use_vlln": true,
34
+ "vl_self_attention_cfg": {
35
+ "attention_head_dim": 64,
36
+ "dropout": 0.2,
37
+ "final_dropout": true,
38
+ "num_attention_heads": 32,
39
+ "num_layers": 4,
40
+ "positional_embeddings": null
41
+ }
42
+ },
43
+ "action_horizon": 16,
44
+ "architectures": [
45
+ "GR00T_N1_5"
46
+ ],
47
+ "attn_implementation": null,
48
+ "backbone_cfg": {
49
+ "eagle_path": "NVEagle/eagle_er-qwen3_1_7B-Siglip2_400M_stage1_5_128gpu_er_v7_1mlp_nops",
50
+ "load_bf16": false,
51
+ "project_to_dim": null,
52
+ "reproject_vision": false,
53
+ "select_layer": 12,
54
+ "tune_llm": false,
55
+ "tune_visual": true,
56
+ "use_flash_attention": true
57
+ },
58
+ "compute_dtype": "bfloat16",
59
+ "hidden_size": 2048,
60
+ "model_dtype": "float32",
61
+ "model_type": "gr00t_n1_5",
62
+ "torch_dtype": "bfloat16",
63
+ "transformers_version": "4.51.3"
64
+ }
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/experiment_cfg/metadata.json ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 0.4850326180458069,
8
+ 1.4967026710510254,
9
+ 0.17114850878715515,
10
+ -0.3953115940093994,
11
+ 0.2793852388858795,
12
+ 2.8790173530578613,
13
+ 0.08779283612966537
14
+ ],
15
+ "min": [
16
+ -0.5562708973884583,
17
+ -0.013451993465423584,
18
+ -0.2617835998535156,
19
+ -2.6419124603271484,
20
+ -0.15718697011470795,
21
+ 1.873246192932129,
22
+ -2.2423808574676514
23
+ ],
24
+ "mean": [
25
+ 0.1307424157857895,
26
+ 0.5650708675384521,
27
+ -0.014015366323292255,
28
+ -1.9048140048980713,
29
+ -0.009533693082630634,
30
+ 2.4560585021972656,
31
+ -0.6259320974349976
32
+ ],
33
+ "std": [
34
+ 0.200121209025383,
35
+ 0.16069450974464417,
36
+ 0.059123557060956955,
37
+ 0.24703119695186615,
38
+ 0.05858810245990753,
39
+ 0.11585842072963715,
40
+ 0.3766859173774719
41
+ ],
42
+ "q01": [
43
+ -0.4516954720020294,
44
+ 0.21395637094974518,
45
+ -0.17290454640984534,
46
+ -2.4090957641601562,
47
+ -0.11889279633760452,
48
+ 2.185713529586792,
49
+ -1.9243584430217744
50
+ ],
51
+ "q99": [
52
+ 0.4381653964519501,
53
+ 1.0068663942813876,
54
+ 0.10277940914034862,
55
+ -1.2384997546672818,
56
+ 0.18799399480223658,
57
+ 2.740046606063843,
58
+ -0.04003881290555
59
+ ]
60
+ },
61
+ "gripper": {
62
+ "max": [
63
+ 0.03999999910593033
64
+ ],
65
+ "min": [
66
+ 0.013646434061229229
67
+ ],
68
+ "mean": [
69
+ 0.020496640354394913
70
+ ],
71
+ "std": [
72
+ 0.007791673298925161
73
+ ],
74
+ "q01": [
75
+ 0.013844946175813675
76
+ ],
77
+ "q99": [
78
+ 0.03999999910593033
79
+ ]
80
+ }
81
+ },
82
+ "action": {
83
+ "single_arm": {
84
+ "max": [
85
+ 0.5285579562187195,
86
+ 1.4984270334243774,
87
+ 0.1840890794992447,
88
+ 2.3694278276172396e-38,
89
+ 0.2826608717441559,
90
+ 2.882380723953247,
91
+ 0.1221635490655899
92
+ ],
93
+ "min": [
94
+ -0.5565145015716553,
95
+ -0.0185831431299448,
96
+ -0.262246310710907,
97
+ -2.641247510910034,
98
+ -0.15789692103862762,
99
+ 0.0,
100
+ -2.245187997817993
101
+ ],
102
+ "mean": [
103
+ 0.16032959520816803,
104
+ 0.569538414478302,
105
+ 0.005173633806407452,
106
+ -1.8937230110168457,
107
+ -0.001417750958353281,
108
+ 2.461942434310913,
109
+ -0.6035827994346619
110
+ ],
111
+ "std": [
112
+ 0.21533340215682983,
113
+ 0.15998932719230652,
114
+ 0.06143704429268837,
115
+ 0.24494345486164093,
116
+ 0.057319220155477524,
117
+ 0.11777312308549881,
118
+ 0.38947156071662903
119
+ ],
120
+ "q01": [
121
+ -0.4519099050760269,
122
+ 0.21485409140586853,
123
+ -0.17316834628582,
124
+ -2.402741022109985,
125
+ -0.11028626263141632,
126
+ 2.186081886291504,
127
+ -1.9257729756832123
128
+ ],
129
+ "q99": [
130
+ 0.48706454038619995,
131
+ 1.0073811304569256,
132
+ 0.11604302376508713,
133
+ -1.2352406561374654,
134
+ 0.18894457504153267,
135
+ 2.7415695333480836,
136
+ 0.012241317890584469
137
+ ]
138
+ },
139
+ "gripper": {
140
+ "max": [
141
+ 1.0
142
+ ],
143
+ "min": [
144
+ -1.0
145
+ ],
146
+ "mean": [
147
+ -0.7483277916908264
148
+ ],
149
+ "std": [
150
+ 0.6628010869026184
151
+ ],
152
+ "q01": [
153
+ -1.0
154
+ ],
155
+ "q99": [
156
+ 1.0
157
+ ]
158
+ }
159
+ }
160
+ },
161
+ "modalities": {
162
+ "video": {
163
+ "front": {
164
+ "resolution": [
165
+ 512,
166
+ 512
167
+ ],
168
+ "channels": 3,
169
+ "fps": 10.0
170
+ }
171
+ },
172
+ "state": {
173
+ "single_arm": {
174
+ "absolute": true,
175
+ "rotation_type": null,
176
+ "shape": [
177
+ 7
178
+ ],
179
+ "continuous": true
180
+ },
181
+ "gripper": {
182
+ "absolute": true,
183
+ "rotation_type": null,
184
+ "shape": [
185
+ 1
186
+ ],
187
+ "continuous": true
188
+ }
189
+ },
190
+ "action": {
191
+ "single_arm": {
192
+ "absolute": true,
193
+ "rotation_type": null,
194
+ "shape": [
195
+ 7
196
+ ],
197
+ "continuous": true
198
+ },
199
+ "gripper": {
200
+ "absolute": true,
201
+ "rotation_type": null,
202
+ "shape": [
203
+ 1
204
+ ],
205
+ "continuous": true
206
+ }
207
+ }
208
+ },
209
+ "embodiment_tag": "new_embodiment"
210
+ }
211
+ }
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7db8d665a5308cd56f79337f96c62142ac6ac86fd6b2d68ff89350eecdab7877
3
+ size 4999367032
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afa72a8af110e40b514adde464162bdc4cdd116480fe303165b53a727a45da91
3
+ size 2586705312
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62102caf7e7658656d3ef7c9fe86cbc1af2fac51772bea97472ed1a322a03f93
3
+ size 8550720062
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e19fc3d9552aecdad9b56407d892348e528dce761333db8fbd3f16b1610c1e1
3
+ size 14244
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d6c5b93324fa72d02f994789c5634c26492fde161591c5c12fb60aea61a68f2
3
+ size 1064
checkpoints/GR00T/PegInsertion/750/checkpoint-100000/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff