tyrleng commited on
Commit
ea084c8
·
verified ·
1 Parent(s): 49d407e

Upload folder using huggingface_hub

Browse files
adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/htx_hfs/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_q",
29
+ "to_k",
30
+ "to_v"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4ec27592f12fa9d8d27a756c2fe4c2a7c0138e49d67e3410891378f70867bf1
3
+ size 52447352
experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 26.38411521911621,
8
+ 53.8775520324707,
9
+ 98.83355712890625,
10
+ 97.96107482910156,
11
+ 13.910284996032715
12
+ ],
13
+ "min": [
14
+ -24.62771987915039,
15
+ -97.95918273925781,
16
+ -62.5841178894043,
17
+ 7.7849860191345215,
18
+ -82.8726806640625
19
+ ],
20
+ "mean": [
21
+ 1.4359424501178135,
22
+ -14.799132693122338,
23
+ 12.745063604994787,
24
+ 65.77702735327829,
25
+ -22.556556278210323
26
+ ],
27
+ "std": [
28
+ 10.927660471650677,
29
+ 52.84280925356305,
30
+ 52.95757307514587,
31
+ 18.220004788429243,
32
+ 16.110957968125668
33
+ ],
34
+ "q01": [
35
+ -16.382637561509558,
36
+ -95.99786083000954,
37
+ -34.09412917342053,
38
+ 36.83884015066295,
39
+ -44.516502948492466
40
+ ],
41
+ "q99": [
42
+ 16.12872118545962,
43
+ 37.9468350507924,
44
+ 98.4659126743856,
45
+ 96.97923508027905,
46
+ -3.599417217414874
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 49.384403228759766
52
+ ],
53
+ "min": [
54
+ 2.4623804092407227
55
+ ],
56
+ "mean": [
57
+ 12.15276024152423
58
+ ],
59
+ "std": [
60
+ 10.538106024943849
61
+ ],
62
+ "q01": [
63
+ 2.723670017749419
64
+ ],
65
+ "q99": [
66
+ 30.10706680889146
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 27.305442810058594,
74
+ 54.183589935302734,
75
+ 99.64539337158203,
76
+ 100.0,
77
+ 14.236706733703613
78
+ ],
79
+ "min": [
80
+ -25.056230545043945,
81
+ -100.0,
82
+ -65.5141830444336,
83
+ 6.324786186218262,
84
+ -82.84734344482422
85
+ ],
86
+ "mean": [
87
+ 1.4937210656009372,
88
+ -15.826116765464835,
89
+ 11.389542859550257,
90
+ 65.58325798896549,
91
+ -22.47620920207392
92
+ ],
93
+ "std": [
94
+ 10.936782945171393,
95
+ 52.26455865603432,
96
+ 53.623327115525385,
97
+ 18.577865660523635,
98
+ 16.175134418673846
99
+ ],
100
+ "q01": [
101
+ -16.71871758023194,
102
+ -96.87557811692852,
103
+ -36.73666969306614,
104
+ 35.36386185937757,
105
+ -44.79180684755351
106
+ ],
107
+ "q99": [
108
+ 16.40704039036922,
109
+ 37.72761073170104,
110
+ 98.44483915552776,
111
+ 97.43717807393482,
112
+ -3.1738026832993746
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 49.772727966308594
118
+ ],
119
+ "min": [
120
+ 0.0
121
+ ],
122
+ "mean": [
123
+ 10.961273960274072
124
+ ],
125
+ "std": [
126
+ 11.533940895963871
127
+ ],
128
+ "q01": [
129
+ 1.8251763963104901
130
+ ],
131
+ "q99": [
132
+ 30.386979082929017
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "front": {
140
+ "resolution": [
141
+ 640,
142
+ 480
143
+ ],
144
+ "channels": 3,
145
+ "fps": 30.0
146
+ },
147
+ "wrist": {
148
+ "resolution": [
149
+ 640,
150
+ 480
151
+ ],
152
+ "channels": 3,
153
+ "fps": 30.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:246e3ce72b0e05b49b0d679a197c6a540979add278b76b6cf27417b4eb6df3d2
3
+ size 104959034
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:234dcd43107c80cdeb25a8372ddaf7b3e31d5af82078372ca6e782f33cf94206
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6066808d56ac8dada07f0f795515f59738ab2aae405c769d611f983355bb27a8
3
+ size 1064
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff