tyrleng commited on
Commit
e963e9f
·
verified ·
1 Parent(s): 07f5db5

Upload folder using huggingface_hub

Browse files
adapter_config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/home/htx_hfs/.cache/huggingface/hub/models--nvidia--GR00T-N1.5-3B/snapshots/869830fc749c35f34771aa5209f923ac57e4564e",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 128,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.1,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 64,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "to_k",
29
+ "to_q",
30
+ "to_v"
31
+ ],
32
+ "target_parameters": null,
33
+ "task_type": "CAUSAL_LM",
34
+ "trainable_token_indices": null,
35
+ "use_dora": false,
36
+ "use_qalora": false,
37
+ "use_rslora": false
38
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0584f3e782bd23c3301bd2ba0970751c6c745c977fd25822c59ea799fcc5ad3
3
+ size 52447352
experiment_cfg/metadata.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "new_embodiment": {
3
+ "statistics": {
4
+ "state": {
5
+ "single_arm": {
6
+ "max": [
7
+ 47.293922424316406,
8
+ 85.06407165527344,
9
+ 98.92182922363281,
10
+ 99.82190704345703,
11
+ 100.0
12
+ ],
13
+ "min": [
14
+ -68.35970306396484,
15
+ -98.93946075439453,
16
+ -96.3162612915039,
17
+ -85.2181625366211,
18
+ -98.4443359375
19
+ ],
20
+ "mean": [
21
+ -6.940913904742286,
22
+ -25.85724106034202,
23
+ 28.523797604707415,
24
+ 52.71536749034758,
25
+ 27.215266155807516
26
+ ],
27
+ "std": [
28
+ 19.75424339384085,
29
+ 51.314506554995106,
30
+ 59.2130534373661,
31
+ 33.87771355357165,
32
+ 54.69567367291825
33
+ ],
34
+ "q01": [
35
+ -27.288997635225563,
36
+ -96.70928320105914,
37
+ -52.875409519629564,
38
+ -9.774115778162933,
39
+ -15.14155975615293
40
+ ],
41
+ "q99": [
42
+ 13.616432030218315,
43
+ 42.441777955836855,
44
+ 98.5565788919979,
45
+ 98.51933238105072,
46
+ 53.88044813818882
47
+ ]
48
+ },
49
+ "gripper": {
50
+ "max": [
51
+ 99.1750717163086
52
+ ],
53
+ "min": [
54
+ 2.1081576347351074
55
+ ],
56
+ "mean": [
57
+ 21.07318527244453
58
+ ],
59
+ "std": [
60
+ 26.338709383209714
61
+ ],
62
+ "q01": [
63
+ 2.6553785652237867
64
+ ],
65
+ "q99": [
66
+ 74.5051196283388
67
+ ]
68
+ }
69
+ },
70
+ "action": {
71
+ "single_arm": {
72
+ "max": [
73
+ 47.793426513671875,
74
+ 83.88998413085938,
75
+ 100.0,
76
+ 100.0,
77
+ 100.0
78
+ ],
79
+ "min": [
80
+ -68.8262939453125,
81
+ -100.0,
82
+ -99.28571319580078,
83
+ -85.4719467163086,
84
+ -98.78630065917969
85
+ ],
86
+ "mean": [
87
+ -6.8667323361293064,
88
+ -26.847045702351373,
89
+ 27.22029425063772,
90
+ 52.47788510953305,
91
+ 27.315668623257867
92
+ ],
93
+ "std": [
94
+ 19.77868276771674,
95
+ 50.54263160019092,
96
+ 60.42032483627416,
97
+ 34.15703516425812,
98
+ 54.738001697399135
99
+ ],
100
+ "q01": [
101
+ -27.469780154840237,
102
+ -96.78522400139514,
103
+ -55.83169491091546,
104
+ -10.587172411380422,
105
+ -15.430792427870232
106
+ ],
107
+ "q99": [
108
+ 13.910071169792403,
109
+ 41.52795310063691,
110
+ 99.86117365643008,
111
+ 99.4607319552107,
112
+ 54.258704182528675
113
+ ]
114
+ },
115
+ "gripper": {
116
+ "max": [
117
+ 100.0
118
+ ],
119
+ "min": [
120
+ 0.0
121
+ ],
122
+ "mean": [
123
+ 20.74569445098748
124
+ ],
125
+ "std": [
126
+ 27.084877952915345
127
+ ],
128
+ "q01": [
129
+ 1.868159578172138
130
+ ],
131
+ "q99": [
132
+ 75.8084468020434
133
+ ]
134
+ }
135
+ }
136
+ },
137
+ "modalities": {
138
+ "video": {
139
+ "front": {
140
+ "resolution": [
141
+ 640,
142
+ 480
143
+ ],
144
+ "channels": 3,
145
+ "fps": 30.0
146
+ },
147
+ "wrist": {
148
+ "resolution": [
149
+ 640,
150
+ 480
151
+ ],
152
+ "channels": 3,
153
+ "fps": 30.0
154
+ }
155
+ },
156
+ "state": {
157
+ "single_arm": {
158
+ "absolute": true,
159
+ "rotation_type": null,
160
+ "shape": [
161
+ 5
162
+ ],
163
+ "continuous": true
164
+ },
165
+ "gripper": {
166
+ "absolute": true,
167
+ "rotation_type": null,
168
+ "shape": [
169
+ 1
170
+ ],
171
+ "continuous": true
172
+ }
173
+ },
174
+ "action": {
175
+ "single_arm": {
176
+ "absolute": true,
177
+ "rotation_type": null,
178
+ "shape": [
179
+ 5
180
+ ],
181
+ "continuous": true
182
+ },
183
+ "gripper": {
184
+ "absolute": true,
185
+ "rotation_type": null,
186
+ "shape": [
187
+ 1
188
+ ],
189
+ "continuous": true
190
+ }
191
+ }
192
+ },
193
+ "embodiment_tag": "new_embodiment"
194
+ }
195
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:897937e0c8f35dfb8dd52dfc52dc7c5f8830686849808cb1c6edabaa8f65e10d
3
+ size 104959034
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff163c37547df0682c8162d102051c32337efe65588e48b71ce691931dfdcfe8
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75e4a20ae2ca44ba35c9de74b6451c85d5b2b2cf79791d071f4fe86be95cc1b2
3
+ size 1064
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff