euijinrnd commited on
Commit
0e85bb2
·
verified ·
1 Parent(s): e6bab54

Add files using upload-large-folder tool

Browse files
config.json ADDED
@@ -0,0 +1,231 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": "c90603846ff2e8fa8c67e1fae4f5ce3d1d3c46e6",
3
+ "_name_or_path": "nvidia/Eagle2-2B",
4
+ "action_dim": 20,
5
+ "action_head": "DiT",
6
+ "action_head_hidden_dim": 1024,
7
+ "action_len": 20,
8
+ "aggregation": "None",
9
+ "architectures": [
10
+ "Eagle2_2BVLA"
11
+ ],
12
+ "auto_map": {},
13
+ "bimanual_action_masking": true,
14
+ "dataset_statistics_path": null,
15
+ "denoiser": "FM",
16
+ "diffusion_batch": 32,
17
+ "dit_size": "DiT-L",
18
+ "downsample_ratio": 0.5,
19
+ "dynamic_image_size": true,
20
+ "efficient_loss": true,
21
+ "enable_cfg": true,
22
+ "force_image_size": 448,
23
+ "global_normalization": true,
24
+ "hz_interpolate": null,
25
+ "image_size": 224,
26
+ "interpolate_gripper": false,
27
+ "keep_aspect_ratio": false,
28
+ "knowledge_insulation": false,
29
+ "llm_config": {
30
+ "_attn_implementation_autoset": true,
31
+ "_name_or_path": "./pretrained/Qwen2_5-1_5B-Instruct",
32
+ "add_cross_attention": false,
33
+ "architectures": [
34
+ "Qwen2ForCausalLM"
35
+ ],
36
+ "attention_dropout": 0.0,
37
+ "auto_map": {
38
+ "AutoConfig": "configuration_qwen2.Qwen2Config",
39
+ "AutoModel": "modeling_qwen2.Qwen2Model",
40
+ "AutoModelForCausalLM": "modeling_qwen2.Qwen2ForCausalLM"
41
+ },
42
+ "bad_words_ids": null,
43
+ "begin_suppress_tokens": null,
44
+ "bos_token_id": 151643,
45
+ "chunk_size_feed_forward": 0,
46
+ "cross_attention_hidden_size": null,
47
+ "decoder_start_token_id": null,
48
+ "diversity_penalty": 0.0,
49
+ "do_sample": false,
50
+ "early_stopping": false,
51
+ "encoder_no_repeat_ngram_size": 0,
52
+ "eos_token_id": 151645,
53
+ "exponential_decay_length_penalty": null,
54
+ "finetuning_task": null,
55
+ "forced_bos_token_id": null,
56
+ "forced_eos_token_id": null,
57
+ "hidden_act": "silu",
58
+ "hidden_size": 1536,
59
+ "id2label": {
60
+ "0": "LABEL_0",
61
+ "1": "LABEL_1"
62
+ },
63
+ "initializer_range": 0.02,
64
+ "intermediate_size": 8960,
65
+ "is_decoder": false,
66
+ "is_encoder_decoder": false,
67
+ "label2id": {
68
+ "LABEL_0": 0,
69
+ "LABEL_1": 1
70
+ },
71
+ "length_penalty": 1.0,
72
+ "max_length": 20,
73
+ "max_position_embeddings": 32768,
74
+ "max_window_layers": 21,
75
+ "min_length": 0,
76
+ "model_type": "qwen2",
77
+ "no_repeat_ngram_size": 0,
78
+ "num_attention_heads": 12,
79
+ "num_beam_groups": 1,
80
+ "num_beams": 1,
81
+ "num_hidden_layers": 28,
82
+ "num_key_value_heads": 2,
83
+ "num_return_sequences": 1,
84
+ "output_attentions": false,
85
+ "output_hidden_states": false,
86
+ "output_scores": false,
87
+ "pad_token_id": null,
88
+ "prefix": null,
89
+ "problem_type": null,
90
+ "pruned_heads": {},
91
+ "remove_invalid_values": false,
92
+ "repetition_penalty": 1.0,
93
+ "return_dict": true,
94
+ "return_dict_in_generate": false,
95
+ "rms_norm_eps": 1e-06,
96
+ "rope_scaling": null,
97
+ "rope_theta": 1000000.0,
98
+ "sep_token_id": null,
99
+ "sliding_window": 32768,
100
+ "suppress_tokens": null,
101
+ "task_specific_params": null,
102
+ "temperature": 1.0,
103
+ "tf_legacy_loss": false,
104
+ "tie_encoder_decoder": false,
105
+ "tie_word_embeddings": true,
106
+ "tokenizer_class": null,
107
+ "top_k": 50,
108
+ "top_p": 1.0,
109
+ "torch_dtype": "bfloat16",
110
+ "torchscript": false,
111
+ "transformers_version": "4.50.0.dev0",
112
+ "typical_p": 1.0,
113
+ "use_bfloat16": false,
114
+ "use_cache": false,
115
+ "use_sliding_window": false,
116
+ "vocab_size": 151674
117
+ },
118
+ "loss_version": "v4",
119
+ "max_dynamic_patch": 12,
120
+ "min_dynamic_patch": 1,
121
+ "mlp_checkpoint": true,
122
+ "model_path": "nvidia/Eagle2-2B",
123
+ "model_type": "Eagle2_2BVLA",
124
+ "modeling": "denoising",
125
+ "normalization": "quantile",
126
+ "num_readouts": 1,
127
+ "pad2square": false,
128
+ "pre_feature_reduction": false,
129
+ "ps_version": "v2",
130
+ "readout_token_as_eos": true,
131
+ "return_text": null,
132
+ "select_layer": -1,
133
+ "state_dim": 20,
134
+ "stopping_token": "|",
135
+ "template": "qwen2-chat",
136
+ "test_denoising_steps": 10,
137
+ "torch_dtype": "bfloat16",
138
+ "train_denoising_steps": 100,
139
+ "transformers_version": null,
140
+ "use_backbone_lora": 0,
141
+ "use_llm_lora": 0,
142
+ "use_thumbnail": true,
143
+ "vision_config": {
144
+ "_attn_implementation_autoset": true,
145
+ "_name_or_path": "",
146
+ "add_cross_attention": false,
147
+ "architectures": [
148
+ "SiglipVisionModel"
149
+ ],
150
+ "attention_dropout": 0.0,
151
+ "auto_map": {
152
+ "AutoConfig": "configuration_siglip.SiglipVisionConfig",
153
+ "AutoModel": "modeling_siglip.SiglipVisionModel"
154
+ },
155
+ "bad_words_ids": null,
156
+ "begin_suppress_tokens": null,
157
+ "bos_token_id": null,
158
+ "chunk_size_feed_forward": 0,
159
+ "cross_attention_hidden_size": null,
160
+ "decoder_start_token_id": null,
161
+ "diversity_penalty": 0.0,
162
+ "do_sample": false,
163
+ "drop_path_rate": 0.1,
164
+ "early_stopping": false,
165
+ "encoder_no_repeat_ngram_size": 0,
166
+ "eos_token_id": null,
167
+ "exponential_decay_length_penalty": null,
168
+ "finetuning_task": null,
169
+ "forced_bos_token_id": null,
170
+ "forced_eos_token_id": null,
171
+ "hidden_act": "gelu_pytorch_tanh",
172
+ "hidden_size": 1152,
173
+ "id2label": {
174
+ "0": "LABEL_0",
175
+ "1": "LABEL_1"
176
+ },
177
+ "image_size": 448,
178
+ "intermediate_size": 4304,
179
+ "is_decoder": false,
180
+ "is_encoder_decoder": false,
181
+ "label2id": {
182
+ "LABEL_0": 0,
183
+ "LABEL_1": 1
184
+ },
185
+ "layer_norm_eps": 1e-06,
186
+ "length_penalty": 1.0,
187
+ "max_length": 20,
188
+ "min_length": 0,
189
+ "model_type": "siglip_vision_model",
190
+ "no_repeat_ngram_size": 0,
191
+ "num_attention_heads": 16,
192
+ "num_beam_groups": 1,
193
+ "num_beams": 1,
194
+ "num_channels": 3,
195
+ "num_hidden_layers": 27,
196
+ "num_image_tokens": 1024,
197
+ "num_return_sequences": 1,
198
+ "output_attentions": false,
199
+ "output_hidden_states": false,
200
+ "output_scores": false,
201
+ "pad_token_id": null,
202
+ "patch_size": 14,
203
+ "prefix": null,
204
+ "problem_type": null,
205
+ "projection_dim": 2048,
206
+ "projector_hidden_act": "gelu_fast",
207
+ "pruned_heads": {},
208
+ "remove_invalid_values": false,
209
+ "repetition_penalty": 1.0,
210
+ "return_dict": true,
211
+ "return_dict_in_generate": false,
212
+ "sep_token_id": null,
213
+ "suppress_tokens": null,
214
+ "task_specific_params": null,
215
+ "temperature": 1.0,
216
+ "tf_legacy_loss": false,
217
+ "tie_encoder_decoder": false,
218
+ "tie_word_embeddings": true,
219
+ "tokenizer_class": null,
220
+ "top_k": 50,
221
+ "top_p": 1.0,
222
+ "torch_dtype": "bfloat16",
223
+ "torchscript": false,
224
+ "transformers_version": "4.50.0.dev0",
225
+ "typical_p": 1.0,
226
+ "use_bfloat16": false,
227
+ "vision_use_head": false
228
+ },
229
+ "vocab_size": 151674,
230
+ "vocab_start": null
231
+ }
dataset_statistics.json ADDED
@@ -0,0 +1,3698 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "fractal20220817_data_gresearch": {
3
+ "action": {
4
+ "mean": [
5
+ 0.5620291829109192,
6
+ -0.08018312603235245,
7
+ 0.7715217471122742,
8
+ -0.10781973600387573,
9
+ -0.6393505334854126,
10
+ -0.5443925261497498,
11
+ -0.8319592475891113,
12
+ 0.2779918909072876,
13
+ -0.12536543607711792,
14
+ 0.06098257377743721
15
+ ],
16
+ "std": [
17
+ 0.12430482357740402,
18
+ 0.11510306596755981,
19
+ 0.24659764766693115,
20
+ 0.3198048770427704,
21
+ 0.2988460063934326,
22
+ 0.30737143754959106,
23
+ 0.23584143817424774,
24
+ 0.3463784456253052,
25
+ 0.2272709161043167,
26
+ 1.022799015045166
27
+ ],
28
+ "max": [
29
+ 0.8761805367469786,
30
+ 1.2049139833450306,
31
+ 1.0974795579910281,
32
+ 1.0,
33
+ 0.9999049079418182,
34
+ 0.9987021696567535,
35
+ 0.9999175441265106,
36
+ 1.0,
37
+ 1.0,
38
+ 1.0001929998397827
39
+ ],
40
+ "min": [
41
+ -0.8384313821792603,
42
+ -0.6293186390399933,
43
+ -0.42908657133579253,
44
+ -0.9999992847442627,
45
+ -0.9996374249458313,
46
+ -0.9969306069612504,
47
+ -0.99928431391716,
48
+ -0.9999619847536088,
49
+ -0.9921636700630188,
50
+ -1.0002006542682649
51
+ ],
52
+ "q01": [
53
+ -0.8384313821792603,
54
+ -0.6293186390399933,
55
+ -0.42908657133579253,
56
+ -0.9999992847442627,
57
+ -0.9996374249458313,
58
+ -0.9969306069612504,
59
+ -0.99928431391716,
60
+ -0.9999619847536088,
61
+ -0.9921636700630188,
62
+ -1.0002006542682649
63
+ ],
64
+ "q99": [
65
+ 0.8761805367469786,
66
+ 1.2049139833450306,
67
+ 1.0974795579910281,
68
+ 1.0,
69
+ 0.9999049079418182,
70
+ 0.9987021696567535,
71
+ 0.9999175441265106,
72
+ 1.0,
73
+ 1.0,
74
+ 1.0001929998397827
75
+ ],
76
+ "mask": [
77
+ true,
78
+ true,
79
+ true,
80
+ false,
81
+ false,
82
+ false,
83
+ false,
84
+ false,
85
+ false,
86
+ false
87
+ ]
88
+ },
89
+ "proprio": {
90
+ "mean": [
91
+ 0.5593820214271545,
92
+ -0.08361946791410446,
93
+ 0.7761924266815186,
94
+ -0.11154364049434662,
95
+ -0.649243175983429,
96
+ -0.5352038145065308,
97
+ -0.8348413109779358,
98
+ 0.27783793210983276,
99
+ -0.12830153107643127,
100
+ -0.14084835350513458
101
+ ],
102
+ "std": [
103
+ 0.1243181824684143,
104
+ 0.11550454795360565,
105
+ 0.24582456052303314,
106
+ 0.31600186228752136,
107
+ 0.2989375591278076,
108
+ 0.3072265088558197,
109
+ 0.23316188156604767,
110
+ 0.34207573533058167,
111
+ 0.22458921372890472,
112
+ 0.9037747979164124
113
+ ],
114
+ "max": [
115
+ 0.8753991436958313,
116
+ 1.2016867852210993,
117
+ 1.0952961683273315,
118
+ 1.0,
119
+ 0.9999044907093048,
120
+ 0.9986884701251983,
121
+ 0.9999175071716309,
122
+ 1.0,
123
+ 1.0,
124
+ 1.0645089149475098
125
+ ],
126
+ "min": [
127
+ -0.8379243296384812,
128
+ -0.6292159473896026,
129
+ -0.4275088250637054,
130
+ -0.9999994039535522,
131
+ -0.999635414481163,
132
+ -0.996793516278267,
133
+ -0.99927858710289,
134
+ -0.9999648547172546,
135
+ -0.9921211671829223,
136
+ -1.0
137
+ ],
138
+ "q01": [
139
+ -0.8379243296384812,
140
+ -0.6292159473896026,
141
+ -0.4275088250637054,
142
+ -0.9999994039535522,
143
+ -0.999635414481163,
144
+ -0.996793516278267,
145
+ -0.99927858710289,
146
+ -0.9999648547172546,
147
+ -0.9921211671829223,
148
+ -1.0
149
+ ],
150
+ "q99": [
151
+ 0.8753991436958313,
152
+ 1.2016867852210993,
153
+ 1.0952961683273315,
154
+ 1.0,
155
+ 0.9999044907093048,
156
+ 0.9986884701251983,
157
+ 0.9999175071716309,
158
+ 1.0,
159
+ 1.0,
160
+ 1.0645089149475098
161
+ ],
162
+ "mask": [
163
+ true,
164
+ true,
165
+ true,
166
+ false,
167
+ false,
168
+ false,
169
+ false,
170
+ false,
171
+ false,
172
+ false
173
+ ]
174
+ },
175
+ "num_transitions": 3699188,
176
+ "num_trajectories": 87212
177
+ },
178
+ "kuka_filtered_gresearch": {
179
+ "action": {
180
+ "mean": [
181
+ 0.5516638159751892,
182
+ 0.04783713445067406,
183
+ 0.13569006323814392,
184
+ -0.9089635014533997,
185
+ -0.16706237196922302,
186
+ -0.006461427081376314,
187
+ -0.1670694798231125,
188
+ 0.9090182185173035,
189
+ 0.00019845239876303822,
190
+ -0.0972195416688919
191
+ ],
192
+ "std": [
193
+ 0.045198213309049606,
194
+ 0.10094085335731506,
195
+ 0.07423478364944458,
196
+ 0.1954759806394577,
197
+ 0.33851009607315063,
198
+ 0.010707848705351353,
199
+ 0.3384891748428345,
200
+ 0.19547049701213837,
201
+ 0.006141520570963621,
202
+ 1.0058562755584717
203
+ ],
204
+ "max": [
205
+ 0.8761805367469786,
206
+ 1.2049139833450306,
207
+ 1.0974795579910281,
208
+ 1.0,
209
+ 0.9999049079418182,
210
+ 0.9987021696567535,
211
+ 0.9999175441265106,
212
+ 1.0,
213
+ 1.0,
214
+ 1.0001929998397827
215
+ ],
216
+ "min": [
217
+ -0.8384313821792603,
218
+ -0.6293186390399933,
219
+ -0.42908657133579253,
220
+ -0.9999992847442627,
221
+ -0.9996374249458313,
222
+ -0.9969306069612504,
223
+ -0.99928431391716,
224
+ -0.9999619847536088,
225
+ -0.9921636700630188,
226
+ -1.0002006542682649
227
+ ],
228
+ "q01": [
229
+ -0.8384313821792603,
230
+ -0.6293186390399933,
231
+ -0.42908657133579253,
232
+ -0.9999992847442627,
233
+ -0.9996374249458313,
234
+ -0.9969306069612504,
235
+ -0.99928431391716,
236
+ -0.9999619847536088,
237
+ -0.9921636700630188,
238
+ -1.0002006542682649
239
+ ],
240
+ "q99": [
241
+ 0.8761805367469786,
242
+ 1.2049139833450306,
243
+ 1.0974795579910281,
244
+ 1.0,
245
+ 0.9999049079418182,
246
+ 0.9987021696567535,
247
+ 0.9999175441265106,
248
+ 1.0,
249
+ 1.0,
250
+ 1.0001929998397827
251
+ ],
252
+ "mask": [
253
+ true,
254
+ true,
255
+ true,
256
+ false,
257
+ false,
258
+ false,
259
+ false,
260
+ false,
261
+ false,
262
+ false
263
+ ]
264
+ },
265
+ "proprio": {
266
+ "mean": [
267
+ 0.5511404871940613,
268
+ 0.04786186292767525,
269
+ 0.1280936300754547,
270
+ -0.9253640174865723,
271
+ -0.13900955021381378,
272
+ -0.004956183955073357,
273
+ -0.13901649415493011,
274
+ 0.9254090785980225,
275
+ 0.00037780386628583074,
276
+ -0.0954216793179512
277
+ ],
278
+ "std": [
279
+ 0.04520808905363083,
280
+ 0.10223246365785599,
281
+ 0.05837938189506531,
282
+ 0.1767014116048813,
283
+ 0.3149489760398865,
284
+ 0.010550693608820438,
285
+ 0.31493160128593445,
286
+ 0.17669369280338287,
287
+ 0.0055999006144702435,
288
+ 1.0052746534347534
289
+ ],
290
+ "max": [
291
+ 0.8753991436958313,
292
+ 1.2016867852210993,
293
+ 1.0952961683273315,
294
+ 1.0,
295
+ 0.9999044907093048,
296
+ 0.9986884701251983,
297
+ 0.9999175071716309,
298
+ 1.0,
299
+ 1.0,
300
+ 1.0645089149475098
301
+ ],
302
+ "min": [
303
+ -0.8379243296384812,
304
+ -0.6292159473896026,
305
+ -0.4275088250637054,
306
+ -0.9999994039535522,
307
+ -0.999635414481163,
308
+ -0.996793516278267,
309
+ -0.99927858710289,
310
+ -0.9999648547172546,
311
+ -0.9921211671829223,
312
+ -1.0
313
+ ],
314
+ "q01": [
315
+ -0.8379243296384812,
316
+ -0.6292159473896026,
317
+ -0.4275088250637054,
318
+ -0.9999994039535522,
319
+ -0.999635414481163,
320
+ -0.996793516278267,
321
+ -0.99927858710289,
322
+ -0.9999648547172546,
323
+ -0.9921211671829223,
324
+ -1.0
325
+ ],
326
+ "q99": [
327
+ 0.8753991436958313,
328
+ 1.2016867852210993,
329
+ 1.0952961683273315,
330
+ 1.0,
331
+ 0.9999044907093048,
332
+ 0.9986884701251983,
333
+ 0.9999175071716309,
334
+ 1.0,
335
+ 1.0,
336
+ 1.0645089149475098
337
+ ],
338
+ "mask": [
339
+ true,
340
+ true,
341
+ true,
342
+ false,
343
+ false,
344
+ false,
345
+ false,
346
+ false,
347
+ false,
348
+ false
349
+ ]
350
+ },
351
+ "num_transitions": 2245999,
352
+ "num_trajectories": 209880
353
+ },
354
+ "bridge_dataset_gresearch": {
355
+ "action": {
356
+ "mean": [
357
+ 0.30980947613716125,
358
+ 0.030704185366630554,
359
+ 0.06348783522844315,
360
+ 0.8430948853492737,
361
+ 0.07754973322153091,
362
+ 0.07554265111684799,
363
+ -0.07994316518306732,
364
+ 0.8534128665924072,
365
+ 0.0064854128286242485,
366
+ 0.13305486738681793
367
+ ],
368
+ "std": [
369
+ 0.06045816093683243,
370
+ 0.09206030517816544,
371
+ 0.05132465064525604,
372
+ 0.2790003716945648,
373
+ 0.42283499240875244,
374
+ 0.16519200801849365,
375
+ 0.4229238033294678,
376
+ 0.2774372398853302,
377
+ 0.1247873306274414,
378
+ 0.9885094165802002
379
+ ],
380
+ "max": [
381
+ 0.8761805367469786,
382
+ 1.2049139833450306,
383
+ 1.0974795579910281,
384
+ 1.0,
385
+ 0.9999049079418182,
386
+ 0.9987021696567535,
387
+ 0.9999175441265106,
388
+ 1.0,
389
+ 1.0,
390
+ 1.0001929998397827
391
+ ],
392
+ "min": [
393
+ -0.8384313821792603,
394
+ -0.6293186390399933,
395
+ -0.42908657133579253,
396
+ -0.9999992847442627,
397
+ -0.9996374249458313,
398
+ -0.9969306069612504,
399
+ -0.99928431391716,
400
+ -0.9999619847536088,
401
+ -0.9921636700630188,
402
+ -1.0002006542682649
403
+ ],
404
+ "q01": [
405
+ -0.8384313821792603,
406
+ -0.6293186390399933,
407
+ -0.42908657133579253,
408
+ -0.9999992847442627,
409
+ -0.9996374249458313,
410
+ -0.9969306069612504,
411
+ -0.99928431391716,
412
+ -0.9999619847536088,
413
+ -0.9921636700630188,
414
+ -1.0002006542682649
415
+ ],
416
+ "q99": [
417
+ 0.8761805367469786,
418
+ 1.2049139833450306,
419
+ 1.0974795579910281,
420
+ 1.0,
421
+ 0.9999049079418182,
422
+ 0.9987021696567535,
423
+ 0.9999175441265106,
424
+ 1.0,
425
+ 1.0,
426
+ 1.0001929998397827
427
+ ],
428
+ "mask": [
429
+ true,
430
+ true,
431
+ true,
432
+ false,
433
+ false,
434
+ false,
435
+ false,
436
+ false,
437
+ false,
438
+ false
439
+ ]
440
+ },
441
+ "proprio": {
442
+ "mean": [
443
+ 0.3095255196094513,
444
+ 0.03058675304055214,
445
+ 0.06373011320829391,
446
+ 0.8441657423973083,
447
+ 0.07736347615718842,
448
+ 0.07511880993843079,
449
+ -0.07974543422460556,
450
+ 0.854377269744873,
451
+ 0.00668002525344491,
452
+ 0.2328014373779297
453
+ ],
454
+ "std": [
455
+ 0.06060030311346054,
456
+ 0.09206640720367432,
457
+ 0.051466673612594604,
458
+ 0.27826520800590515,
459
+ 0.4215569496154785,
460
+ 0.16456280648708344,
461
+ 0.42165014147758484,
462
+ 0.2767714560031891,
463
+ 0.12426231801509857,
464
+ 0.6662362813949585
465
+ ],
466
+ "max": [
467
+ 0.8753991436958313,
468
+ 1.2016867852210993,
469
+ 1.0952961683273315,
470
+ 1.0,
471
+ 0.9999044907093048,
472
+ 0.9986884701251983,
473
+ 0.9999175071716309,
474
+ 1.0,
475
+ 1.0,
476
+ 1.0645089149475098
477
+ ],
478
+ "min": [
479
+ -0.8379243296384812,
480
+ -0.6292159473896026,
481
+ -0.4275088250637054,
482
+ -0.9999994039535522,
483
+ -0.999635414481163,
484
+ -0.996793516278267,
485
+ -0.99927858710289,
486
+ -0.9999648547172546,
487
+ -0.9921211671829223,
488
+ -1.0
489
+ ],
490
+ "q01": [
491
+ -0.8379243296384812,
492
+ -0.6292159473896026,
493
+ -0.4275088250637054,
494
+ -0.9999994039535522,
495
+ -0.999635414481163,
496
+ -0.996793516278267,
497
+ -0.99927858710289,
498
+ -0.9999648547172546,
499
+ -0.9921211671829223,
500
+ -1.0
501
+ ],
502
+ "q99": [
503
+ 0.8753991436958313,
504
+ 1.2016867852210993,
505
+ 1.0952961683273315,
506
+ 1.0,
507
+ 0.9999044907093048,
508
+ 0.9986884701251983,
509
+ 0.9999175071716309,
510
+ 1.0,
511
+ 1.0,
512
+ 1.0645089149475098
513
+ ],
514
+ "mask": [
515
+ true,
516
+ true,
517
+ true,
518
+ false,
519
+ false,
520
+ false,
521
+ false,
522
+ false,
523
+ false,
524
+ false
525
+ ]
526
+ },
527
+ "num_transitions": 2075399,
528
+ "num_trajectories": 60064
529
+ },
530
+ "taco_play_gresearch": {
531
+ "action": {
532
+ "mean": [
533
+ 0.37266838550567627,
534
+ 0.13337592780590057,
535
+ 0.3828725516796112,
536
+ 0.683295726776123,
537
+ 0.3260023295879364,
538
+ 0.1700427234172821,
539
+ 0.3300667107105255,
540
+ -0.7032512426376343,
541
+ -0.02541291154921055,
542
+ -0.1530049741268158
543
+ ],
544
+ "std": [
545
+ 0.11484064906835556,
546
+ 0.26921647787094116,
547
+ 0.11261918395757675,
548
+ 0.3818320631980896,
549
+ 0.4699958264827728,
550
+ 0.1768525391817093,
551
+ 0.4743078947067261,
552
+ 0.3875681757926941,
553
+ 0.14425930380821228,
554
+ 0.9883240461349487
555
+ ],
556
+ "max": [
557
+ 0.8761805367469786,
558
+ 1.2049139833450306,
559
+ 1.0974795579910281,
560
+ 1.0,
561
+ 0.9999049079418182,
562
+ 0.9987021696567535,
563
+ 0.9999175441265106,
564
+ 1.0,
565
+ 1.0,
566
+ 1.0001929998397827
567
+ ],
568
+ "min": [
569
+ -0.8384313821792603,
570
+ -0.6293186390399933,
571
+ -0.42908657133579253,
572
+ -0.9999992847442627,
573
+ -0.9996374249458313,
574
+ -0.9969306069612504,
575
+ -0.99928431391716,
576
+ -0.9999619847536088,
577
+ -0.9921636700630188,
578
+ -1.0002006542682649
579
+ ],
580
+ "q01": [
581
+ -0.8384313821792603,
582
+ -0.6293186390399933,
583
+ -0.42908657133579253,
584
+ -0.9999992847442627,
585
+ -0.9996374249458313,
586
+ -0.9969306069612504,
587
+ -0.99928431391716,
588
+ -0.9999619847536088,
589
+ -0.9921636700630188,
590
+ -1.0002006542682649
591
+ ],
592
+ "q99": [
593
+ 0.8761805367469786,
594
+ 1.2049139833450306,
595
+ 1.0974795579910281,
596
+ 1.0,
597
+ 0.9999049079418182,
598
+ 0.9987021696567535,
599
+ 0.9999175441265106,
600
+ 1.0,
601
+ 1.0,
602
+ 1.0001929998397827
603
+ ],
604
+ "mask": [
605
+ true,
606
+ true,
607
+ true,
608
+ false,
609
+ false,
610
+ false,
611
+ false,
612
+ false,
613
+ false,
614
+ false
615
+ ]
616
+ },
617
+ "proprio": {
618
+ "mean": [
619
+ 0.3726738393306732,
620
+ 0.13310405611991882,
621
+ 0.3827139437198639,
622
+ 0.6834419369697571,
623
+ 0.3259745240211487,
624
+ 0.16991664469242096,
625
+ 0.33001208305358887,
626
+ -0.7033524513244629,
627
+ -0.025529393926262856,
628
+ 0.1655077338218689
629
+ ],
630
+ "std": [
631
+ 0.11486992239952087,
632
+ 0.2694161832332611,
633
+ 0.11262532323598862,
634
+ 0.38189199566841125,
635
+ 0.469816118478775,
636
+ 0.17681437730789185,
637
+ 0.4741390347480774,
638
+ 0.3876173496246338,
639
+ 0.14429394900798798,
640
+ 0.7380104064941406
641
+ ],
642
+ "max": [
643
+ 0.8753991436958313,
644
+ 1.2016867852210993,
645
+ 1.0952961683273315,
646
+ 1.0,
647
+ 0.9999044907093048,
648
+ 0.9986884701251983,
649
+ 0.9999175071716309,
650
+ 1.0,
651
+ 1.0,
652
+ 1.0645089149475098
653
+ ],
654
+ "min": [
655
+ -0.8379243296384812,
656
+ -0.6292159473896026,
657
+ -0.4275088250637054,
658
+ -0.9999994039535522,
659
+ -0.999635414481163,
660
+ -0.996793516278267,
661
+ -0.99927858710289,
662
+ -0.9999648547172546,
663
+ -0.9921211671829223,
664
+ -1.0
665
+ ],
666
+ "q01": [
667
+ -0.8379243296384812,
668
+ -0.6292159473896026,
669
+ -0.4275088250637054,
670
+ -0.9999994039535522,
671
+ -0.999635414481163,
672
+ -0.996793516278267,
673
+ -0.99927858710289,
674
+ -0.9999648547172546,
675
+ -0.9921211671829223,
676
+ -1.0
677
+ ],
678
+ "q99": [
679
+ 0.8753991436958313,
680
+ 1.2016867852210993,
681
+ 1.0952961683273315,
682
+ 1.0,
683
+ 0.9999044907093048,
684
+ 0.9986884701251983,
685
+ 0.9999175071716309,
686
+ 1.0,
687
+ 1.0,
688
+ 1.0645089149475098
689
+ ],
690
+ "mask": [
691
+ true,
692
+ true,
693
+ true,
694
+ false,
695
+ false,
696
+ false,
697
+ false,
698
+ false,
699
+ false,
700
+ false
701
+ ]
702
+ },
703
+ "num_transitions": 234195,
704
+ "num_trajectories": 3603
705
+ },
706
+ "jaco_play_gresearch": {
707
+ "action": {
708
+ "mean": [
709
+ -0.07594405114650726,
710
+ -0.4307216703891754,
711
+ 0.27456924319267273,
712
+ 1.0,
713
+ 0.0,
714
+ 0.0,
715
+ 0.0,
716
+ 0.0,
717
+ 1.0,
718
+ -0.30496877431869507
719
+ ],
720
+ "std": [
721
+ 0.14000719785690308,
722
+ 0.09076625853776932,
723
+ 0.052180223166942596,
724
+ 0.0,
725
+ 0.0,
726
+ 0.0,
727
+ 0.0,
728
+ 0.0,
729
+ 0.0,
730
+ 0.9525790214538574
731
+ ],
732
+ "max": [
733
+ 0.8761805367469786,
734
+ 1.2049139833450306,
735
+ 1.0974795579910281,
736
+ 1.0,
737
+ 0.9999049079418182,
738
+ 0.9987021696567535,
739
+ 0.9999175441265106,
740
+ 1.0,
741
+ 1.0,
742
+ 1.0001929998397827
743
+ ],
744
+ "min": [
745
+ -0.8384313821792603,
746
+ -0.6293186390399933,
747
+ -0.42908657133579253,
748
+ -0.9999992847442627,
749
+ -0.9996374249458313,
750
+ -0.9969306069612504,
751
+ -0.99928431391716,
752
+ -0.9999619847536088,
753
+ -0.9921636700630188,
754
+ -1.0002006542682649
755
+ ],
756
+ "q01": [
757
+ -0.8384313821792603,
758
+ -0.6293186390399933,
759
+ -0.42908657133579253,
760
+ -0.9999992847442627,
761
+ -0.9996374249458313,
762
+ -0.9969306069612504,
763
+ -0.99928431391716,
764
+ -0.9999619847536088,
765
+ -0.9921636700630188,
766
+ -1.0002006542682649
767
+ ],
768
+ "q99": [
769
+ 0.8761805367469786,
770
+ 1.2049139833450306,
771
+ 1.0974795579910281,
772
+ 1.0,
773
+ 0.9999049079418182,
774
+ 0.9987021696567535,
775
+ 0.9999175441265106,
776
+ 1.0,
777
+ 1.0,
778
+ 1.0001929998397827
779
+ ],
780
+ "mask": [
781
+ true,
782
+ true,
783
+ true,
784
+ false,
785
+ false,
786
+ false,
787
+ false,
788
+ false,
789
+ false,
790
+ false
791
+ ]
792
+ },
793
+ "proprio": {
794
+ "mean": [
795
+ -0.07597272843122482,
796
+ -0.43056127429008484,
797
+ 0.27473437786102295,
798
+ 1.0,
799
+ 0.0,
800
+ 0.0,
801
+ 0.0,
802
+ 0.0,
803
+ 1.0,
804
+ 0.033545784652233124
805
+ ],
806
+ "std": [
807
+ 0.13988833129405975,
808
+ 0.09080637991428375,
809
+ 0.05218466371297836,
810
+ 0.0,
811
+ 0.0,
812
+ 0.0,
813
+ 0.0,
814
+ 0.0,
815
+ 0.0,
816
+ 0.46983420848846436
817
+ ],
818
+ "max": [
819
+ 0.8753991436958313,
820
+ 1.2016867852210993,
821
+ 1.0952961683273315,
822
+ 1.0,
823
+ 0.9999044907093048,
824
+ 0.9986884701251983,
825
+ 0.9999175071716309,
826
+ 1.0,
827
+ 1.0,
828
+ 1.0645089149475098
829
+ ],
830
+ "min": [
831
+ -0.8379243296384812,
832
+ -0.6292159473896026,
833
+ -0.4275088250637054,
834
+ -0.9999994039535522,
835
+ -0.999635414481163,
836
+ -0.996793516278267,
837
+ -0.99927858710289,
838
+ -0.9999648547172546,
839
+ -0.9921211671829223,
840
+ -1.0
841
+ ],
842
+ "q01": [
843
+ -0.8379243296384812,
844
+ -0.6292159473896026,
845
+ -0.4275088250637054,
846
+ -0.9999994039535522,
847
+ -0.999635414481163,
848
+ -0.996793516278267,
849
+ -0.99927858710289,
850
+ -0.9999648547172546,
851
+ -0.9921211671829223,
852
+ -1.0
853
+ ],
854
+ "q99": [
855
+ 0.8753991436958313,
856
+ 1.2016867852210993,
857
+ 1.0952961683273315,
858
+ 1.0,
859
+ 0.9999044907093048,
860
+ 0.9986884701251983,
861
+ 0.9999175071716309,
862
+ 1.0,
863
+ 1.0,
864
+ 1.0645089149475098
865
+ ],
866
+ "mask": [
867
+ true,
868
+ true,
869
+ true,
870
+ false,
871
+ false,
872
+ false,
873
+ false,
874
+ false,
875
+ false,
876
+ false
877
+ ]
878
+ },
879
+ "num_transitions": 76880,
880
+ "num_trajectories": 1085
881
+ },
882
+ "viola_gresearch": {
883
+ "action": {
884
+ "mean": [
885
+ 0.5466106534004211,
886
+ 0.005938321817666292,
887
+ 0.23128610849380493,
888
+ 0.5923478007316589,
889
+ -0.4513634741306305,
890
+ 0.005500443279743195,
891
+ -0.4338614344596863,
892
+ -0.5633255243301392,
893
+ 0.15282785892486572,
894
+ 0.4637015163898468
895
+ ],
896
+ "std": [
897
+ 0.06075085327029228,
898
+ 0.12955749034881592,
899
+ 0.13219115138053894,
900
+ 0.4747803807258606,
901
+ 0.45891299843788147,
902
+ 0.09686105698347092,
903
+ 0.44180333614349365,
904
+ 0.46353647112846375,
905
+ 0.2470567524433136,
906
+ 0.8861218690872192
907
+ ],
908
+ "max": [
909
+ 0.8761805367469786,
910
+ 1.2049139833450306,
911
+ 1.0974795579910281,
912
+ 1.0,
913
+ 0.9999049079418182,
914
+ 0.9987021696567535,
915
+ 0.9999175441265106,
916
+ 1.0,
917
+ 1.0,
918
+ 1.0001929998397827
919
+ ],
920
+ "min": [
921
+ -0.8384313821792603,
922
+ -0.6293186390399933,
923
+ -0.42908657133579253,
924
+ -0.9999992847442627,
925
+ -0.9996374249458313,
926
+ -0.9969306069612504,
927
+ -0.99928431391716,
928
+ -0.9999619847536088,
929
+ -0.9921636700630188,
930
+ -1.0002006542682649
931
+ ],
932
+ "q01": [
933
+ -0.8384313821792603,
934
+ -0.6293186390399933,
935
+ -0.42908657133579253,
936
+ -0.9999992847442627,
937
+ -0.9996374249458313,
938
+ -0.9969306069612504,
939
+ -0.99928431391716,
940
+ -0.9999619847536088,
941
+ -0.9921636700630188,
942
+ -1.0002006542682649
943
+ ],
944
+ "q99": [
945
+ 0.8761805367469786,
946
+ 1.2049139833450306,
947
+ 1.0974795579910281,
948
+ 1.0,
949
+ 0.9999049079418182,
950
+ 0.9987021696567535,
951
+ 0.9999175441265106,
952
+ 1.0,
953
+ 1.0,
954
+ 1.0001929998397827
955
+ ],
956
+ "mask": [
957
+ true,
958
+ true,
959
+ true,
960
+ false,
961
+ false,
962
+ false,
963
+ false,
964
+ false,
965
+ false,
966
+ false
967
+ ]
968
+ },
969
+ "proprio": {
970
+ "mean": [
971
+ 0.5462588667869568,
972
+ 0.0062553794123232365,
973
+ 0.23125682771205902,
974
+ 0.5936321020126343,
975
+ -0.4500076472759247,
976
+ 0.005611719563603401,
977
+ -0.4325021803379059,
978
+ -0.5636327266693115,
979
+ 0.1527891457080841,
980
+ 0.3015170991420746
981
+ ],
982
+ "std": [
983
+ 0.06195013225078583,
984
+ 0.12935087084770203,
985
+ 0.1321941763162613,
986
+ 0.47453510761260986,
987
+ 0.45886990427970886,
988
+ 0.09670286625623703,
989
+ 0.4416916072368622,
990
+ 0.46456196904182434,
991
+ 0.24704748392105103,
992
+ 0.7458924055099487
993
+ ],
994
+ "max": [
995
+ 0.8753991436958313,
996
+ 1.2016867852210993,
997
+ 1.0952961683273315,
998
+ 1.0,
999
+ 0.9999044907093048,
1000
+ 0.9986884701251983,
1001
+ 0.9999175071716309,
1002
+ 1.0,
1003
+ 1.0,
1004
+ 1.0645089149475098
1005
+ ],
1006
+ "min": [
1007
+ -0.8379243296384812,
1008
+ -0.6292159473896026,
1009
+ -0.4275088250637054,
1010
+ -0.9999994039535522,
1011
+ -0.999635414481163,
1012
+ -0.996793516278267,
1013
+ -0.99927858710289,
1014
+ -0.9999648547172546,
1015
+ -0.9921211671829223,
1016
+ -1.0
1017
+ ],
1018
+ "q01": [
1019
+ -0.8379243296384812,
1020
+ -0.6292159473896026,
1021
+ -0.4275088250637054,
1022
+ -0.9999994039535522,
1023
+ -0.999635414481163,
1024
+ -0.996793516278267,
1025
+ -0.99927858710289,
1026
+ -0.9999648547172546,
1027
+ -0.9921211671829223,
1028
+ -1.0
1029
+ ],
1030
+ "q99": [
1031
+ 0.8753991436958313,
1032
+ 1.2016867852210993,
1033
+ 1.0952961683273315,
1034
+ 1.0,
1035
+ 0.9999044907093048,
1036
+ 0.9986884701251983,
1037
+ 0.9999175071716309,
1038
+ 1.0,
1039
+ 1.0,
1040
+ 1.0645089149475098
1041
+ ],
1042
+ "mask": [
1043
+ true,
1044
+ true,
1045
+ true,
1046
+ false,
1047
+ false,
1048
+ false,
1049
+ false,
1050
+ false,
1051
+ false,
1052
+ false
1053
+ ]
1054
+ },
1055
+ "num_transitions": 76174,
1056
+ "num_trajectories": 150
1057
+ },
1058
+ "berkeley_autolab_ur5_gresearch": {
1059
+ "action": {
1060
+ "mean": [
1061
+ 0.45475509762763977,
1062
+ 0.05735519900918007,
1063
+ -0.0355774350464344,
1064
+ -0.020543565973639488,
1065
+ 0.9803922176361084,
1066
+ -0.0030599436722695827,
1067
+ 0.9692681431770325,
1068
+ 0.020198913291096687,
1069
+ 0.008216173388063908,
1070
+ 0.25202447175979614
1071
+ ],
1072
+ "std": [
1073
+ 0.0786939263343811,
1074
+ 0.12533214688301086,
1075
+ 0.08071372658014297,
1076
+ 0.17048366367816925,
1077
+ 0.036416441202163696,
1078
+ 0.09088648855686188,
1079
+ 0.06161711737513542,
1080
+ 0.16556577384471893,
1081
+ 0.17069849371910095,
1082
+ 0.9679319858551025
1083
+ ],
1084
+ "max": [
1085
+ 0.8761805367469786,
1086
+ 1.2049139833450306,
1087
+ 1.0974795579910281,
1088
+ 1.0,
1089
+ 0.9999049079418182,
1090
+ 0.9987021696567535,
1091
+ 0.9999175441265106,
1092
+ 1.0,
1093
+ 1.0,
1094
+ 1.0001929998397827
1095
+ ],
1096
+ "min": [
1097
+ -0.8384313821792603,
1098
+ -0.6293186390399933,
1099
+ -0.42908657133579253,
1100
+ -0.9999992847442627,
1101
+ -0.9996374249458313,
1102
+ -0.9969306069612504,
1103
+ -0.99928431391716,
1104
+ -0.9999619847536088,
1105
+ -0.9921636700630188,
1106
+ -1.0002006542682649
1107
+ ],
1108
+ "q01": [
1109
+ -0.8384313821792603,
1110
+ -0.6293186390399933,
1111
+ -0.42908657133579253,
1112
+ -0.9999992847442627,
1113
+ -0.9996374249458313,
1114
+ -0.9969306069612504,
1115
+ -0.99928431391716,
1116
+ -0.9999619847536088,
1117
+ -0.9921636700630188,
1118
+ -1.0002006542682649
1119
+ ],
1120
+ "q99": [
1121
+ 0.8761805367469786,
1122
+ 1.2049139833450306,
1123
+ 1.0974795579910281,
1124
+ 1.0,
1125
+ 0.9999049079418182,
1126
+ 0.9987021696567535,
1127
+ 0.9999175441265106,
1128
+ 1.0,
1129
+ 1.0,
1130
+ 1.0001929998397827
1131
+ ],
1132
+ "mask": [
1133
+ true,
1134
+ true,
1135
+ true,
1136
+ false,
1137
+ false,
1138
+ false,
1139
+ false,
1140
+ false,
1141
+ false,
1142
+ false
1143
+ ]
1144
+ },
1145
+ "proprio": {
1146
+ "mean": [
1147
+ 0.45388421416282654,
1148
+ 0.05689476057887077,
1149
+ -0.03582395985722542,
1150
+ -0.02117013931274414,
1151
+ 0.9805601835250854,
1152
+ -0.0032330502290278673,
1153
+ 0.9695110321044922,
1154
+ 0.020909857004880905,
1155
+ 0.008249479345977306,
1156
+ -0.25177690386772156
1157
+ ],
1158
+ "std": [
1159
+ 0.07929588854312897,
1160
+ 0.12502582371234894,
1161
+ 0.08041266351938248,
1162
+ 0.16969986259937286,
1163
+ 0.0362262949347496,
1164
+ 0.09050362557172775,
1165
+ 0.061446838080883026,
1166
+ 0.16480131447315216,
1167
+ 0.17005135118961334,
1168
+ 0.9680400490760803
1169
+ ],
1170
+ "max": [
1171
+ 0.8753991436958313,
1172
+ 1.2016867852210993,
1173
+ 1.0952961683273315,
1174
+ 1.0,
1175
+ 0.9999044907093048,
1176
+ 0.9986884701251983,
1177
+ 0.9999175071716309,
1178
+ 1.0,
1179
+ 1.0,
1180
+ 1.0645089149475098
1181
+ ],
1182
+ "min": [
1183
+ -0.8379243296384812,
1184
+ -0.6292159473896026,
1185
+ -0.4275088250637054,
1186
+ -0.9999994039535522,
1187
+ -0.999635414481163,
1188
+ -0.996793516278267,
1189
+ -0.99927858710289,
1190
+ -0.9999648547172546,
1191
+ -0.9921211671829223,
1192
+ -1.0
1193
+ ],
1194
+ "q01": [
1195
+ -0.8379243296384812,
1196
+ -0.6292159473896026,
1197
+ -0.4275088250637054,
1198
+ -0.9999994039535522,
1199
+ -0.999635414481163,
1200
+ -0.996793516278267,
1201
+ -0.99927858710289,
1202
+ -0.9999648547172546,
1203
+ -0.9921211671829223,
1204
+ -1.0
1205
+ ],
1206
+ "q99": [
1207
+ 0.8753991436958313,
1208
+ 1.2016867852210993,
1209
+ 1.0952961683273315,
1210
+ 1.0,
1211
+ 0.9999044907093048,
1212
+ 0.9986884701251983,
1213
+ 0.9999175071716309,
1214
+ 1.0,
1215
+ 1.0,
1216
+ 1.0645089149475098
1217
+ ],
1218
+ "mask": [
1219
+ true,
1220
+ true,
1221
+ true,
1222
+ false,
1223
+ false,
1224
+ false,
1225
+ false,
1226
+ false,
1227
+ false,
1228
+ false
1229
+ ]
1230
+ },
1231
+ "num_transitions": 96939,
1232
+ "num_trajectories": 1000
1233
+ },
1234
+ "stanford_hydra_dataset_converted_externally_to_rlds_gresearch": {
1235
+ "action": {
1236
+ "mean": [
1237
+ 0.42426154017448425,
1238
+ -0.01393541507422924,
1239
+ 0.3763691782951355,
1240
+ 0.4628583490848541,
1241
+ -0.05695318803191185,
1242
+ 0.04738074913620949,
1243
+ -0.06379242241382599,
1244
+ -0.37509685754776,
1245
+ -0.20774757862091064,
1246
+ -0.0230943001806736
1247
+ ],
1248
+ "std": [
1249
+ 0.10466139763593674,
1250
+ 0.12751954793930054,
1251
+ 0.11140365898609161,
1252
+ 0.5973678231239319,
1253
+ 0.5371988415718079,
1254
+ 0.3671499490737915,
1255
+ 0.6135614514350891,
1256
+ 0.5570010542869568,
1257
+ 0.35400402545928955,
1258
+ 0.9995129108428955
1259
+ ],
1260
+ "max": [
1261
+ 0.8761805367469786,
1262
+ 1.2049139833450306,
1263
+ 1.0974795579910281,
1264
+ 1.0,
1265
+ 0.9999049079418182,
1266
+ 0.9987021696567535,
1267
+ 0.9999175441265106,
1268
+ 1.0,
1269
+ 1.0,
1270
+ 1.0001929998397827
1271
+ ],
1272
+ "min": [
1273
+ -0.8384313821792603,
1274
+ -0.6293186390399933,
1275
+ -0.42908657133579253,
1276
+ -0.9999992847442627,
1277
+ -0.9996374249458313,
1278
+ -0.9969306069612504,
1279
+ -0.99928431391716,
1280
+ -0.9999619847536088,
1281
+ -0.9921636700630188,
1282
+ -1.0002006542682649
1283
+ ],
1284
+ "q01": [
1285
+ -0.8384313821792603,
1286
+ -0.6293186390399933,
1287
+ -0.42908657133579253,
1288
+ -0.9999992847442627,
1289
+ -0.9996374249458313,
1290
+ -0.9969306069612504,
1291
+ -0.99928431391716,
1292
+ -0.9999619847536088,
1293
+ -0.9921636700630188,
1294
+ -1.0002006542682649
1295
+ ],
1296
+ "q99": [
1297
+ 0.8761805367469786,
1298
+ 1.2049139833450306,
1299
+ 1.0974795579910281,
1300
+ 1.0,
1301
+ 0.9999049079418182,
1302
+ 0.9987021696567535,
1303
+ 0.9999175441265106,
1304
+ 1.0,
1305
+ 1.0,
1306
+ 1.0001929998397827
1307
+ ],
1308
+ "mask": [
1309
+ true,
1310
+ true,
1311
+ true,
1312
+ false,
1313
+ false,
1314
+ false,
1315
+ false,
1316
+ false,
1317
+ false,
1318
+ false
1319
+ ]
1320
+ },
1321
+ "proprio": {
1322
+ "mean": [
1323
+ 0.4238300919532776,
1324
+ -0.013957408256828785,
1325
+ 0.3767646253108978,
1326
+ 0.4618605375289917,
1327
+ -0.05737191066145897,
1328
+ 0.04753109812736511,
1329
+ -0.0640297383069992,
1330
+ -0.37453097105026245,
1331
+ -0.20720084011554718,
1332
+ 0.2113121896982193
1333
+ ],
1334
+ "std": [
1335
+ 0.10451750457286835,
1336
+ 0.12747645378112793,
1337
+ 0.11160647869110107,
1338
+ 0.598028302192688,
1339
+ 0.5374307632446289,
1340
+ 0.3669152557849884,
1341
+ 0.6136991381645203,
1342
+ 0.5576587319374084,
1343
+ 0.3535982668399811,
1344
+ 0.7844396233558655
1345
+ ],
1346
+ "max": [
1347
+ 0.8753991436958313,
1348
+ 1.2016867852210993,
1349
+ 1.0952961683273315,
1350
+ 1.0,
1351
+ 0.9999044907093048,
1352
+ 0.9986884701251983,
1353
+ 0.9999175071716309,
1354
+ 1.0,
1355
+ 1.0,
1356
+ 1.0645089149475098
1357
+ ],
1358
+ "min": [
1359
+ -0.8379243296384812,
1360
+ -0.6292159473896026,
1361
+ -0.4275088250637054,
1362
+ -0.9999994039535522,
1363
+ -0.999635414481163,
1364
+ -0.996793516278267,
1365
+ -0.99927858710289,
1366
+ -0.9999648547172546,
1367
+ -0.9921211671829223,
1368
+ -1.0
1369
+ ],
1370
+ "q01": [
1371
+ -0.8379243296384812,
1372
+ -0.6292159473896026,
1373
+ -0.4275088250637054,
1374
+ -0.9999994039535522,
1375
+ -0.999635414481163,
1376
+ -0.996793516278267,
1377
+ -0.99927858710289,
1378
+ -0.9999648547172546,
1379
+ -0.9921211671829223,
1380
+ -1.0
1381
+ ],
1382
+ "q99": [
1383
+ 0.8753991436958313,
1384
+ 1.2016867852210993,
1385
+ 1.0952961683273315,
1386
+ 1.0,
1387
+ 0.9999044907093048,
1388
+ 0.9986884701251983,
1389
+ 0.9999175071716309,
1390
+ 1.0,
1391
+ 1.0,
1392
+ 1.0645089149475098
1393
+ ],
1394
+ "mask": [
1395
+ true,
1396
+ true,
1397
+ true,
1398
+ false,
1399
+ false,
1400
+ false,
1401
+ false,
1402
+ false,
1403
+ false,
1404
+ false
1405
+ ]
1406
+ },
1407
+ "num_transitions": 357664,
1408
+ "num_trajectories": 570
1409
+ },
1410
+ "austin_buds_dataset_converted_externally_to_rlds_gresearch": {
1411
+ "action": {
1412
+ "mean": [
1413
+ 0.5590491890907288,
1414
+ -0.002368063433095813,
1415
+ 0.12580959498882294,
1416
+ 0.9949961304664612,
1417
+ 0.011333368718624115,
1418
+ -0.027318278327584267,
1419
+ 0.010495728813111782,
1420
+ -0.9905137419700623,
1421
+ 0.0034327588509768248,
1422
+ -0.29980623722076416
1423
+ ],
1424
+ "std": [
1425
+ 0.08964027464389801,
1426
+ 0.14624646306037903,
1427
+ 0.07679233700037003,
1428
+ 0.00393949868157506,
1429
+ 0.04501992091536522,
1430
+ 0.0840773954987526,
1431
+ 0.04606208577752113,
1432
+ 0.1232115849852562,
1433
+ 0.03879213705658913,
1434
+ 0.9539803862571716
1435
+ ],
1436
+ "max": [
1437
+ 0.8761805367469786,
1438
+ 1.2049139833450306,
1439
+ 1.0974795579910281,
1440
+ 1.0,
1441
+ 0.9999049079418182,
1442
+ 0.9987021696567535,
1443
+ 0.9999175441265106,
1444
+ 1.0,
1445
+ 1.0,
1446
+ 1.0001929998397827
1447
+ ],
1448
+ "min": [
1449
+ -0.8384313821792603,
1450
+ -0.6293186390399933,
1451
+ -0.42908657133579253,
1452
+ -0.9999992847442627,
1453
+ -0.9996374249458313,
1454
+ -0.9969306069612504,
1455
+ -0.99928431391716,
1456
+ -0.9999619847536088,
1457
+ -0.9921636700630188,
1458
+ -1.0002006542682649
1459
+ ],
1460
+ "q01": [
1461
+ -0.8384313821792603,
1462
+ -0.6293186390399933,
1463
+ -0.42908657133579253,
1464
+ -0.9999992847442627,
1465
+ -0.9996374249458313,
1466
+ -0.9969306069612504,
1467
+ -0.99928431391716,
1468
+ -0.9999619847536088,
1469
+ -0.9921636700630188,
1470
+ -1.0002006542682649
1471
+ ],
1472
+ "q99": [
1473
+ 0.8761805367469786,
1474
+ 1.2049139833450306,
1475
+ 1.0974795579910281,
1476
+ 1.0,
1477
+ 0.9999049079418182,
1478
+ 0.9987021696567535,
1479
+ 0.9999175441265106,
1480
+ 1.0,
1481
+ 1.0,
1482
+ 1.0001929998397827
1483
+ ],
1484
+ "mask": [
1485
+ true,
1486
+ true,
1487
+ true,
1488
+ false,
1489
+ false,
1490
+ false,
1491
+ false,
1492
+ false,
1493
+ false,
1494
+ false
1495
+ ]
1496
+ },
1497
+ "proprio": {
1498
+ "mean": [
1499
+ 0.5588383078575134,
1500
+ -0.0022611774038523436,
1501
+ 0.12596501410007477,
1502
+ 0.9950018525123596,
1503
+ 0.011318260803818703,
1504
+ -0.027189675718545914,
1505
+ 0.010479051619768143,
1506
+ -0.9905157089233398,
1507
+ 0.003402459667995572,
1508
+ -0.03792320191860199
1509
+ ],
1510
+ "std": [
1511
+ 0.08969210833311081,
1512
+ 0.14624181389808655,
1513
+ 0.0769542008638382,
1514
+ 0.003941199276596308,
1515
+ 0.04499692842364311,
1516
+ 0.0840696468949318,
1517
+ 0.04603803902864456,
1518
+ 0.12321168929338455,
1519
+ 0.038784872740507126,
1520
+ 0.7493138313293457
1521
+ ],
1522
+ "max": [
1523
+ 0.8753991436958313,
1524
+ 1.2016867852210993,
1525
+ 1.0952961683273315,
1526
+ 1.0,
1527
+ 0.9999044907093048,
1528
+ 0.9986884701251983,
1529
+ 0.9999175071716309,
1530
+ 1.0,
1531
+ 1.0,
1532
+ 1.0645089149475098
1533
+ ],
1534
+ "min": [
1535
+ -0.8379243296384812,
1536
+ -0.6292159473896026,
1537
+ -0.4275088250637054,
1538
+ -0.9999994039535522,
1539
+ -0.999635414481163,
1540
+ -0.996793516278267,
1541
+ -0.99927858710289,
1542
+ -0.9999648547172546,
1543
+ -0.9921211671829223,
1544
+ -1.0
1545
+ ],
1546
+ "q01": [
1547
+ -0.8379243296384812,
1548
+ -0.6292159473896026,
1549
+ -0.4275088250637054,
1550
+ -0.9999994039535522,
1551
+ -0.999635414481163,
1552
+ -0.996793516278267,
1553
+ -0.99927858710289,
1554
+ -0.9999648547172546,
1555
+ -0.9921211671829223,
1556
+ -1.0
1557
+ ],
1558
+ "q99": [
1559
+ 0.8753991436958313,
1560
+ 1.2016867852210993,
1561
+ 1.0952961683273315,
1562
+ 1.0,
1563
+ 0.9999044907093048,
1564
+ 0.9986884701251983,
1565
+ 0.9999175071716309,
1566
+ 1.0,
1567
+ 1.0,
1568
+ 1.0645089149475098
1569
+ ],
1570
+ "mask": [
1571
+ true,
1572
+ true,
1573
+ true,
1574
+ false,
1575
+ false,
1576
+ false,
1577
+ false,
1578
+ false,
1579
+ false,
1580
+ false
1581
+ ]
1582
+ },
1583
+ "num_transitions": 34062,
1584
+ "num_trajectories": 50
1585
+ },
1586
+ "nyu_franka_play_dataset_converted_externally_to_rlds_gresearch": {
1587
+ "action": {
1588
+ "mean": [
1589
+ 0.3039938807487488,
1590
+ 0.415892630815506,
1591
+ 0.4642658531665802,
1592
+ 0.2071198672056198,
1593
+ 0.779800295829773,
1594
+ -0.022626692429184914,
1595
+ 0.07495637983083725,
1596
+ -0.04023102670907974,
1597
+ 0.7098985314369202,
1598
+ 0.44032958149909973
1599
+ ],
1600
+ "std": [
1601
+ 0.11779714375734329,
1602
+ 0.11434336751699448,
1603
+ 0.1649951934814453,
1604
+ 0.2833395302295685,
1605
+ 0.19798970222473145,
1606
+ 0.47856229543685913,
1607
+ 0.4094158113002777,
1608
+ 0.5040473937988281,
1609
+ 0.2590758800506592,
1610
+ 0.8976481556892395
1611
+ ],
1612
+ "max": [
1613
+ 0.8761805367469786,
1614
+ 1.2049139833450306,
1615
+ 1.0974795579910281,
1616
+ 1.0,
1617
+ 0.9999049079418182,
1618
+ 0.9987021696567535,
1619
+ 0.9999175441265106,
1620
+ 1.0,
1621
+ 1.0,
1622
+ 1.0001929998397827
1623
+ ],
1624
+ "min": [
1625
+ -0.8384313821792603,
1626
+ -0.6293186390399933,
1627
+ -0.42908657133579253,
1628
+ -0.9999992847442627,
1629
+ -0.9996374249458313,
1630
+ -0.9969306069612504,
1631
+ -0.99928431391716,
1632
+ -0.9999619847536088,
1633
+ -0.9921636700630188,
1634
+ -1.0002006542682649
1635
+ ],
1636
+ "q01": [
1637
+ -0.8384313821792603,
1638
+ -0.6293186390399933,
1639
+ -0.42908657133579253,
1640
+ -0.9999992847442627,
1641
+ -0.9996374249458313,
1642
+ -0.9969306069612504,
1643
+ -0.99928431391716,
1644
+ -0.9999619847536088,
1645
+ -0.9921636700630188,
1646
+ -1.0002006542682649
1647
+ ],
1648
+ "q99": [
1649
+ 0.8761805367469786,
1650
+ 1.2049139833450306,
1651
+ 1.0974795579910281,
1652
+ 1.0,
1653
+ 0.9999049079418182,
1654
+ 0.9987021696567535,
1655
+ 0.9999175441265106,
1656
+ 1.0,
1657
+ 1.0,
1658
+ 1.0001929998397827
1659
+ ],
1660
+ "mask": [
1661
+ true,
1662
+ true,
1663
+ true,
1664
+ false,
1665
+ false,
1666
+ false,
1667
+ false,
1668
+ false,
1669
+ false,
1670
+ false
1671
+ ]
1672
+ },
1673
+ "proprio": {
1674
+ "mean": [
1675
+ 0.30296146869659424,
1676
+ 0.41601383686065674,
1677
+ 0.4639338552951813,
1678
+ 0.20585064589977264,
1679
+ 0.779963493347168,
1680
+ -0.024647140875458717,
1681
+ 0.07426593452692032,
1682
+ -0.03759022429585457,
1683
+ 0.7105168104171753,
1684
+ 0.44032958149909973
1685
+ ],
1686
+ "std": [
1687
+ 0.11822924762964249,
1688
+ 0.11411069333553314,
1689
+ 0.16491974890232086,
1690
+ 0.28281328082084656,
1691
+ 0.197642520070076,
1692
+ 0.4791986644268036,
1693
+ 0.40874630212783813,
1694
+ 0.5044371485710144,
1695
+ 0.2582760751247406,
1696
+ 0.8976481556892395
1697
+ ],
1698
+ "max": [
1699
+ 0.8753991436958313,
1700
+ 1.2016867852210993,
1701
+ 1.0952961683273315,
1702
+ 1.0,
1703
+ 0.9999044907093048,
1704
+ 0.9986884701251983,
1705
+ 0.9999175071716309,
1706
+ 1.0,
1707
+ 1.0,
1708
+ 1.0645089149475098
1709
+ ],
1710
+ "min": [
1711
+ -0.8379243296384812,
1712
+ -0.6292159473896026,
1713
+ -0.4275088250637054,
1714
+ -0.9999994039535522,
1715
+ -0.999635414481163,
1716
+ -0.996793516278267,
1717
+ -0.99927858710289,
1718
+ -0.9999648547172546,
1719
+ -0.9921211671829223,
1720
+ -1.0
1721
+ ],
1722
+ "q01": [
1723
+ -0.8379243296384812,
1724
+ -0.6292159473896026,
1725
+ -0.4275088250637054,
1726
+ -0.9999994039535522,
1727
+ -0.999635414481163,
1728
+ -0.996793516278267,
1729
+ -0.99927858710289,
1730
+ -0.9999648547172546,
1731
+ -0.9921211671829223,
1732
+ -1.0
1733
+ ],
1734
+ "q99": [
1735
+ 0.8753991436958313,
1736
+ 1.2016867852210993,
1737
+ 1.0952961683273315,
1738
+ 1.0,
1739
+ 0.9999044907093048,
1740
+ 0.9986884701251983,
1741
+ 0.9999175071716309,
1742
+ 1.0,
1743
+ 1.0,
1744
+ 1.0645089149475098
1745
+ ],
1746
+ "mask": [
1747
+ true,
1748
+ true,
1749
+ true,
1750
+ false,
1751
+ false,
1752
+ false,
1753
+ false,
1754
+ false,
1755
+ false,
1756
+ false
1757
+ ]
1758
+ },
1759
+ "num_transitions": 44419,
1760
+ "num_trajectories": 456
1761
+ },
1762
+ "furniture_bench_dataset_converted_externally_to_rlds_gresearch": {
1763
+ "action": {
1764
+ "mean": [
1765
+ 0.5624285936355591,
1766
+ 0.0499618798494339,
1767
+ 0.08099052309989929,
1768
+ 0.6089181900024414,
1769
+ -0.03151005879044533,
1770
+ -0.022212982177734375,
1771
+ -0.03453856334090233,
1772
+ -0.6460461616516113,
1773
+ -0.04919625446200371,
1774
+ -0.023842008784413338
1775
+ ],
1776
+ "std": [
1777
+ 0.07705103605985641,
1778
+ 0.07908549904823303,
1779
+ 0.04069368913769722,
1780
+ 0.40301641821861267,
1781
+ 0.6237525939941406,
1782
+ 0.27449631690979004,
1783
+ 0.6290705800056458,
1784
+ 0.42472779750823975,
1785
+ 0.08097054064273834,
1786
+ 0.9997770190238953
1787
+ ],
1788
+ "max": [
1789
+ 0.8761805367469786,
1790
+ 1.2049139833450306,
1791
+ 1.0974795579910281,
1792
+ 1.0,
1793
+ 0.9999049079418182,
1794
+ 0.9987021696567535,
1795
+ 0.9999175441265106,
1796
+ 1.0,
1797
+ 1.0,
1798
+ 1.0001929998397827
1799
+ ],
1800
+ "min": [
1801
+ -0.8384313821792603,
1802
+ -0.6293186390399933,
1803
+ -0.42908657133579253,
1804
+ -0.9999992847442627,
1805
+ -0.9996374249458313,
1806
+ -0.9969306069612504,
1807
+ -0.99928431391716,
1808
+ -0.9999619847536088,
1809
+ -0.9921636700630188,
1810
+ -1.0002006542682649
1811
+ ],
1812
+ "q01": [
1813
+ -0.8384313821792603,
1814
+ -0.6293186390399933,
1815
+ -0.42908657133579253,
1816
+ -0.9999992847442627,
1817
+ -0.9996374249458313,
1818
+ -0.9969306069612504,
1819
+ -0.99928431391716,
1820
+ -0.9999619847536088,
1821
+ -0.9921636700630188,
1822
+ -1.0002006542682649
1823
+ ],
1824
+ "q99": [
1825
+ 0.8761805367469786,
1826
+ 1.2049139833450306,
1827
+ 1.0974795579910281,
1828
+ 1.0,
1829
+ 0.9999049079418182,
1830
+ 0.9987021696567535,
1831
+ 0.9999175441265106,
1832
+ 1.0,
1833
+ 1.0,
1834
+ 1.0001929998397827
1835
+ ],
1836
+ "mask": [
1837
+ true,
1838
+ true,
1839
+ true,
1840
+ false,
1841
+ false,
1842
+ false,
1843
+ false,
1844
+ false,
1845
+ false,
1846
+ false
1847
+ ]
1848
+ },
1849
+ "proprio": {
1850
+ "mean": [
1851
+ 0.5623790621757507,
1852
+ 0.04993896186351776,
1853
+ 0.08099812269210815,
1854
+ 0.6094604730606079,
1855
+ -0.029752522706985474,
1856
+ -0.022167107090353966,
1857
+ -0.032779522240161896,
1858
+ -0.6465969085693359,
1859
+ -0.049149055033922195,
1860
+ 0.14255355298519135
1861
+ ],
1862
+ "std": [
1863
+ 0.07702553272247314,
1864
+ 0.07905887067317963,
1865
+ 0.040680669248104095,
1866
+ 0.4023666977882385,
1867
+ 0.6235852837562561,
1868
+ 0.27445435523986816,
1869
+ 0.6287874579429626,
1870
+ 0.42406368255615234,
1871
+ 0.08097553998231888,
1872
+ 0.5384380221366882
1873
+ ],
1874
+ "max": [
1875
+ 0.8753991436958313,
1876
+ 1.2016867852210993,
1877
+ 1.0952961683273315,
1878
+ 1.0,
1879
+ 0.9999044907093048,
1880
+ 0.9986884701251983,
1881
+ 0.9999175071716309,
1882
+ 1.0,
1883
+ 1.0,
1884
+ 1.0645089149475098
1885
+ ],
1886
+ "min": [
1887
+ -0.8379243296384812,
1888
+ -0.6292159473896026,
1889
+ -0.4275088250637054,
1890
+ -0.9999994039535522,
1891
+ -0.999635414481163,
1892
+ -0.996793516278267,
1893
+ -0.99927858710289,
1894
+ -0.9999648547172546,
1895
+ -0.9921211671829223,
1896
+ -1.0
1897
+ ],
1898
+ "q01": [
1899
+ -0.8379243296384812,
1900
+ -0.6292159473896026,
1901
+ -0.4275088250637054,
1902
+ -0.9999994039535522,
1903
+ -0.999635414481163,
1904
+ -0.996793516278267,
1905
+ -0.99927858710289,
1906
+ -0.9999648547172546,
1907
+ -0.9921211671829223,
1908
+ -1.0
1909
+ ],
1910
+ "q99": [
1911
+ 0.8753991436958313,
1912
+ 1.2016867852210993,
1913
+ 1.0952961683273315,
1914
+ 1.0,
1915
+ 0.9999044907093048,
1916
+ 0.9986884701251983,
1917
+ 0.9999175071716309,
1918
+ 1.0,
1919
+ 1.0,
1920
+ 1.0645089149475098
1921
+ ],
1922
+ "mask": [
1923
+ true,
1924
+ true,
1925
+ true,
1926
+ false,
1927
+ false,
1928
+ false,
1929
+ false,
1930
+ false,
1931
+ false,
1932
+ false
1933
+ ]
1934
+ },
1935
+ "num_transitions": 3639794,
1936
+ "num_trajectories": 4744
1937
+ },
1938
+ "austin_sailor_dataset_converted_externally_to_rlds_gresearch": {
1939
+ "action": {
1940
+ "mean": [
1941
+ 0.5368196964263916,
1942
+ -0.08607572317123413,
1943
+ 0.1113031804561615,
1944
+ 0.8149315714836121,
1945
+ -0.03179684281349182,
1946
+ -0.013138960115611553,
1947
+ -0.031681980937719345,
1948
+ -0.8167121410369873,
1949
+ 0.00663051987066865,
1950
+ 0.051545400172472
1951
+ ],
1952
+ "std": [
1953
+ 0.08000274747610092,
1954
+ 0.13987071812152863,
1955
+ 0.056283172219991684,
1956
+ 0.22642314434051514,
1957
+ 0.5262997150421143,
1958
+ 0.08116751909255981,
1959
+ 0.5274008512496948,
1960
+ 0.2268417626619339,
1961
+ 0.05055464431643486,
1962
+ 0.9994269609451294
1963
+ ],
1964
+ "max": [
1965
+ 0.8761805367469786,
1966
+ 1.2049139833450306,
1967
+ 1.0974795579910281,
1968
+ 1.0,
1969
+ 0.9999049079418182,
1970
+ 0.9987021696567535,
1971
+ 0.9999175441265106,
1972
+ 1.0,
1973
+ 1.0,
1974
+ 1.0001929998397827
1975
+ ],
1976
+ "min": [
1977
+ -0.8384313821792603,
1978
+ -0.6293186390399933,
1979
+ -0.42908657133579253,
1980
+ -0.9999992847442627,
1981
+ -0.9996374249458313,
1982
+ -0.9969306069612504,
1983
+ -0.99928431391716,
1984
+ -0.9999619847536088,
1985
+ -0.9921636700630188,
1986
+ -1.0002006542682649
1987
+ ],
1988
+ "q01": [
1989
+ -0.8384313821792603,
1990
+ -0.6293186390399933,
1991
+ -0.42908657133579253,
1992
+ -0.9999992847442627,
1993
+ -0.9996374249458313,
1994
+ -0.9969306069612504,
1995
+ -0.99928431391716,
1996
+ -0.9999619847536088,
1997
+ -0.9921636700630188,
1998
+ -1.0002006542682649
1999
+ ],
2000
+ "q99": [
2001
+ 0.8761805367469786,
2002
+ 1.2049139833450306,
2003
+ 1.0974795579910281,
2004
+ 1.0,
2005
+ 0.9999049079418182,
2006
+ 0.9987021696567535,
2007
+ 0.9999175441265106,
2008
+ 1.0,
2009
+ 1.0,
2010
+ 1.0001929998397827
2011
+ ],
2012
+ "mask": [
2013
+ true,
2014
+ true,
2015
+ true,
2016
+ false,
2017
+ false,
2018
+ false,
2019
+ false,
2020
+ false,
2021
+ false,
2022
+ false
2023
+ ]
2024
+ },
2025
+ "proprio": {
2026
+ "mean": [
2027
+ 0.5367389917373657,
2028
+ -0.08604224026203156,
2029
+ 0.1113586351275444,
2030
+ 0.815004825592041,
2031
+ -0.03175605088472366,
2032
+ -0.013147763907909393,
2033
+ -0.03164118900895119,
2034
+ -0.8167852163314819,
2035
+ 0.006632022559642792,
2036
+ 0.3188406527042389
2037
+ ],
2038
+ "std": [
2039
+ 0.08000334352254868,
2040
+ 0.13984552025794983,
2041
+ 0.05637868493795395,
2042
+ 0.22643202543258667,
2043
+ 0.5261872410774231,
2044
+ 0.08115336298942566,
2045
+ 0.5272873640060425,
2046
+ 0.22685186564922333,
2047
+ 0.050542689859867096,
2048
+ 0.6499426364898682
2049
+ ],
2050
+ "max": [
2051
+ 0.8753991436958313,
2052
+ 1.2016867852210993,
2053
+ 1.0952961683273315,
2054
+ 1.0,
2055
+ 0.9999044907093048,
2056
+ 0.9986884701251983,
2057
+ 0.9999175071716309,
2058
+ 1.0,
2059
+ 1.0,
2060
+ 1.0645089149475098
2061
+ ],
2062
+ "min": [
2063
+ -0.8379243296384812,
2064
+ -0.6292159473896026,
2065
+ -0.4275088250637054,
2066
+ -0.9999994039535522,
2067
+ -0.999635414481163,
2068
+ -0.996793516278267,
2069
+ -0.99927858710289,
2070
+ -0.9999648547172546,
2071
+ -0.9921211671829223,
2072
+ -1.0
2073
+ ],
2074
+ "q01": [
2075
+ -0.8379243296384812,
2076
+ -0.6292159473896026,
2077
+ -0.4275088250637054,
2078
+ -0.9999994039535522,
2079
+ -0.999635414481163,
2080
+ -0.996793516278267,
2081
+ -0.99927858710289,
2082
+ -0.9999648547172546,
2083
+ -0.9921211671829223,
2084
+ -1.0
2085
+ ],
2086
+ "q99": [
2087
+ 0.8753991436958313,
2088
+ 1.2016867852210993,
2089
+ 1.0952961683273315,
2090
+ 1.0,
2091
+ 0.9999044907093048,
2092
+ 0.9986884701251983,
2093
+ 0.9999175071716309,
2094
+ 1.0,
2095
+ 1.0,
2096
+ 1.0645089149475098
2097
+ ],
2098
+ "mask": [
2099
+ true,
2100
+ true,
2101
+ true,
2102
+ false,
2103
+ false,
2104
+ false,
2105
+ false,
2106
+ false,
2107
+ false,
2108
+ false
2109
+ ]
2110
+ },
2111
+ "num_transitions": 352854,
2112
+ "num_trajectories": 240
2113
+ },
2114
+ "austin_sirius_dataset_converted_externally_to_rlds_gresearch": {
2115
+ "action": {
2116
+ "mean": [
2117
+ 0.4491329789161682,
2118
+ 0.094222791492939,
2119
+ 0.17133495211601257,
2120
+ 0.7470101714134216,
2121
+ -0.3794350326061249,
2122
+ 0.012982295826077461,
2123
+ -0.37987712025642395,
2124
+ -0.7126607298851013,
2125
+ -0.014358571730554104,
2126
+ -0.13706059753894806
2127
+ ],
2128
+ "std": [
2129
+ 0.0758960172533989,
2130
+ 0.08460639417171478,
2131
+ 0.047928906977176666,
2132
+ 0.35279470682144165,
2133
+ 0.41457077860832214,
2134
+ 0.04725702106952667,
2135
+ 0.4147632420063019,
2136
+ 0.41906556487083435,
2137
+ 0.015075068920850754,
2138
+ 0.9914436936378479
2139
+ ],
2140
+ "max": [
2141
+ 0.8761805367469786,
2142
+ 1.2049139833450306,
2143
+ 1.0974795579910281,
2144
+ 1.0,
2145
+ 0.9999049079418182,
2146
+ 0.9987021696567535,
2147
+ 0.9999175441265106,
2148
+ 1.0,
2149
+ 1.0,
2150
+ 1.0001929998397827
2151
+ ],
2152
+ "min": [
2153
+ -0.8384313821792603,
2154
+ -0.6293186390399933,
2155
+ -0.42908657133579253,
2156
+ -0.9999992847442627,
2157
+ -0.9996374249458313,
2158
+ -0.9969306069612504,
2159
+ -0.99928431391716,
2160
+ -0.9999619847536088,
2161
+ -0.9921636700630188,
2162
+ -1.0002006542682649
2163
+ ],
2164
+ "q01": [
2165
+ -0.8384313821792603,
2166
+ -0.6293186390399933,
2167
+ -0.42908657133579253,
2168
+ -0.9999992847442627,
2169
+ -0.9996374249458313,
2170
+ -0.9969306069612504,
2171
+ -0.99928431391716,
2172
+ -0.9999619847536088,
2173
+ -0.9921636700630188,
2174
+ -1.0002006542682649
2175
+ ],
2176
+ "q99": [
2177
+ 0.8761805367469786,
2178
+ 1.2049139833450306,
2179
+ 1.0974795579910281,
2180
+ 1.0,
2181
+ 0.9999049079418182,
2182
+ 0.9987021696567535,
2183
+ 0.9999175441265106,
2184
+ 1.0,
2185
+ 1.0,
2186
+ 1.0001929998397827
2187
+ ],
2188
+ "mask": [
2189
+ true,
2190
+ true,
2191
+ true,
2192
+ false,
2193
+ false,
2194
+ false,
2195
+ false,
2196
+ false,
2197
+ false,
2198
+ false
2199
+ ]
2200
+ },
2201
+ "proprio": {
2202
+ "mean": [
2203
+ 0.448244571685791,
2204
+ 0.09392176568508148,
2205
+ 0.17099416255950928,
2206
+ 0.7470197677612305,
2207
+ -0.37939515709877014,
2208
+ 0.012941493652760983,
2209
+ -0.37983736395835876,
2210
+ -0.7088543176651001,
2211
+ -0.014346957206726074,
2212
+ 0.5362510681152344
2213
+ ],
2214
+ "std": [
2215
+ 0.07837049663066864,
2216
+ 0.08452176302671432,
2217
+ 0.048490703105926514,
2218
+ 0.35280194878578186,
2219
+ 0.4145856201648712,
2220
+ 0.04725063219666481,
2221
+ 0.41477733850479126,
2222
+ 0.425743967294693,
2223
+ 0.015073223039507866,
2224
+ 0.4376317262649536
2225
+ ],
2226
+ "max": [
2227
+ 0.8753991436958313,
2228
+ 1.2016867852210993,
2229
+ 1.0952961683273315,
2230
+ 1.0,
2231
+ 0.9999044907093048,
2232
+ 0.9986884701251983,
2233
+ 0.9999175071716309,
2234
+ 1.0,
2235
+ 1.0,
2236
+ 1.0645089149475098
2237
+ ],
2238
+ "min": [
2239
+ -0.8379243296384812,
2240
+ -0.6292159473896026,
2241
+ -0.4275088250637054,
2242
+ -0.9999994039535522,
2243
+ -0.999635414481163,
2244
+ -0.996793516278267,
2245
+ -0.99927858710289,
2246
+ -0.9999648547172546,
2247
+ -0.9921211671829223,
2248
+ -1.0
2249
+ ],
2250
+ "q01": [
2251
+ -0.8379243296384812,
2252
+ -0.6292159473896026,
2253
+ -0.4275088250637054,
2254
+ -0.9999994039535522,
2255
+ -0.999635414481163,
2256
+ -0.996793516278267,
2257
+ -0.99927858710289,
2258
+ -0.9999648547172546,
2259
+ -0.9921211671829223,
2260
+ -1.0
2261
+ ],
2262
+ "q99": [
2263
+ 0.8753991436958313,
2264
+ 1.2016867852210993,
2265
+ 1.0952961683273315,
2266
+ 1.0,
2267
+ 0.9999044907093048,
2268
+ 0.9986884701251983,
2269
+ 0.9999175071716309,
2270
+ 1.0,
2271
+ 1.0,
2272
+ 1.0645089149475098
2273
+ ],
2274
+ "mask": [
2275
+ true,
2276
+ true,
2277
+ true,
2278
+ false,
2279
+ false,
2280
+ false,
2281
+ false,
2282
+ false,
2283
+ false,
2284
+ false
2285
+ ]
2286
+ },
2287
+ "num_transitions": 279380,
2288
+ "num_trajectories": 559
2289
+ },
2290
+ "dlr_edan_shared_control_converted_externally_to_rlds_gresearch": {
2291
+ "action": {
2292
+ "mean": [
2293
+ -0.5640354752540588,
2294
+ 0.4938611388206482,
2295
+ 0.10734671354293823,
2296
+ -0.384106308221817,
2297
+ 0.06716357916593552,
2298
+ 0.6988685131072998,
2299
+ 0.36692702770233154,
2300
+ -0.10909400880336761,
2301
+ 0.3807893395423889,
2302
+ 0.4134179651737213
2303
+ ],
2304
+ "std": [
2305
+ 0.17244893312454224,
2306
+ 0.14638741314411163,
2307
+ 0.3275325894355774,
2308
+ 0.3478024899959564,
2309
+ 0.2861262857913971,
2310
+ 0.39584824442863464,
2311
+ 0.6165788769721985,
2312
+ 0.37950125336647034,
2313
+ 0.4292682111263275,
2314
+ 0.910550594329834
2315
+ ],
2316
+ "max": [
2317
+ 0.8761805367469786,
2318
+ 1.2049139833450306,
2319
+ 1.0974795579910281,
2320
+ 1.0,
2321
+ 0.9999049079418182,
2322
+ 0.9987021696567535,
2323
+ 0.9999175441265106,
2324
+ 1.0,
2325
+ 1.0,
2326
+ 1.0001929998397827
2327
+ ],
2328
+ "min": [
2329
+ -0.8384313821792603,
2330
+ -0.6293186390399933,
2331
+ -0.42908657133579253,
2332
+ -0.9999992847442627,
2333
+ -0.9996374249458313,
2334
+ -0.9969306069612504,
2335
+ -0.99928431391716,
2336
+ -0.9999619847536088,
2337
+ -0.9921636700630188,
2338
+ -1.0002006542682649
2339
+ ],
2340
+ "q01": [
2341
+ -0.8384313821792603,
2342
+ -0.6293186390399933,
2343
+ -0.42908657133579253,
2344
+ -0.9999992847442627,
2345
+ -0.9996374249458313,
2346
+ -0.9969306069612504,
2347
+ -0.99928431391716,
2348
+ -0.9999619847536088,
2349
+ -0.9921636700630188,
2350
+ -1.0002006542682649
2351
+ ],
2352
+ "q99": [
2353
+ 0.8761805367469786,
2354
+ 1.2049139833450306,
2355
+ 1.0974795579910281,
2356
+ 1.0,
2357
+ 0.9999049079418182,
2358
+ 0.9987021696567535,
2359
+ 0.9999175441265106,
2360
+ 1.0,
2361
+ 1.0,
2362
+ 1.0001929998397827
2363
+ ],
2364
+ "mask": [
2365
+ true,
2366
+ true,
2367
+ true,
2368
+ false,
2369
+ false,
2370
+ false,
2371
+ false,
2372
+ false,
2373
+ false,
2374
+ false
2375
+ ]
2376
+ },
2377
+ "proprio": {
2378
+ "mean": [
2379
+ -0.5664793848991394,
2380
+ 0.49272674322128296,
2381
+ 0.1093696877360344,
2382
+ -0.3843309283256531,
2383
+ 0.060900215059518814,
2384
+ 0.7039747834205627,
2385
+ 0.3725402057170868,
2386
+ -0.10422468185424805,
2387
+ 0.37737715244293213,
2388
+ -0.4134179651737213
2389
+ ],
2390
+ "std": [
2391
+ 0.1705676019191742,
2392
+ 0.14644132554531097,
2393
+ 0.3261846899986267,
2394
+ 0.34728819131851196,
2395
+ 0.27886962890625,
2396
+ 0.39321795105934143,
2397
+ 0.6173756718635559,
2398
+ 0.3803780972957611,
2399
+ 0.4267268776893616,
2400
+ 0.910550594329834
2401
+ ],
2402
+ "max": [
2403
+ 0.8753991436958313,
2404
+ 1.2016867852210993,
2405
+ 1.0952961683273315,
2406
+ 1.0,
2407
+ 0.9999044907093048,
2408
+ 0.9986884701251983,
2409
+ 0.9999175071716309,
2410
+ 1.0,
2411
+ 1.0,
2412
+ 1.0645089149475098
2413
+ ],
2414
+ "min": [
2415
+ -0.8379243296384812,
2416
+ -0.6292159473896026,
2417
+ -0.4275088250637054,
2418
+ -0.9999994039535522,
2419
+ -0.999635414481163,
2420
+ -0.996793516278267,
2421
+ -0.99927858710289,
2422
+ -0.9999648547172546,
2423
+ -0.9921211671829223,
2424
+ -1.0
2425
+ ],
2426
+ "q01": [
2427
+ -0.8379243296384812,
2428
+ -0.6292159473896026,
2429
+ -0.4275088250637054,
2430
+ -0.9999994039535522,
2431
+ -0.999635414481163,
2432
+ -0.996793516278267,
2433
+ -0.99927858710289,
2434
+ -0.9999648547172546,
2435
+ -0.9921211671829223,
2436
+ -1.0
2437
+ ],
2438
+ "q99": [
2439
+ 0.8753991436958313,
2440
+ 1.2016867852210993,
2441
+ 1.0952961683273315,
2442
+ 1.0,
2443
+ 0.9999044907093048,
2444
+ 0.9986884701251983,
2445
+ 0.9999175071716309,
2446
+ 1.0,
2447
+ 1.0,
2448
+ 1.0645089149475098
2449
+ ],
2450
+ "mask": [
2451
+ true,
2452
+ true,
2453
+ true,
2454
+ false,
2455
+ false,
2456
+ false,
2457
+ false,
2458
+ false,
2459
+ false,
2460
+ false
2461
+ ]
2462
+ },
2463
+ "num_transitions": 8824,
2464
+ "num_trajectories": 104
2465
+ },
2466
+ "utaustin_mutex_gresearch": {
2467
+ "action": {
2468
+ "mean": [
2469
+ 0.4329838454723358,
2470
+ -0.10120747238397598,
2471
+ 0.22025811672210693,
2472
+ 0.8212123513221741,
2473
+ -0.21459075808525085,
2474
+ -0.0665178969502449,
2475
+ -0.24166706204414368,
2476
+ -0.6632139086723328,
2477
+ -0.30717793107032776,
2478
+ -0.003665544791147113
2479
+ ],
2480
+ "std": [
2481
+ 0.043886806815862656,
2482
+ 0.20657485723495483,
2483
+ 0.12638156116008759,
2484
+ 0.2851228713989258,
2485
+ 0.42492440342903137,
2486
+ 0.1168530210852623,
2487
+ 0.38834014534950256,
2488
+ 0.3347749710083008,
2489
+ 0.38038182258605957,
2490
+ 1.0000113248825073
2491
+ ],
2492
+ "max": [
2493
+ 0.8761805367469786,
2494
+ 1.2049139833450306,
2495
+ 1.0974795579910281,
2496
+ 1.0,
2497
+ 0.9999049079418182,
2498
+ 0.9987021696567535,
2499
+ 0.9999175441265106,
2500
+ 1.0,
2501
+ 1.0,
2502
+ 1.0001929998397827
2503
+ ],
2504
+ "min": [
2505
+ -0.8384313821792603,
2506
+ -0.6293186390399933,
2507
+ -0.42908657133579253,
2508
+ -0.9999992847442627,
2509
+ -0.9996374249458313,
2510
+ -0.9969306069612504,
2511
+ -0.99928431391716,
2512
+ -0.9999619847536088,
2513
+ -0.9921636700630188,
2514
+ -1.0002006542682649
2515
+ ],
2516
+ "q01": [
2517
+ -0.8384313821792603,
2518
+ -0.6293186390399933,
2519
+ -0.42908657133579253,
2520
+ -0.9999992847442627,
2521
+ -0.9996374249458313,
2522
+ -0.9969306069612504,
2523
+ -0.99928431391716,
2524
+ -0.9999619847536088,
2525
+ -0.9921636700630188,
2526
+ -1.0002006542682649
2527
+ ],
2528
+ "q99": [
2529
+ 0.8761805367469786,
2530
+ 1.2049139833450306,
2531
+ 1.0974795579910281,
2532
+ 1.0,
2533
+ 0.9999049079418182,
2534
+ 0.9987021696567535,
2535
+ 0.9999175441265106,
2536
+ 1.0,
2537
+ 1.0,
2538
+ 1.0001929998397827
2539
+ ],
2540
+ "mask": [
2541
+ true,
2542
+ true,
2543
+ true,
2544
+ false,
2545
+ false,
2546
+ false,
2547
+ false,
2548
+ false,
2549
+ false,
2550
+ false
2551
+ ]
2552
+ },
2553
+ "proprio": {
2554
+ "mean": [
2555
+ 0.4331544041633606,
2556
+ -0.10096580535173416,
2557
+ 0.2204408496618271,
2558
+ 0.8224666714668274,
2559
+ -0.21331606805324554,
2560
+ -0.06597500294446945,
2561
+ -0.24025921523571014,
2562
+ -0.6650930047035217,
2563
+ -0.3062214255332947,
2564
+ 0.2226639986038208
2565
+ ],
2566
+ "std": [
2567
+ 0.043805938214063644,
2568
+ 0.20559944212436676,
2569
+ 0.12614838778972626,
2570
+ 0.2843463718891144,
2571
+ 0.4238165318965912,
2572
+ 0.1166481301188469,
2573
+ 0.387310653924942,
2574
+ 0.33454954624176025,
2575
+ 0.3800017535686493,
2576
+ 0.8144067525863647
2577
+ ],
2578
+ "max": [
2579
+ 0.8753991436958313,
2580
+ 1.2016867852210993,
2581
+ 1.0952961683273315,
2582
+ 1.0,
2583
+ 0.9999044907093048,
2584
+ 0.9986884701251983,
2585
+ 0.9999175071716309,
2586
+ 1.0,
2587
+ 1.0,
2588
+ 1.0645089149475098
2589
+ ],
2590
+ "min": [
2591
+ -0.8379243296384812,
2592
+ -0.6292159473896026,
2593
+ -0.4275088250637054,
2594
+ -0.9999994039535522,
2595
+ -0.999635414481163,
2596
+ -0.996793516278267,
2597
+ -0.99927858710289,
2598
+ -0.9999648547172546,
2599
+ -0.9921211671829223,
2600
+ -1.0
2601
+ ],
2602
+ "q01": [
2603
+ -0.8379243296384812,
2604
+ -0.6292159473896026,
2605
+ -0.4275088250637054,
2606
+ -0.9999994039535522,
2607
+ -0.999635414481163,
2608
+ -0.996793516278267,
2609
+ -0.99927858710289,
2610
+ -0.9999648547172546,
2611
+ -0.9921211671829223,
2612
+ -1.0
2613
+ ],
2614
+ "q99": [
2615
+ 0.8753991436958313,
2616
+ 1.2016867852210993,
2617
+ 1.0952961683273315,
2618
+ 1.0,
2619
+ 0.9999044907093048,
2620
+ 0.9986884701251983,
2621
+ 0.9999175071716309,
2622
+ 1.0,
2623
+ 1.0,
2624
+ 1.0645089149475098
2625
+ ],
2626
+ "mask": [
2627
+ true,
2628
+ true,
2629
+ true,
2630
+ false,
2631
+ false,
2632
+ false,
2633
+ false,
2634
+ false,
2635
+ false,
2636
+ false
2637
+ ]
2638
+ },
2639
+ "num_transitions": 360383,
2640
+ "num_trajectories": 1500
2641
+ },
2642
+ "berkeley_fanuc_manipulation_gresearch": {
2643
+ "action": {
2644
+ "mean": [
2645
+ 0.5453283786773682,
2646
+ 0.003412138670682907,
2647
+ 0.2424720674753189,
2648
+ 0.8885376453399658,
2649
+ 0.02352900244295597,
2650
+ 0.028720814734697342,
2651
+ 0.02244826778769493,
2652
+ -0.8494833111763,
2653
+ 0.07993770390748978,
2654
+ 0.39541465044021606
2655
+ ],
2656
+ "std": [
2657
+ 0.07659938931465149,
2658
+ 0.1454569697380066,
2659
+ 0.13793079555034637,
2660
+ 0.25514063239097595,
2661
+ 0.3272935152053833,
2662
+ 0.1920861005783081,
2663
+ 0.3559061586856842,
2664
+ 0.27330872416496277,
2665
+ 0.2648014426231384,
2666
+ 0.9186042547225952
2667
+ ],
2668
+ "max": [
2669
+ 0.8761805367469786,
2670
+ 1.2049139833450306,
2671
+ 1.0974795579910281,
2672
+ 1.0,
2673
+ 0.9999049079418182,
2674
+ 0.9987021696567535,
2675
+ 0.9999175441265106,
2676
+ 1.0,
2677
+ 1.0,
2678
+ 1.0001929998397827
2679
+ ],
2680
+ "min": [
2681
+ -0.8384313821792603,
2682
+ -0.6293186390399933,
2683
+ -0.42908657133579253,
2684
+ -0.9999992847442627,
2685
+ -0.9996374249458313,
2686
+ -0.9969306069612504,
2687
+ -0.99928431391716,
2688
+ -0.9999619847536088,
2689
+ -0.9921636700630188,
2690
+ -1.0002006542682649
2691
+ ],
2692
+ "q01": [
2693
+ -0.8384313821792603,
2694
+ -0.6293186390399933,
2695
+ -0.42908657133579253,
2696
+ -0.9999992847442627,
2697
+ -0.9996374249458313,
2698
+ -0.9969306069612504,
2699
+ -0.99928431391716,
2700
+ -0.9999619847536088,
2701
+ -0.9921636700630188,
2702
+ -1.0002006542682649
2703
+ ],
2704
+ "q99": [
2705
+ 0.8761805367469786,
2706
+ 1.2049139833450306,
2707
+ 1.0974795579910281,
2708
+ 1.0,
2709
+ 0.9999049079418182,
2710
+ 0.9987021696567535,
2711
+ 0.9999175441265106,
2712
+ 1.0,
2713
+ 1.0,
2714
+ 1.0001929998397827
2715
+ ],
2716
+ "mask": [
2717
+ true,
2718
+ true,
2719
+ true,
2720
+ false,
2721
+ false,
2722
+ false,
2723
+ false,
2724
+ false,
2725
+ false,
2726
+ false
2727
+ ]
2728
+ },
2729
+ "proprio": {
2730
+ "mean": [
2731
+ 0.5446352958679199,
2732
+ 0.0037373844534158707,
2733
+ 0.24398808181285858,
2734
+ 0.8888180255889893,
2735
+ 0.02343420498073101,
2736
+ 0.028684617951512337,
2737
+ 0.02236282266676426,
2738
+ -0.8502753973007202,
2739
+ 0.0793125182390213,
2740
+ 0.39541465044021606
2741
+ ],
2742
+ "std": [
2743
+ 0.07649148255586624,
2744
+ 0.14486142992973328,
2745
+ 0.13899101316928864,
2746
+ 0.25504106283187866,
2747
+ 0.3266383707523346,
2748
+ 0.1920391470193863,
2749
+ 0.355115681886673,
2750
+ 0.27290859818458557,
2751
+ 0.263923317193985,
2752
+ 0.9186042547225952
2753
+ ],
2754
+ "max": [
2755
+ 0.8753991436958313,
2756
+ 1.2016867852210993,
2757
+ 1.0952961683273315,
2758
+ 1.0,
2759
+ 0.9999044907093048,
2760
+ 0.9986884701251983,
2761
+ 0.9999175071716309,
2762
+ 1.0,
2763
+ 1.0,
2764
+ 1.0645089149475098
2765
+ ],
2766
+ "min": [
2767
+ -0.8379243296384812,
2768
+ -0.6292159473896026,
2769
+ -0.4275088250637054,
2770
+ -0.9999994039535522,
2771
+ -0.999635414481163,
2772
+ -0.996793516278267,
2773
+ -0.99927858710289,
2774
+ -0.9999648547172546,
2775
+ -0.9921211671829223,
2776
+ -1.0
2777
+ ],
2778
+ "q01": [
2779
+ -0.8379243296384812,
2780
+ -0.6292159473896026,
2781
+ -0.4275088250637054,
2782
+ -0.9999994039535522,
2783
+ -0.999635414481163,
2784
+ -0.996793516278267,
2785
+ -0.99927858710289,
2786
+ -0.9999648547172546,
2787
+ -0.9921211671829223,
2788
+ -1.0
2789
+ ],
2790
+ "q99": [
2791
+ 0.8753991436958313,
2792
+ 1.2016867852210993,
2793
+ 1.0952961683273315,
2794
+ 1.0,
2795
+ 0.9999044907093048,
2796
+ 0.9986884701251983,
2797
+ 0.9999175071716309,
2798
+ 1.0,
2799
+ 1.0,
2800
+ 1.0645089149475098
2801
+ ],
2802
+ "mask": [
2803
+ true,
2804
+ true,
2805
+ true,
2806
+ false,
2807
+ false,
2808
+ false,
2809
+ false,
2810
+ false,
2811
+ false,
2812
+ false
2813
+ ]
2814
+ },
2815
+ "num_transitions": 62198,
2816
+ "num_trajectories": 415
2817
+ },
2818
+ "cmu_stretch_gresearch": {
2819
+ "action": {
2820
+ "mean": [
2821
+ 0.1465548425912857,
2822
+ 0.0,
2823
+ 0.7584614753723145,
2824
+ 1.0,
2825
+ 0.0,
2826
+ 0.0,
2827
+ 0.0,
2828
+ 1.0,
2829
+ 0.0,
2830
+ -0.1982637345790863
2831
+ ],
2832
+ "std": [
2833
+ 0.0994541123509407,
2834
+ 0.0,
2835
+ 0.15765506029129028,
2836
+ 0.0,
2837
+ 0.0,
2838
+ 0.0,
2839
+ 0.0,
2840
+ 0.0,
2841
+ 0.0,
2842
+ 0.9800659418106079
2843
+ ],
2844
+ "max": [
2845
+ 0.8761805367469786,
2846
+ 1.2049139833450306,
2847
+ 1.0974795579910281,
2848
+ 1.0,
2849
+ 0.9999049079418182,
2850
+ 0.9987021696567535,
2851
+ 0.9999175441265106,
2852
+ 1.0,
2853
+ 1.0,
2854
+ 1.0001929998397827
2855
+ ],
2856
+ "min": [
2857
+ -0.8384313821792603,
2858
+ -0.6293186390399933,
2859
+ -0.42908657133579253,
2860
+ -0.9999992847442627,
2861
+ -0.9996374249458313,
2862
+ -0.9969306069612504,
2863
+ -0.99928431391716,
2864
+ -0.9999619847536088,
2865
+ -0.9921636700630188,
2866
+ -1.0002006542682649
2867
+ ],
2868
+ "q01": [
2869
+ -0.8384313821792603,
2870
+ -0.6293186390399933,
2871
+ -0.42908657133579253,
2872
+ -0.9999992847442627,
2873
+ -0.9996374249458313,
2874
+ -0.9969306069612504,
2875
+ -0.99928431391716,
2876
+ -0.9999619847536088,
2877
+ -0.9921636700630188,
2878
+ -1.0002006542682649
2879
+ ],
2880
+ "q99": [
2881
+ 0.8761805367469786,
2882
+ 1.2049139833450306,
2883
+ 1.0974795579910281,
2884
+ 1.0,
2885
+ 0.9999049079418182,
2886
+ 0.9987021696567535,
2887
+ 0.9999175441265106,
2888
+ 1.0,
2889
+ 1.0,
2890
+ 1.0001929998397827
2891
+ ],
2892
+ "mask": [
2893
+ true,
2894
+ true,
2895
+ true,
2896
+ false,
2897
+ false,
2898
+ false,
2899
+ false,
2900
+ false,
2901
+ false,
2902
+ false
2903
+ ]
2904
+ },
2905
+ "proprio": {
2906
+ "mean": [
2907
+ 0.14618979394435883,
2908
+ 0.0,
2909
+ 0.7568053007125854,
2910
+ 1.0,
2911
+ 0.0,
2912
+ 0.0,
2913
+ 0.0,
2914
+ 1.0,
2915
+ 0.0,
2916
+ -0.17334973812103271
2917
+ ],
2918
+ "std": [
2919
+ 0.09959966689348221,
2920
+ 0.0,
2921
+ 0.15819641947746277,
2922
+ 0.0,
2923
+ 0.0,
2924
+ 0.0,
2925
+ 0.0,
2926
+ 0.0,
2927
+ 0.0,
2928
+ 0.9756740927696228
2929
+ ],
2930
+ "max": [
2931
+ 0.8753991436958313,
2932
+ 1.2016867852210993,
2933
+ 1.0952961683273315,
2934
+ 1.0,
2935
+ 0.9999044907093048,
2936
+ 0.9986884701251983,
2937
+ 0.9999175071716309,
2938
+ 1.0,
2939
+ 1.0,
2940
+ 1.0645089149475098
2941
+ ],
2942
+ "min": [
2943
+ -0.8379243296384812,
2944
+ -0.6292159473896026,
2945
+ -0.4275088250637054,
2946
+ -0.9999994039535522,
2947
+ -0.999635414481163,
2948
+ -0.996793516278267,
2949
+ -0.99927858710289,
2950
+ -0.9999648547172546,
2951
+ -0.9921211671829223,
2952
+ -1.0
2953
+ ],
2954
+ "q01": [
2955
+ -0.8379243296384812,
2956
+ -0.6292159473896026,
2957
+ -0.4275088250637054,
2958
+ -0.9999994039535522,
2959
+ -0.999635414481163,
2960
+ -0.996793516278267,
2961
+ -0.99927858710289,
2962
+ -0.9999648547172546,
2963
+ -0.9921211671829223,
2964
+ -1.0
2965
+ ],
2966
+ "q99": [
2967
+ 0.8753991436958313,
2968
+ 1.2016867852210993,
2969
+ 1.0952961683273315,
2970
+ 1.0,
2971
+ 0.9999044907093048,
2972
+ 0.9986884701251983,
2973
+ 0.9999175071716309,
2974
+ 1.0,
2975
+ 1.0,
2976
+ 1.0645089149475098
2977
+ ],
2978
+ "mask": [
2979
+ true,
2980
+ true,
2981
+ true,
2982
+ false,
2983
+ false,
2984
+ false,
2985
+ false,
2986
+ false,
2987
+ false,
2988
+ false
2989
+ ]
2990
+ },
2991
+ "num_transitions": 24881,
2992
+ "num_trajectories": 135
2993
+ },
2994
+ "bc_z_filtered_gresearch": {
2995
+ "action": {
2996
+ "mean": [
2997
+ 0.01686379313468933,
2998
+ 0.10985267162322998,
2999
+ 0.7889096140861511,
3000
+ 0.5185964703559875,
3001
+ -0.08276072144508362,
3002
+ 0.173233762383461,
3003
+ 0.16470178961753845,
3004
+ 0.4580680727958679,
3005
+ -0.2728700339794159,
3006
+ 0.6686875224113464
3007
+ ],
3008
+ "std": [
3009
+ 0.1836080700159073,
3010
+ 0.09609317779541016,
3011
+ 0.08680785447359085,
3012
+ 0.46468544006347656,
3013
+ 0.5311728715896606,
3014
+ 0.44270169734954834,
3015
+ 0.6481146812438965,
3016
+ 0.39129871129989624,
3017
+ 0.3275664746761322,
3018
+ 0.7226802706718445
3019
+ ],
3020
+ "max": [
3021
+ 0.8761805367469786,
3022
+ 1.2049139833450306,
3023
+ 1.0974795579910281,
3024
+ 1.0,
3025
+ 0.9999049079418182,
3026
+ 0.9987021696567535,
3027
+ 0.9999175441265106,
3028
+ 1.0,
3029
+ 1.0,
3030
+ 1.0001929998397827
3031
+ ],
3032
+ "min": [
3033
+ -0.8384313821792603,
3034
+ -0.6293186390399933,
3035
+ -0.42908657133579253,
3036
+ -0.9999992847442627,
3037
+ -0.9996374249458313,
3038
+ -0.9969306069612504,
3039
+ -0.99928431391716,
3040
+ -0.9999619847536088,
3041
+ -0.9921636700630188,
3042
+ -1.0002006542682649
3043
+ ],
3044
+ "q01": [
3045
+ -0.8384313821792603,
3046
+ -0.6293186390399933,
3047
+ -0.42908657133579253,
3048
+ -0.9999992847442627,
3049
+ -0.9996374249458313,
3050
+ -0.9969306069612504,
3051
+ -0.99928431391716,
3052
+ -0.9999619847536088,
3053
+ -0.9921636700630188,
3054
+ -1.0002006542682649
3055
+ ],
3056
+ "q99": [
3057
+ 0.8761805367469786,
3058
+ 1.2049139833450306,
3059
+ 1.0974795579910281,
3060
+ 1.0,
3061
+ 0.9999049079418182,
3062
+ 0.9987021696567535,
3063
+ 0.9999175441265106,
3064
+ 1.0,
3065
+ 1.0,
3066
+ 1.0001929998397827
3067
+ ],
3068
+ "mask": [
3069
+ true,
3070
+ true,
3071
+ true,
3072
+ false,
3073
+ false,
3074
+ false,
3075
+ false,
3076
+ false,
3077
+ false,
3078
+ false
3079
+ ]
3080
+ },
3081
+ "proprio": {
3082
+ "mean": [
3083
+ 0.017874393612146378,
3084
+ 0.10983800888061523,
3085
+ 0.78847336769104,
3086
+ 0.5196760892868042,
3087
+ -0.08152023702859879,
3088
+ 0.17145685851573944,
3089
+ 0.16265232861042023,
3090
+ 0.4587028920650482,
3091
+ -0.27345868945121765,
3092
+ -0.09970496594905853
3093
+ ],
3094
+ "std": [
3095
+ 0.18378695845603943,
3096
+ 0.09588225930929184,
3097
+ 0.08683514595031738,
3098
+ 0.46486517786979675,
3099
+ 0.5308765769004822,
3100
+ 0.44253969192504883,
3101
+ 0.6481422781944275,
3102
+ 0.39143168926239014,
3103
+ 0.32712316513061523,
3104
+ 0.8921455144882202
3105
+ ],
3106
+ "max": [
3107
+ 0.8753991436958313,
3108
+ 1.2016867852210993,
3109
+ 1.0952961683273315,
3110
+ 1.0,
3111
+ 0.9999044907093048,
3112
+ 0.9986884701251983,
3113
+ 0.9999175071716309,
3114
+ 1.0,
3115
+ 1.0,
3116
+ 1.0645089149475098
3117
+ ],
3118
+ "min": [
3119
+ -0.8379243296384812,
3120
+ -0.6292159473896026,
3121
+ -0.4275088250637054,
3122
+ -0.9999994039535522,
3123
+ -0.999635414481163,
3124
+ -0.996793516278267,
3125
+ -0.99927858710289,
3126
+ -0.9999648547172546,
3127
+ -0.9921211671829223,
3128
+ -1.0
3129
+ ],
3130
+ "q01": [
3131
+ -0.8379243296384812,
3132
+ -0.6292159473896026,
3133
+ -0.4275088250637054,
3134
+ -0.9999994039535522,
3135
+ -0.999635414481163,
3136
+ -0.996793516278267,
3137
+ -0.99927858710289,
3138
+ -0.9999648547172546,
3139
+ -0.9921211671829223,
3140
+ -1.0
3141
+ ],
3142
+ "q99": [
3143
+ 0.8753991436958313,
3144
+ 1.2016867852210993,
3145
+ 1.0952961683273315,
3146
+ 1.0,
3147
+ 0.9999044907093048,
3148
+ 0.9986884701251983,
3149
+ 0.9999175071716309,
3150
+ 1.0,
3151
+ 1.0,
3152
+ 1.0645089149475098
3153
+ ],
3154
+ "mask": [
3155
+ true,
3156
+ true,
3157
+ true,
3158
+ false,
3159
+ false,
3160
+ false,
3161
+ false,
3162
+ false,
3163
+ false,
3164
+ false
3165
+ ]
3166
+ },
3167
+ "num_transitions": 5972271,
3168
+ "num_trajectories": 43264
3169
+ },
3170
+ "fmb_gresearch": {
3171
+ "action": {
3172
+ "mean": [
3173
+ 0.5149797797203064,
3174
+ -0.044962573796510696,
3175
+ 0.17422549426555634,
3176
+ 0.7991654872894287,
3177
+ 0.006531611550599337,
3178
+ -0.5243619680404663,
3179
+ 0.3250426948070526,
3180
+ 0.7283564805984497,
3181
+ 0.5478328466415405,
3182
+ 0.46337518095970154
3183
+ ],
3184
+ "std": [
3185
+ 0.0708208978176117,
3186
+ 0.13683798909187317,
3187
+ 0.0761876031756401,
3188
+ 0.09095892310142517,
3189
+ 0.08736640214920044,
3190
+ 0.2576131522655487,
3191
+ 0.1862049400806427,
3192
+ 0.09788890928030014,
3193
+ 0.13817831873893738,
3194
+ 0.8886659741401672
3195
+ ],
3196
+ "max": [
3197
+ 0.8761805367469786,
3198
+ 1.2049139833450306,
3199
+ 1.0974795579910281,
3200
+ 1.0,
3201
+ 0.9999049079418182,
3202
+ 0.9987021696567535,
3203
+ 0.9999175441265106,
3204
+ 1.0,
3205
+ 1.0,
3206
+ 1.0001929998397827
3207
+ ],
3208
+ "min": [
3209
+ -0.8384313821792603,
3210
+ -0.6293186390399933,
3211
+ -0.42908657133579253,
3212
+ -0.9999992847442627,
3213
+ -0.9996374249458313,
3214
+ -0.9969306069612504,
3215
+ -0.99928431391716,
3216
+ -0.9999619847536088,
3217
+ -0.9921636700630188,
3218
+ -1.0002006542682649
3219
+ ],
3220
+ "q01": [
3221
+ -0.8384313821792603,
3222
+ -0.6293186390399933,
3223
+ -0.42908657133579253,
3224
+ -0.9999992847442627,
3225
+ -0.9996374249458313,
3226
+ -0.9969306069612504,
3227
+ -0.99928431391716,
3228
+ -0.9999619847536088,
3229
+ -0.9921636700630188,
3230
+ -1.0002006542682649
3231
+ ],
3232
+ "q99": [
3233
+ 0.8761805367469786,
3234
+ 1.2049139833450306,
3235
+ 1.0974795579910281,
3236
+ 1.0,
3237
+ 0.9999049079418182,
3238
+ 0.9987021696567535,
3239
+ 0.9999175441265106,
3240
+ 1.0,
3241
+ 1.0,
3242
+ 1.0001929998397827
3243
+ ],
3244
+ "mask": [
3245
+ true,
3246
+ true,
3247
+ true,
3248
+ false,
3249
+ false,
3250
+ false,
3251
+ false,
3252
+ false,
3253
+ false,
3254
+ false
3255
+ ]
3256
+ },
3257
+ "proprio": {
3258
+ "mean": [
3259
+ 0.514910101890564,
3260
+ -0.04377448558807373,
3261
+ 0.17520515620708466,
3262
+ 0.7986150979995728,
3263
+ 0.006624775938689709,
3264
+ -0.5256816148757935,
3265
+ 0.32580259442329407,
3266
+ 0.7288216948509216,
3267
+ 0.5471109747886658,
3268
+ 0.4554476737976074
3269
+ ],
3270
+ "std": [
3271
+ 0.07067131996154785,
3272
+ 0.1367093026638031,
3273
+ 0.07637086510658264,
3274
+ 0.09053026884794235,
3275
+ 0.08733317255973816,
3276
+ 0.25671660900115967,
3277
+ 0.185720294713974,
3278
+ 0.09751445800065994,
3279
+ 0.13750971853733063,
3280
+ 0.8919597864151001
3281
+ ],
3282
+ "max": [
3283
+ 0.8753991436958313,
3284
+ 1.2016867852210993,
3285
+ 1.0952961683273315,
3286
+ 1.0,
3287
+ 0.9999044907093048,
3288
+ 0.9986884701251983,
3289
+ 0.9999175071716309,
3290
+ 1.0,
3291
+ 1.0,
3292
+ 1.0645089149475098
3293
+ ],
3294
+ "min": [
3295
+ -0.8379243296384812,
3296
+ -0.6292159473896026,
3297
+ -0.4275088250637054,
3298
+ -0.9999994039535522,
3299
+ -0.999635414481163,
3300
+ -0.996793516278267,
3301
+ -0.99927858710289,
3302
+ -0.9999648547172546,
3303
+ -0.9921211671829223,
3304
+ -1.0
3305
+ ],
3306
+ "q01": [
3307
+ -0.8379243296384812,
3308
+ -0.6292159473896026,
3309
+ -0.4275088250637054,
3310
+ -0.9999994039535522,
3311
+ -0.999635414481163,
3312
+ -0.996793516278267,
3313
+ -0.99927858710289,
3314
+ -0.9999648547172546,
3315
+ -0.9921211671829223,
3316
+ -1.0
3317
+ ],
3318
+ "q99": [
3319
+ 0.8753991436958313,
3320
+ 1.2016867852210993,
3321
+ 1.0952961683273315,
3322
+ 1.0,
3323
+ 0.9999044907093048,
3324
+ 0.9986884701251983,
3325
+ 0.9999175071716309,
3326
+ 1.0,
3327
+ 1.0,
3328
+ 1.0645089149475098
3329
+ ],
3330
+ "mask": [
3331
+ true,
3332
+ true,
3333
+ true,
3334
+ false,
3335
+ false,
3336
+ false,
3337
+ false,
3338
+ false,
3339
+ false,
3340
+ false
3341
+ ]
3342
+ },
3343
+ "num_transitions": 1128729,
3344
+ "num_trajectories": 8611
3345
+ },
3346
+ "dobbe_gresearch": {
3347
+ "action": {
3348
+ "mean": [
3349
+ 0.0061858659610152245,
3350
+ 0.26042863726615906,
3351
+ -0.018509425222873688,
3352
+ 0.8032262921333313,
3353
+ 0.011619718745350838,
3354
+ 0.08878090232610703,
3355
+ 0.005480863619595766,
3356
+ 0.7486934065818787,
3357
+ -0.03153136000037193,
3358
+ 0.6124216318130493
3359
+ ],
3360
+ "std": [
3361
+ 0.7869429588317871,
3362
+ 0.5556016564369202,
3363
+ 2.2371985912323,
3364
+ 0.3514711260795593,
3365
+ 0.24396458268165588,
3366
+ 0.4074695408344269,
3367
+ 0.34791073203086853,
3368
+ 0.5444973707199097,
3369
+ 0.15351136028766632,
3370
+ 0.4484878182411194
3371
+ ],
3372
+ "max": [
3373
+ 0.8761805367469786,
3374
+ 1.2049139833450306,
3375
+ 1.0974795579910281,
3376
+ 1.0,
3377
+ 0.9999049079418182,
3378
+ 0.9987021696567535,
3379
+ 0.9999175441265106,
3380
+ 1.0,
3381
+ 1.0,
3382
+ 1.0001929998397827
3383
+ ],
3384
+ "min": [
3385
+ -0.8384313821792603,
3386
+ -0.6293186390399933,
3387
+ -0.42908657133579253,
3388
+ -0.9999992847442627,
3389
+ -0.9996374249458313,
3390
+ -0.9969306069612504,
3391
+ -0.99928431391716,
3392
+ -0.9999619847536088,
3393
+ -0.9921636700630188,
3394
+ -1.0002006542682649
3395
+ ],
3396
+ "q01": [
3397
+ -0.8384313821792603,
3398
+ -0.6293186390399933,
3399
+ -0.42908657133579253,
3400
+ -0.9999992847442627,
3401
+ -0.9996374249458313,
3402
+ -0.9969306069612504,
3403
+ -0.99928431391716,
3404
+ -0.9999619847536088,
3405
+ -0.9921636700630188,
3406
+ -1.0002006542682649
3407
+ ],
3408
+ "q99": [
3409
+ 0.8761805367469786,
3410
+ 1.2049139833450306,
3411
+ 1.0974795579910281,
3412
+ 1.0,
3413
+ 0.9999049079418182,
3414
+ 0.9987021696567535,
3415
+ 0.9999175441265106,
3416
+ 1.0,
3417
+ 1.0,
3418
+ 1.0001929998397827
3419
+ ],
3420
+ "mask": [
3421
+ true,
3422
+ true,
3423
+ true,
3424
+ false,
3425
+ false,
3426
+ false,
3427
+ false,
3428
+ false,
3429
+ false,
3430
+ false
3431
+ ]
3432
+ },
3433
+ "proprio": {
3434
+ "mean": [
3435
+ 0.006298029329627752,
3436
+ 0.2592981457710266,
3437
+ -0.018407080322504044,
3438
+ 0.804182767868042,
3439
+ 0.011643029749393463,
3440
+ 0.08833134174346924,
3441
+ 0.005384726449847221,
3442
+ 0.7499514222145081,
3443
+ -0.03142579272389412,
3444
+ 0.2645684480667114
3445
+ ],
3446
+ "std": [
3447
+ 0.7855807542800903,
3448
+ 0.5538082718849182,
3449
+ 2.2336723804473877,
3450
+ 0.35101377964019775,
3451
+ 0.2431618571281433,
3452
+ 0.40661847591400146,
3453
+ 0.3470095694065094,
3454
+ 0.5436821579933167,
3455
+ 0.15308144688606262,
3456
+ 0.7936993837356567
3457
+ ],
3458
+ "max": [
3459
+ 0.8753991436958313,
3460
+ 1.2016867852210993,
3461
+ 1.0952961683273315,
3462
+ 1.0,
3463
+ 0.9999044907093048,
3464
+ 0.9986884701251983,
3465
+ 0.9999175071716309,
3466
+ 1.0,
3467
+ 1.0,
3468
+ 1.0645089149475098
3469
+ ],
3470
+ "min": [
3471
+ -0.8379243296384812,
3472
+ -0.6292159473896026,
3473
+ -0.4275088250637054,
3474
+ -0.9999994039535522,
3475
+ -0.999635414481163,
3476
+ -0.996793516278267,
3477
+ -0.99927858710289,
3478
+ -0.9999648547172546,
3479
+ -0.9921211671829223,
3480
+ -1.0
3481
+ ],
3482
+ "q01": [
3483
+ -0.8379243296384812,
3484
+ -0.6292159473896026,
3485
+ -0.4275088250637054,
3486
+ -0.9999994039535522,
3487
+ -0.999635414481163,
3488
+ -0.996793516278267,
3489
+ -0.99927858710289,
3490
+ -0.9999648547172546,
3491
+ -0.9921211671829223,
3492
+ -1.0
3493
+ ],
3494
+ "q99": [
3495
+ 0.8753991436958313,
3496
+ 1.2016867852210993,
3497
+ 1.0952961683273315,
3498
+ 1.0,
3499
+ 0.9999044907093048,
3500
+ 0.9986884701251983,
3501
+ 0.9999175071716309,
3502
+ 1.0,
3503
+ 1.0,
3504
+ 1.0645089149475098
3505
+ ],
3506
+ "mask": [
3507
+ true,
3508
+ true,
3509
+ true,
3510
+ false,
3511
+ false,
3512
+ false,
3513
+ false,
3514
+ false,
3515
+ false,
3516
+ false
3517
+ ]
3518
+ },
3519
+ "num_transitions": 1134703,
3520
+ "num_trajectories": 5208
3521
+ },
3522
+ "droid_gresearch": {
3523
+ "action": {
3524
+ "mean": [
3525
+ 0.5361732244491577,
3526
+ 0.001534969313070178,
3527
+ 0.3150203227996826,
3528
+ 0.6224499344825745,
3529
+ -0.03386346250772476,
3530
+ 0.07832133769989014,
3531
+ -0.042715393006801605,
3532
+ -0.6224499940872192,
3533
+ 0.017628736793994904,
3534
+ -0.0924207791686058
3535
+ ],
3536
+ "std": [
3537
+ 0.11631692945957184,
3538
+ 0.17411848902702332,
3539
+ 0.161155104637146,
3540
+ 0.3620404601097107,
3541
+ 0.48013269901275635,
3542
+ 0.3059387505054474,
3543
+ 0.4865337312221527,
3544
+ 0.3760391175746918,
3545
+ 0.2518838942050934,
3546
+ 0.8853684067726135
3547
+ ],
3548
+ "max": [
3549
+ 0.8761805367469786,
3550
+ 1.2049139833450306,
3551
+ 1.0974795579910281,
3552
+ 1.0,
3553
+ 0.9999049079418182,
3554
+ 0.9987021696567535,
3555
+ 0.9999175441265106,
3556
+ 1.0,
3557
+ 1.0,
3558
+ 1.0001929998397827
3559
+ ],
3560
+ "min": [
3561
+ -0.8384313821792603,
3562
+ -0.6293186390399933,
3563
+ -0.42908657133579253,
3564
+ -0.9999992847442627,
3565
+ -0.9996374249458313,
3566
+ -0.9969306069612504,
3567
+ -0.99928431391716,
3568
+ -0.9999619847536088,
3569
+ -0.9921636700630188,
3570
+ -1.0002006542682649
3571
+ ],
3572
+ "q01": [
3573
+ -0.8384313821792603,
3574
+ -0.6293186390399933,
3575
+ -0.42908657133579253,
3576
+ -0.9999992847442627,
3577
+ -0.9996374249458313,
3578
+ -0.9969306069612504,
3579
+ -0.99928431391716,
3580
+ -0.9999619847536088,
3581
+ -0.9921636700630188,
3582
+ -1.0002006542682649
3583
+ ],
3584
+ "q99": [
3585
+ 0.8761805367469786,
3586
+ 1.2049139833450306,
3587
+ 1.0974795579910281,
3588
+ 1.0,
3589
+ 0.9999049079418182,
3590
+ 0.9987021696567535,
3591
+ 0.9999175441265106,
3592
+ 1.0,
3593
+ 1.0,
3594
+ 1.0001929998397827
3595
+ ],
3596
+ "mask": [
3597
+ true,
3598
+ true,
3599
+ true,
3600
+ false,
3601
+ false,
3602
+ false,
3603
+ false,
3604
+ false,
3605
+ false,
3606
+ false
3607
+ ]
3608
+ },
3609
+ "proprio": {
3610
+ "mean": [
3611
+ 0.5352964997291565,
3612
+ 0.0015270119765773416,
3613
+ 0.31521719694137573,
3614
+ 0.6224499344825745,
3615
+ -0.03372149169445038,
3616
+ 0.0780143290758133,
3617
+ -0.04255221411585808,
3618
+ -0.622450053691864,
3619
+ 0.017578328028321266,
3620
+ -0.27697455883026123
3621
+ ],
3622
+ "std": [
3623
+ 0.11647670716047287,
3624
+ 0.17398209869861603,
3625
+ 0.16120275855064392,
3626
+ 0.3619924485683441,
3627
+ 0.4795165956020355,
3628
+ 0.3056667149066925,
3629
+ 0.48590150475502014,
3630
+ 0.37595245242118835,
3631
+ 0.25170600414276123,
3632
+ 0.8289845585823059
3633
+ ],
3634
+ "max": [
3635
+ 0.8753991436958313,
3636
+ 1.2016867852210993,
3637
+ 1.0952961683273315,
3638
+ 1.0,
3639
+ 0.9999044907093048,
3640
+ 0.9986884701251983,
3641
+ 0.9999175071716309,
3642
+ 1.0,
3643
+ 1.0,
3644
+ 1.0645089149475098
3645
+ ],
3646
+ "min": [
3647
+ -0.8379243296384812,
3648
+ -0.6292159473896026,
3649
+ -0.4275088250637054,
3650
+ -0.9999994039535522,
3651
+ -0.999635414481163,
3652
+ -0.996793516278267,
3653
+ -0.99927858710289,
3654
+ -0.9999648547172546,
3655
+ -0.9921211671829223,
3656
+ -1.0
3657
+ ],
3658
+ "q01": [
3659
+ -0.8379243296384812,
3660
+ -0.6292159473896026,
3661
+ -0.4275088250637054,
3662
+ -0.9999994039535522,
3663
+ -0.999635414481163,
3664
+ -0.996793516278267,
3665
+ -0.99927858710289,
3666
+ -0.9999648547172546,
3667
+ -0.9921211671829223,
3668
+ -1.0
3669
+ ],
3670
+ "q99": [
3671
+ 0.8753991436958313,
3672
+ 1.2016867852210993,
3673
+ 1.0952961683273315,
3674
+ 1.0,
3675
+ 0.9999044907093048,
3676
+ 0.9986884701251983,
3677
+ 0.9999175071716309,
3678
+ 1.0,
3679
+ 1.0,
3680
+ 1.0645089149475098
3681
+ ],
3682
+ "mask": [
3683
+ true,
3684
+ true,
3685
+ true,
3686
+ false,
3687
+ false,
3688
+ false,
3689
+ false,
3690
+ false,
3691
+ false,
3692
+ false
3693
+ ]
3694
+ },
3695
+ "num_transitions": 26952093,
3696
+ "num_trajectories": 92233
3697
+ }
3698
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77aeab26f3d0c175502b4a0a817aada9d15819c443489f04d65aa104856849fa
3
+ size 4985048272
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d20245b20352ccb66cc5f254156ad3935d4e8380cc46bbb478154cf35c55aa8a
3
+ size 177538408
model.safetensors.index.json ADDED
@@ -0,0 +1,996 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 5162458672
4
+ },
5
+ "weight_map": {
6
+ "action_head.combine.bias": "model-00002-of-00002.safetensors",
7
+ "action_head.combine.weight": "model-00002-of-00002.safetensors",
8
+ "action_head.net.blocks.0.attn.proj.bias": "model-00001-of-00002.safetensors",
9
+ "action_head.net.blocks.0.attn.proj.weight": "model-00001-of-00002.safetensors",
10
+ "action_head.net.blocks.0.attn.qkv.bias": "model-00001-of-00002.safetensors",
11
+ "action_head.net.blocks.0.attn.qkv.weight": "model-00001-of-00002.safetensors",
12
+ "action_head.net.blocks.0.mlp.fc1.bias": "model-00001-of-00002.safetensors",
13
+ "action_head.net.blocks.0.mlp.fc1.weight": "model-00001-of-00002.safetensors",
14
+ "action_head.net.blocks.0.mlp.fc2.bias": "model-00001-of-00002.safetensors",
15
+ "action_head.net.blocks.0.mlp.fc2.weight": "model-00001-of-00002.safetensors",
16
+ "action_head.net.blocks.1.attn.proj.bias": "model-00001-of-00002.safetensors",
17
+ "action_head.net.blocks.1.attn.proj.weight": "model-00001-of-00002.safetensors",
18
+ "action_head.net.blocks.1.attn.qkv.bias": "model-00001-of-00002.safetensors",
19
+ "action_head.net.blocks.1.attn.qkv.weight": "model-00001-of-00002.safetensors",
20
+ "action_head.net.blocks.1.mlp.fc1.bias": "model-00001-of-00002.safetensors",
21
+ "action_head.net.blocks.1.mlp.fc1.weight": "model-00001-of-00002.safetensors",
22
+ "action_head.net.blocks.1.mlp.fc2.bias": "model-00001-of-00002.safetensors",
23
+ "action_head.net.blocks.1.mlp.fc2.weight": "model-00001-of-00002.safetensors",
24
+ "action_head.net.blocks.10.attn.proj.bias": "model-00001-of-00002.safetensors",
25
+ "action_head.net.blocks.10.attn.proj.weight": "model-00001-of-00002.safetensors",
26
+ "action_head.net.blocks.10.attn.qkv.bias": "model-00001-of-00002.safetensors",
27
+ "action_head.net.blocks.10.attn.qkv.weight": "model-00001-of-00002.safetensors",
28
+ "action_head.net.blocks.10.mlp.fc1.bias": "model-00001-of-00002.safetensors",
29
+ "action_head.net.blocks.10.mlp.fc1.weight": "model-00001-of-00002.safetensors",
30
+ "action_head.net.blocks.10.mlp.fc2.bias": "model-00001-of-00002.safetensors",
31
+ "action_head.net.blocks.10.mlp.fc2.weight": "model-00001-of-00002.safetensors",
32
+ "action_head.net.blocks.11.attn.proj.bias": "model-00001-of-00002.safetensors",
33
+ "action_head.net.blocks.11.attn.proj.weight": "model-00001-of-00002.safetensors",
34
+ "action_head.net.blocks.11.attn.qkv.bias": "model-00001-of-00002.safetensors",
35
+ "action_head.net.blocks.11.attn.qkv.weight": "model-00001-of-00002.safetensors",
36
+ "action_head.net.blocks.11.mlp.fc1.bias": "model-00001-of-00002.safetensors",
37
+ "action_head.net.blocks.11.mlp.fc1.weight": "model-00001-of-00002.safetensors",
38
+ "action_head.net.blocks.11.mlp.fc2.bias": "model-00001-of-00002.safetensors",
39
+ "action_head.net.blocks.11.mlp.fc2.weight": "model-00001-of-00002.safetensors",
40
+ "action_head.net.blocks.12.attn.proj.bias": "model-00001-of-00002.safetensors",
41
+ "action_head.net.blocks.12.attn.proj.weight": "model-00001-of-00002.safetensors",
42
+ "action_head.net.blocks.12.attn.qkv.bias": "model-00001-of-00002.safetensors",
43
+ "action_head.net.blocks.12.attn.qkv.weight": "model-00001-of-00002.safetensors",
44
+ "action_head.net.blocks.12.mlp.fc1.bias": "model-00001-of-00002.safetensors",
45
+ "action_head.net.blocks.12.mlp.fc1.weight": "model-00001-of-00002.safetensors",
46
+ "action_head.net.blocks.12.mlp.fc2.bias": "model-00001-of-00002.safetensors",
47
+ "action_head.net.blocks.12.mlp.fc2.weight": "model-00001-of-00002.safetensors",
48
+ "action_head.net.blocks.13.attn.proj.bias": "model-00001-of-00002.safetensors",
49
+ "action_head.net.blocks.13.attn.proj.weight": "model-00001-of-00002.safetensors",
50
+ "action_head.net.blocks.13.attn.qkv.bias": "model-00001-of-00002.safetensors",
51
+ "action_head.net.blocks.13.attn.qkv.weight": "model-00001-of-00002.safetensors",
52
+ "action_head.net.blocks.13.mlp.fc1.bias": "model-00001-of-00002.safetensors",
53
+ "action_head.net.blocks.13.mlp.fc1.weight": "model-00001-of-00002.safetensors",
54
+ "action_head.net.blocks.13.mlp.fc2.bias": "model-00001-of-00002.safetensors",
55
+ "action_head.net.blocks.13.mlp.fc2.weight": "model-00001-of-00002.safetensors",
56
+ "action_head.net.blocks.14.attn.proj.bias": "model-00001-of-00002.safetensors",
57
+ "action_head.net.blocks.14.attn.proj.weight": "model-00001-of-00002.safetensors",
58
+ "action_head.net.blocks.14.attn.qkv.bias": "model-00001-of-00002.safetensors",
59
+ "action_head.net.blocks.14.attn.qkv.weight": "model-00001-of-00002.safetensors",
60
+ "action_head.net.blocks.14.mlp.fc1.bias": "model-00001-of-00002.safetensors",
61
+ "action_head.net.blocks.14.mlp.fc1.weight": "model-00001-of-00002.safetensors",
62
+ "action_head.net.blocks.14.mlp.fc2.bias": "model-00001-of-00002.safetensors",
63
+ "action_head.net.blocks.14.mlp.fc2.weight": "model-00001-of-00002.safetensors",
64
+ "action_head.net.blocks.15.attn.proj.bias": "model-00001-of-00002.safetensors",
65
+ "action_head.net.blocks.15.attn.proj.weight": "model-00001-of-00002.safetensors",
66
+ "action_head.net.blocks.15.attn.qkv.bias": "model-00001-of-00002.safetensors",
67
+ "action_head.net.blocks.15.attn.qkv.weight": "model-00001-of-00002.safetensors",
68
+ "action_head.net.blocks.15.mlp.fc1.bias": "model-00001-of-00002.safetensors",
69
+ "action_head.net.blocks.15.mlp.fc1.weight": "model-00001-of-00002.safetensors",
70
+ "action_head.net.blocks.15.mlp.fc2.bias": "model-00001-of-00002.safetensors",
71
+ "action_head.net.blocks.15.mlp.fc2.weight": "model-00001-of-00002.safetensors",
72
+ "action_head.net.blocks.16.attn.proj.bias": "model-00001-of-00002.safetensors",
73
+ "action_head.net.blocks.16.attn.proj.weight": "model-00001-of-00002.safetensors",
74
+ "action_head.net.blocks.16.attn.qkv.bias": "model-00001-of-00002.safetensors",
75
+ "action_head.net.blocks.16.attn.qkv.weight": "model-00001-of-00002.safetensors",
76
+ "action_head.net.blocks.16.mlp.fc1.bias": "model-00001-of-00002.safetensors",
77
+ "action_head.net.blocks.16.mlp.fc1.weight": "model-00001-of-00002.safetensors",
78
+ "action_head.net.blocks.16.mlp.fc2.bias": "model-00001-of-00002.safetensors",
79
+ "action_head.net.blocks.16.mlp.fc2.weight": "model-00001-of-00002.safetensors",
80
+ "action_head.net.blocks.17.attn.proj.bias": "model-00001-of-00002.safetensors",
81
+ "action_head.net.blocks.17.attn.proj.weight": "model-00001-of-00002.safetensors",
82
+ "action_head.net.blocks.17.attn.qkv.bias": "model-00001-of-00002.safetensors",
83
+ "action_head.net.blocks.17.attn.qkv.weight": "model-00001-of-00002.safetensors",
84
+ "action_head.net.blocks.17.mlp.fc1.bias": "model-00001-of-00002.safetensors",
85
+ "action_head.net.blocks.17.mlp.fc1.weight": "model-00001-of-00002.safetensors",
86
+ "action_head.net.blocks.17.mlp.fc2.bias": "model-00001-of-00002.safetensors",
87
+ "action_head.net.blocks.17.mlp.fc2.weight": "model-00001-of-00002.safetensors",
88
+ "action_head.net.blocks.18.attn.proj.bias": "model-00001-of-00002.safetensors",
89
+ "action_head.net.blocks.18.attn.proj.weight": "model-00001-of-00002.safetensors",
90
+ "action_head.net.blocks.18.attn.qkv.bias": "model-00001-of-00002.safetensors",
91
+ "action_head.net.blocks.18.attn.qkv.weight": "model-00001-of-00002.safetensors",
92
+ "action_head.net.blocks.18.mlp.fc1.bias": "model-00001-of-00002.safetensors",
93
+ "action_head.net.blocks.18.mlp.fc1.weight": "model-00001-of-00002.safetensors",
94
+ "action_head.net.blocks.18.mlp.fc2.bias": "model-00001-of-00002.safetensors",
95
+ "action_head.net.blocks.18.mlp.fc2.weight": "model-00001-of-00002.safetensors",
96
+ "action_head.net.blocks.19.attn.proj.bias": "model-00001-of-00002.safetensors",
97
+ "action_head.net.blocks.19.attn.proj.weight": "model-00001-of-00002.safetensors",
98
+ "action_head.net.blocks.19.attn.qkv.bias": "model-00001-of-00002.safetensors",
99
+ "action_head.net.blocks.19.attn.qkv.weight": "model-00001-of-00002.safetensors",
100
+ "action_head.net.blocks.19.mlp.fc1.bias": "model-00001-of-00002.safetensors",
101
+ "action_head.net.blocks.19.mlp.fc1.weight": "model-00001-of-00002.safetensors",
102
+ "action_head.net.blocks.19.mlp.fc2.bias": "model-00001-of-00002.safetensors",
103
+ "action_head.net.blocks.19.mlp.fc2.weight": "model-00001-of-00002.safetensors",
104
+ "action_head.net.blocks.2.attn.proj.bias": "model-00001-of-00002.safetensors",
105
+ "action_head.net.blocks.2.attn.proj.weight": "model-00001-of-00002.safetensors",
106
+ "action_head.net.blocks.2.attn.qkv.bias": "model-00001-of-00002.safetensors",
107
+ "action_head.net.blocks.2.attn.qkv.weight": "model-00001-of-00002.safetensors",
108
+ "action_head.net.blocks.2.mlp.fc1.bias": "model-00001-of-00002.safetensors",
109
+ "action_head.net.blocks.2.mlp.fc1.weight": "model-00001-of-00002.safetensors",
110
+ "action_head.net.blocks.2.mlp.fc2.bias": "model-00001-of-00002.safetensors",
111
+ "action_head.net.blocks.2.mlp.fc2.weight": "model-00001-of-00002.safetensors",
112
+ "action_head.net.blocks.20.attn.proj.bias": "model-00001-of-00002.safetensors",
113
+ "action_head.net.blocks.20.attn.proj.weight": "model-00001-of-00002.safetensors",
114
+ "action_head.net.blocks.20.attn.qkv.bias": "model-00001-of-00002.safetensors",
115
+ "action_head.net.blocks.20.attn.qkv.weight": "model-00001-of-00002.safetensors",
116
+ "action_head.net.blocks.20.mlp.fc1.bias": "model-00001-of-00002.safetensors",
117
+ "action_head.net.blocks.20.mlp.fc1.weight": "model-00001-of-00002.safetensors",
118
+ "action_head.net.blocks.20.mlp.fc2.bias": "model-00002-of-00002.safetensors",
119
+ "action_head.net.blocks.20.mlp.fc2.weight": "model-00002-of-00002.safetensors",
120
+ "action_head.net.blocks.21.attn.proj.bias": "model-00002-of-00002.safetensors",
121
+ "action_head.net.blocks.21.attn.proj.weight": "model-00002-of-00002.safetensors",
122
+ "action_head.net.blocks.21.attn.qkv.bias": "model-00002-of-00002.safetensors",
123
+ "action_head.net.blocks.21.attn.qkv.weight": "model-00002-of-00002.safetensors",
124
+ "action_head.net.blocks.21.mlp.fc1.bias": "model-00002-of-00002.safetensors",
125
+ "action_head.net.blocks.21.mlp.fc1.weight": "model-00002-of-00002.safetensors",
126
+ "action_head.net.blocks.21.mlp.fc2.bias": "model-00002-of-00002.safetensors",
127
+ "action_head.net.blocks.21.mlp.fc2.weight": "model-00002-of-00002.safetensors",
128
+ "action_head.net.blocks.22.attn.proj.bias": "model-00002-of-00002.safetensors",
129
+ "action_head.net.blocks.22.attn.proj.weight": "model-00002-of-00002.safetensors",
130
+ "action_head.net.blocks.22.attn.qkv.bias": "model-00002-of-00002.safetensors",
131
+ "action_head.net.blocks.22.attn.qkv.weight": "model-00002-of-00002.safetensors",
132
+ "action_head.net.blocks.22.mlp.fc1.bias": "model-00002-of-00002.safetensors",
133
+ "action_head.net.blocks.22.mlp.fc1.weight": "model-00002-of-00002.safetensors",
134
+ "action_head.net.blocks.22.mlp.fc2.bias": "model-00002-of-00002.safetensors",
135
+ "action_head.net.blocks.22.mlp.fc2.weight": "model-00002-of-00002.safetensors",
136
+ "action_head.net.blocks.23.attn.proj.bias": "model-00002-of-00002.safetensors",
137
+ "action_head.net.blocks.23.attn.proj.weight": "model-00002-of-00002.safetensors",
138
+ "action_head.net.blocks.23.attn.qkv.bias": "model-00002-of-00002.safetensors",
139
+ "action_head.net.blocks.23.attn.qkv.weight": "model-00002-of-00002.safetensors",
140
+ "action_head.net.blocks.23.mlp.fc1.bias": "model-00002-of-00002.safetensors",
141
+ "action_head.net.blocks.23.mlp.fc1.weight": "model-00002-of-00002.safetensors",
142
+ "action_head.net.blocks.23.mlp.fc2.bias": "model-00002-of-00002.safetensors",
143
+ "action_head.net.blocks.23.mlp.fc2.weight": "model-00002-of-00002.safetensors",
144
+ "action_head.net.blocks.3.attn.proj.bias": "model-00001-of-00002.safetensors",
145
+ "action_head.net.blocks.3.attn.proj.weight": "model-00001-of-00002.safetensors",
146
+ "action_head.net.blocks.3.attn.qkv.bias": "model-00001-of-00002.safetensors",
147
+ "action_head.net.blocks.3.attn.qkv.weight": "model-00001-of-00002.safetensors",
148
+ "action_head.net.blocks.3.mlp.fc1.bias": "model-00001-of-00002.safetensors",
149
+ "action_head.net.blocks.3.mlp.fc1.weight": "model-00001-of-00002.safetensors",
150
+ "action_head.net.blocks.3.mlp.fc2.bias": "model-00001-of-00002.safetensors",
151
+ "action_head.net.blocks.3.mlp.fc2.weight": "model-00001-of-00002.safetensors",
152
+ "action_head.net.blocks.4.attn.proj.bias": "model-00001-of-00002.safetensors",
153
+ "action_head.net.blocks.4.attn.proj.weight": "model-00001-of-00002.safetensors",
154
+ "action_head.net.blocks.4.attn.qkv.bias": "model-00001-of-00002.safetensors",
155
+ "action_head.net.blocks.4.attn.qkv.weight": "model-00001-of-00002.safetensors",
156
+ "action_head.net.blocks.4.mlp.fc1.bias": "model-00001-of-00002.safetensors",
157
+ "action_head.net.blocks.4.mlp.fc1.weight": "model-00001-of-00002.safetensors",
158
+ "action_head.net.blocks.4.mlp.fc2.bias": "model-00001-of-00002.safetensors",
159
+ "action_head.net.blocks.4.mlp.fc2.weight": "model-00001-of-00002.safetensors",
160
+ "action_head.net.blocks.5.attn.proj.bias": "model-00001-of-00002.safetensors",
161
+ "action_head.net.blocks.5.attn.proj.weight": "model-00001-of-00002.safetensors",
162
+ "action_head.net.blocks.5.attn.qkv.bias": "model-00001-of-00002.safetensors",
163
+ "action_head.net.blocks.5.attn.qkv.weight": "model-00001-of-00002.safetensors",
164
+ "action_head.net.blocks.5.mlp.fc1.bias": "model-00001-of-00002.safetensors",
165
+ "action_head.net.blocks.5.mlp.fc1.weight": "model-00001-of-00002.safetensors",
166
+ "action_head.net.blocks.5.mlp.fc2.bias": "model-00001-of-00002.safetensors",
167
+ "action_head.net.blocks.5.mlp.fc2.weight": "model-00001-of-00002.safetensors",
168
+ "action_head.net.blocks.6.attn.proj.bias": "model-00001-of-00002.safetensors",
169
+ "action_head.net.blocks.6.attn.proj.weight": "model-00001-of-00002.safetensors",
170
+ "action_head.net.blocks.6.attn.qkv.bias": "model-00001-of-00002.safetensors",
171
+ "action_head.net.blocks.6.attn.qkv.weight": "model-00001-of-00002.safetensors",
172
+ "action_head.net.blocks.6.mlp.fc1.bias": "model-00001-of-00002.safetensors",
173
+ "action_head.net.blocks.6.mlp.fc1.weight": "model-00001-of-00002.safetensors",
174
+ "action_head.net.blocks.6.mlp.fc2.bias": "model-00001-of-00002.safetensors",
175
+ "action_head.net.blocks.6.mlp.fc2.weight": "model-00001-of-00002.safetensors",
176
+ "action_head.net.blocks.7.attn.proj.bias": "model-00001-of-00002.safetensors",
177
+ "action_head.net.blocks.7.attn.proj.weight": "model-00001-of-00002.safetensors",
178
+ "action_head.net.blocks.7.attn.qkv.bias": "model-00001-of-00002.safetensors",
179
+ "action_head.net.blocks.7.attn.qkv.weight": "model-00001-of-00002.safetensors",
180
+ "action_head.net.blocks.7.mlp.fc1.bias": "model-00001-of-00002.safetensors",
181
+ "action_head.net.blocks.7.mlp.fc1.weight": "model-00001-of-00002.safetensors",
182
+ "action_head.net.blocks.7.mlp.fc2.bias": "model-00001-of-00002.safetensors",
183
+ "action_head.net.blocks.7.mlp.fc2.weight": "model-00001-of-00002.safetensors",
184
+ "action_head.net.blocks.8.attn.proj.bias": "model-00001-of-00002.safetensors",
185
+ "action_head.net.blocks.8.attn.proj.weight": "model-00001-of-00002.safetensors",
186
+ "action_head.net.blocks.8.attn.qkv.bias": "model-00001-of-00002.safetensors",
187
+ "action_head.net.blocks.8.attn.qkv.weight": "model-00001-of-00002.safetensors",
188
+ "action_head.net.blocks.8.mlp.fc1.bias": "model-00001-of-00002.safetensors",
189
+ "action_head.net.blocks.8.mlp.fc1.weight": "model-00001-of-00002.safetensors",
190
+ "action_head.net.blocks.8.mlp.fc2.bias": "model-00001-of-00002.safetensors",
191
+ "action_head.net.blocks.8.mlp.fc2.weight": "model-00001-of-00002.safetensors",
192
+ "action_head.net.blocks.9.attn.proj.bias": "model-00001-of-00002.safetensors",
193
+ "action_head.net.blocks.9.attn.proj.weight": "model-00001-of-00002.safetensors",
194
+ "action_head.net.blocks.9.attn.qkv.bias": "model-00001-of-00002.safetensors",
195
+ "action_head.net.blocks.9.attn.qkv.weight": "model-00001-of-00002.safetensors",
196
+ "action_head.net.blocks.9.mlp.fc1.bias": "model-00001-of-00002.safetensors",
197
+ "action_head.net.blocks.9.mlp.fc1.weight": "model-00001-of-00002.safetensors",
198
+ "action_head.net.blocks.9.mlp.fc2.bias": "model-00001-of-00002.safetensors",
199
+ "action_head.net.blocks.9.mlp.fc2.weight": "model-00001-of-00002.safetensors",
200
+ "action_head.net.final_layer.linear.bias": "model-00002-of-00002.safetensors",
201
+ "action_head.net.final_layer.linear.weight": "model-00002-of-00002.safetensors",
202
+ "action_head.net.positional_embedding": "model-00001-of-00002.safetensors",
203
+ "action_head.net.t_embedder.mlp.0.bias": "model-00001-of-00002.safetensors",
204
+ "action_head.net.t_embedder.mlp.0.weight": "model-00001-of-00002.safetensors",
205
+ "action_head.net.t_embedder.mlp.2.bias": "model-00001-of-00002.safetensors",
206
+ "action_head.net.t_embedder.mlp.2.weight": "model-00001-of-00002.safetensors",
207
+ "action_head.net.x_embedder.linear.bias": "model-00001-of-00002.safetensors",
208
+ "action_head.net.x_embedder.linear.weight": "model-00001-of-00002.safetensors",
209
+ "action_head.net.z_embedder.linear.bias": "model-00001-of-00002.safetensors",
210
+ "action_head.net.z_embedder.linear.weight": "model-00001-of-00002.safetensors",
211
+ "action_head.net.z_embedder.uncondition": "model-00001-of-00002.safetensors",
212
+ "embed_arm_state.bias": "model-00001-of-00002.safetensors",
213
+ "embed_arm_state.weight": "model-00001-of-00002.safetensors",
214
+ "language_model.model.embed_tokens.weight": "model-00001-of-00002.safetensors",
215
+ "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
216
+ "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
217
+ "language_model.model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
218
+ "language_model.model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
219
+ "language_model.model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
220
+ "language_model.model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
221
+ "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
222
+ "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
223
+ "language_model.model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
224
+ "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
225
+ "language_model.model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
226
+ "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
227
+ "language_model.model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
228
+ "language_model.model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
229
+ "language_model.model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
230
+ "language_model.model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
231
+ "language_model.model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
232
+ "language_model.model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
233
+ "language_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
234
+ "language_model.model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
235
+ "language_model.model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
236
+ "language_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
237
+ "language_model.model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
238
+ "language_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
239
+ "language_model.model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
240
+ "language_model.model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
241
+ "language_model.model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
242
+ "language_model.model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
243
+ "language_model.model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
244
+ "language_model.model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
245
+ "language_model.model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
246
+ "language_model.model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
247
+ "language_model.model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
248
+ "language_model.model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
249
+ "language_model.model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
250
+ "language_model.model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
251
+ "language_model.model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
252
+ "language_model.model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
253
+ "language_model.model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
254
+ "language_model.model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
255
+ "language_model.model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
256
+ "language_model.model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
257
+ "language_model.model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
258
+ "language_model.model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
259
+ "language_model.model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
260
+ "language_model.model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
261
+ "language_model.model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
262
+ "language_model.model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
263
+ "language_model.model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
264
+ "language_model.model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
265
+ "language_model.model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
266
+ "language_model.model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
267
+ "language_model.model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
268
+ "language_model.model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
269
+ "language_model.model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
270
+ "language_model.model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
271
+ "language_model.model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
272
+ "language_model.model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
273
+ "language_model.model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
274
+ "language_model.model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
275
+ "language_model.model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
276
+ "language_model.model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
277
+ "language_model.model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
278
+ "language_model.model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
279
+ "language_model.model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
280
+ "language_model.model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
281
+ "language_model.model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
282
+ "language_model.model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
283
+ "language_model.model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
284
+ "language_model.model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
285
+ "language_model.model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
286
+ "language_model.model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
287
+ "language_model.model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
288
+ "language_model.model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
289
+ "language_model.model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
290
+ "language_model.model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
291
+ "language_model.model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
292
+ "language_model.model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
293
+ "language_model.model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
294
+ "language_model.model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
295
+ "language_model.model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
296
+ "language_model.model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
297
+ "language_model.model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
298
+ "language_model.model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
299
+ "language_model.model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
300
+ "language_model.model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
301
+ "language_model.model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
302
+ "language_model.model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
303
+ "language_model.model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
304
+ "language_model.model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
305
+ "language_model.model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
306
+ "language_model.model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
307
+ "language_model.model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
308
+ "language_model.model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
309
+ "language_model.model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
310
+ "language_model.model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
311
+ "language_model.model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
312
+ "language_model.model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
313
+ "language_model.model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
314
+ "language_model.model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
315
+ "language_model.model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
316
+ "language_model.model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
317
+ "language_model.model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
318
+ "language_model.model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
319
+ "language_model.model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
320
+ "language_model.model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
321
+ "language_model.model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
322
+ "language_model.model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
323
+ "language_model.model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
324
+ "language_model.model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
325
+ "language_model.model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
326
+ "language_model.model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
327
+ "language_model.model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
328
+ "language_model.model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
329
+ "language_model.model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
330
+ "language_model.model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
331
+ "language_model.model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
332
+ "language_model.model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
333
+ "language_model.model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
334
+ "language_model.model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
335
+ "language_model.model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
336
+ "language_model.model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
337
+ "language_model.model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
338
+ "language_model.model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
339
+ "language_model.model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
340
+ "language_model.model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
341
+ "language_model.model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
342
+ "language_model.model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
343
+ "language_model.model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
344
+ "language_model.model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
345
+ "language_model.model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
346
+ "language_model.model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
347
+ "language_model.model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
348
+ "language_model.model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
349
+ "language_model.model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
350
+ "language_model.model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
351
+ "language_model.model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
352
+ "language_model.model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
353
+ "language_model.model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
354
+ "language_model.model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
355
+ "language_model.model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
356
+ "language_model.model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
357
+ "language_model.model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
358
+ "language_model.model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
359
+ "language_model.model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
360
+ "language_model.model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
361
+ "language_model.model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
362
+ "language_model.model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
363
+ "language_model.model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
364
+ "language_model.model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
365
+ "language_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
366
+ "language_model.model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
367
+ "language_model.model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
368
+ "language_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
369
+ "language_model.model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
370
+ "language_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
371
+ "language_model.model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
372
+ "language_model.model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
373
+ "language_model.model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
374
+ "language_model.model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
375
+ "language_model.model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
376
+ "language_model.model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
377
+ "language_model.model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
378
+ "language_model.model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
379
+ "language_model.model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
380
+ "language_model.model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
381
+ "language_model.model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
382
+ "language_model.model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
383
+ "language_model.model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
384
+ "language_model.model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
385
+ "language_model.model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
386
+ "language_model.model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
387
+ "language_model.model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
388
+ "language_model.model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
389
+ "language_model.model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
390
+ "language_model.model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
391
+ "language_model.model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
392
+ "language_model.model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
393
+ "language_model.model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
394
+ "language_model.model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
395
+ "language_model.model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
396
+ "language_model.model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
397
+ "language_model.model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
398
+ "language_model.model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
399
+ "language_model.model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
400
+ "language_model.model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
401
+ "language_model.model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
402
+ "language_model.model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
403
+ "language_model.model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
404
+ "language_model.model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
405
+ "language_model.model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
406
+ "language_model.model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
407
+ "language_model.model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
408
+ "language_model.model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
409
+ "language_model.model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
410
+ "language_model.model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
411
+ "language_model.model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
412
+ "language_model.model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
413
+ "language_model.model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
414
+ "language_model.model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
415
+ "language_model.model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
416
+ "language_model.model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
417
+ "language_model.model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
418
+ "language_model.model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
419
+ "language_model.model.layers.24.input_layernorm.weight": "model-00001-of-00002.safetensors",
420
+ "language_model.model.layers.24.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
421
+ "language_model.model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
422
+ "language_model.model.layers.24.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
423
+ "language_model.model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
424
+ "language_model.model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
425
+ "language_model.model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
426
+ "language_model.model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
427
+ "language_model.model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
428
+ "language_model.model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
429
+ "language_model.model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
430
+ "language_model.model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
431
+ "language_model.model.layers.25.input_layernorm.weight": "model-00001-of-00002.safetensors",
432
+ "language_model.model.layers.25.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
433
+ "language_model.model.layers.25.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
434
+ "language_model.model.layers.25.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
435
+ "language_model.model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
436
+ "language_model.model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
437
+ "language_model.model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
438
+ "language_model.model.layers.25.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
439
+ "language_model.model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
440
+ "language_model.model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
441
+ "language_model.model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
442
+ "language_model.model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
443
+ "language_model.model.layers.26.input_layernorm.weight": "model-00001-of-00002.safetensors",
444
+ "language_model.model.layers.26.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
445
+ "language_model.model.layers.26.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
446
+ "language_model.model.layers.26.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
447
+ "language_model.model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
448
+ "language_model.model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
449
+ "language_model.model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
450
+ "language_model.model.layers.26.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
451
+ "language_model.model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
452
+ "language_model.model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
453
+ "language_model.model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
454
+ "language_model.model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
455
+ "language_model.model.layers.27.input_layernorm.weight": "model-00001-of-00002.safetensors",
456
+ "language_model.model.layers.27.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
457
+ "language_model.model.layers.27.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
458
+ "language_model.model.layers.27.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
459
+ "language_model.model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
460
+ "language_model.model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
461
+ "language_model.model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
462
+ "language_model.model.layers.27.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
463
+ "language_model.model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
464
+ "language_model.model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
465
+ "language_model.model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
466
+ "language_model.model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
467
+ "language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
468
+ "language_model.model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
469
+ "language_model.model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
470
+ "language_model.model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
471
+ "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
472
+ "language_model.model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
473
+ "language_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
474
+ "language_model.model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
475
+ "language_model.model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
476
+ "language_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
477
+ "language_model.model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
478
+ "language_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
479
+ "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
480
+ "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
481
+ "language_model.model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
482
+ "language_model.model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
483
+ "language_model.model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
484
+ "language_model.model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
485
+ "language_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
486
+ "language_model.model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
487
+ "language_model.model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
488
+ "language_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
489
+ "language_model.model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
490
+ "language_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
491
+ "language_model.model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
492
+ "language_model.model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
493
+ "language_model.model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
494
+ "language_model.model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
495
+ "language_model.model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
496
+ "language_model.model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
497
+ "language_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
498
+ "language_model.model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
499
+ "language_model.model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
500
+ "language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
501
+ "language_model.model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
502
+ "language_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
503
+ "language_model.model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
504
+ "language_model.model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
505
+ "language_model.model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
506
+ "language_model.model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
507
+ "language_model.model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
508
+ "language_model.model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
509
+ "language_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
510
+ "language_model.model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
511
+ "language_model.model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
512
+ "language_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
513
+ "language_model.model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
514
+ "language_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
515
+ "language_model.model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
516
+ "language_model.model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
517
+ "language_model.model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
518
+ "language_model.model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
519
+ "language_model.model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
520
+ "language_model.model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
521
+ "language_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
522
+ "language_model.model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
523
+ "language_model.model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
524
+ "language_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
525
+ "language_model.model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
526
+ "language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
527
+ "language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
528
+ "language_model.model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
529
+ "language_model.model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
530
+ "language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
531
+ "language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
532
+ "language_model.model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
533
+ "language_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
534
+ "language_model.model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
535
+ "language_model.model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
536
+ "language_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
537
+ "language_model.model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
538
+ "language_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
539
+ "language_model.model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
540
+ "language_model.model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
541
+ "language_model.model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
542
+ "language_model.model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
543
+ "language_model.model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
544
+ "language_model.model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
545
+ "language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
546
+ "language_model.model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
547
+ "language_model.model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
548
+ "language_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
549
+ "language_model.model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
550
+ "language_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
551
+ "language_model.model.norm.weight": "model-00001-of-00002.safetensors",
552
+ "mlp1.0.bias": "model-00001-of-00002.safetensors",
553
+ "mlp1.0.weight": "model-00001-of-00002.safetensors",
554
+ "mlp1.1.bias": "model-00001-of-00002.safetensors",
555
+ "mlp1.1.weight": "model-00001-of-00002.safetensors",
556
+ "mlp1.3.bias": "model-00001-of-00002.safetensors",
557
+ "mlp1.3.weight": "model-00001-of-00002.safetensors",
558
+ "vision_model.vision_model.embeddings.patch_embedding.bias": "model-00001-of-00002.safetensors",
559
+ "vision_model.vision_model.embeddings.patch_embedding.weight": "model-00001-of-00002.safetensors",
560
+ "vision_model.vision_model.embeddings.position_embedding.weight": "model-00001-of-00002.safetensors",
561
+ "vision_model.vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00002.safetensors",
562
+ "vision_model.vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00002.safetensors",
563
+ "vision_model.vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00002.safetensors",
564
+ "vision_model.vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00002.safetensors",
565
+ "vision_model.vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00002.safetensors",
566
+ "vision_model.vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00002.safetensors",
567
+ "vision_model.vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00002.safetensors",
568
+ "vision_model.vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00002.safetensors",
569
+ "vision_model.vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
570
+ "vision_model.vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
571
+ "vision_model.vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
572
+ "vision_model.vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
573
+ "vision_model.vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
574
+ "vision_model.vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
575
+ "vision_model.vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
576
+ "vision_model.vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
577
+ "vision_model.vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00002.safetensors",
578
+ "vision_model.vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00002.safetensors",
579
+ "vision_model.vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00002.safetensors",
580
+ "vision_model.vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00002.safetensors",
581
+ "vision_model.vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00002.safetensors",
582
+ "vision_model.vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00002.safetensors",
583
+ "vision_model.vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00002.safetensors",
584
+ "vision_model.vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00002.safetensors",
585
+ "vision_model.vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
586
+ "vision_model.vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
587
+ "vision_model.vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
588
+ "vision_model.vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
589
+ "vision_model.vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
590
+ "vision_model.vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
591
+ "vision_model.vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
592
+ "vision_model.vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
593
+ "vision_model.vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00002.safetensors",
594
+ "vision_model.vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00002.safetensors",
595
+ "vision_model.vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00002.safetensors",
596
+ "vision_model.vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00002.safetensors",
597
+ "vision_model.vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00002.safetensors",
598
+ "vision_model.vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00002.safetensors",
599
+ "vision_model.vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00002.safetensors",
600
+ "vision_model.vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00002.safetensors",
601
+ "vision_model.vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
602
+ "vision_model.vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
603
+ "vision_model.vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
604
+ "vision_model.vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
605
+ "vision_model.vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
606
+ "vision_model.vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
607
+ "vision_model.vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
608
+ "vision_model.vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
609
+ "vision_model.vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00002.safetensors",
610
+ "vision_model.vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00002.safetensors",
611
+ "vision_model.vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00002.safetensors",
612
+ "vision_model.vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00002.safetensors",
613
+ "vision_model.vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00002.safetensors",
614
+ "vision_model.vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00002.safetensors",
615
+ "vision_model.vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00002.safetensors",
616
+ "vision_model.vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00002.safetensors",
617
+ "vision_model.vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
618
+ "vision_model.vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
619
+ "vision_model.vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
620
+ "vision_model.vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
621
+ "vision_model.vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
622
+ "vision_model.vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
623
+ "vision_model.vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
624
+ "vision_model.vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
625
+ "vision_model.vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00002.safetensors",
626
+ "vision_model.vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00002.safetensors",
627
+ "vision_model.vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00002.safetensors",
628
+ "vision_model.vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00002.safetensors",
629
+ "vision_model.vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00002.safetensors",
630
+ "vision_model.vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00002.safetensors",
631
+ "vision_model.vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00002.safetensors",
632
+ "vision_model.vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00002.safetensors",
633
+ "vision_model.vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
634
+ "vision_model.vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
635
+ "vision_model.vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
636
+ "vision_model.vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
637
+ "vision_model.vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
638
+ "vision_model.vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
639
+ "vision_model.vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
640
+ "vision_model.vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
641
+ "vision_model.vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00002.safetensors",
642
+ "vision_model.vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00002.safetensors",
643
+ "vision_model.vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00002.safetensors",
644
+ "vision_model.vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00002.safetensors",
645
+ "vision_model.vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00002.safetensors",
646
+ "vision_model.vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00002.safetensors",
647
+ "vision_model.vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00002.safetensors",
648
+ "vision_model.vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00002.safetensors",
649
+ "vision_model.vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
650
+ "vision_model.vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
651
+ "vision_model.vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
652
+ "vision_model.vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
653
+ "vision_model.vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
654
+ "vision_model.vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
655
+ "vision_model.vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
656
+ "vision_model.vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
657
+ "vision_model.vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00002.safetensors",
658
+ "vision_model.vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00002.safetensors",
659
+ "vision_model.vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00002.safetensors",
660
+ "vision_model.vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00002.safetensors",
661
+ "vision_model.vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00002.safetensors",
662
+ "vision_model.vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00002.safetensors",
663
+ "vision_model.vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00002.safetensors",
664
+ "vision_model.vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00002.safetensors",
665
+ "vision_model.vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
666
+ "vision_model.vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
667
+ "vision_model.vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
668
+ "vision_model.vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
669
+ "vision_model.vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
670
+ "vision_model.vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
671
+ "vision_model.vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
672
+ "vision_model.vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
673
+ "vision_model.vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00002.safetensors",
674
+ "vision_model.vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00002.safetensors",
675
+ "vision_model.vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00002.safetensors",
676
+ "vision_model.vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00002.safetensors",
677
+ "vision_model.vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00002.safetensors",
678
+ "vision_model.vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00002.safetensors",
679
+ "vision_model.vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00002.safetensors",
680
+ "vision_model.vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00002.safetensors",
681
+ "vision_model.vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
682
+ "vision_model.vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
683
+ "vision_model.vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
684
+ "vision_model.vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
685
+ "vision_model.vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
686
+ "vision_model.vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
687
+ "vision_model.vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
688
+ "vision_model.vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
689
+ "vision_model.vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00002.safetensors",
690
+ "vision_model.vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00002.safetensors",
691
+ "vision_model.vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00002.safetensors",
692
+ "vision_model.vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00002.safetensors",
693
+ "vision_model.vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00002.safetensors",
694
+ "vision_model.vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00002.safetensors",
695
+ "vision_model.vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00002.safetensors",
696
+ "vision_model.vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00002.safetensors",
697
+ "vision_model.vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
698
+ "vision_model.vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
699
+ "vision_model.vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
700
+ "vision_model.vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
701
+ "vision_model.vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
702
+ "vision_model.vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
703
+ "vision_model.vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
704
+ "vision_model.vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
705
+ "vision_model.vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00002.safetensors",
706
+ "vision_model.vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00002.safetensors",
707
+ "vision_model.vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00002.safetensors",
708
+ "vision_model.vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00002.safetensors",
709
+ "vision_model.vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00002.safetensors",
710
+ "vision_model.vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00002.safetensors",
711
+ "vision_model.vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00002.safetensors",
712
+ "vision_model.vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00002.safetensors",
713
+ "vision_model.vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
714
+ "vision_model.vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
715
+ "vision_model.vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
716
+ "vision_model.vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
717
+ "vision_model.vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
718
+ "vision_model.vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
719
+ "vision_model.vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
720
+ "vision_model.vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
721
+ "vision_model.vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00002.safetensors",
722
+ "vision_model.vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00002.safetensors",
723
+ "vision_model.vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00002.safetensors",
724
+ "vision_model.vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00002.safetensors",
725
+ "vision_model.vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00002.safetensors",
726
+ "vision_model.vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00002.safetensors",
727
+ "vision_model.vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00002.safetensors",
728
+ "vision_model.vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00002.safetensors",
729
+ "vision_model.vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
730
+ "vision_model.vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
731
+ "vision_model.vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
732
+ "vision_model.vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
733
+ "vision_model.vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
734
+ "vision_model.vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
735
+ "vision_model.vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
736
+ "vision_model.vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
737
+ "vision_model.vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00002.safetensors",
738
+ "vision_model.vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00002.safetensors",
739
+ "vision_model.vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00002.safetensors",
740
+ "vision_model.vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00002.safetensors",
741
+ "vision_model.vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00002.safetensors",
742
+ "vision_model.vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00002.safetensors",
743
+ "vision_model.vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00002.safetensors",
744
+ "vision_model.vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00002.safetensors",
745
+ "vision_model.vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
746
+ "vision_model.vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
747
+ "vision_model.vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
748
+ "vision_model.vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
749
+ "vision_model.vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
750
+ "vision_model.vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
751
+ "vision_model.vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
752
+ "vision_model.vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
753
+ "vision_model.vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00002.safetensors",
754
+ "vision_model.vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00002.safetensors",
755
+ "vision_model.vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00002.safetensors",
756
+ "vision_model.vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00002.safetensors",
757
+ "vision_model.vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00002.safetensors",
758
+ "vision_model.vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00002.safetensors",
759
+ "vision_model.vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00002.safetensors",
760
+ "vision_model.vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00002.safetensors",
761
+ "vision_model.vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
762
+ "vision_model.vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
763
+ "vision_model.vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
764
+ "vision_model.vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
765
+ "vision_model.vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
766
+ "vision_model.vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
767
+ "vision_model.vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
768
+ "vision_model.vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
769
+ "vision_model.vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00002.safetensors",
770
+ "vision_model.vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00002.safetensors",
771
+ "vision_model.vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00002.safetensors",
772
+ "vision_model.vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00002.safetensors",
773
+ "vision_model.vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00002.safetensors",
774
+ "vision_model.vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00002.safetensors",
775
+ "vision_model.vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00002.safetensors",
776
+ "vision_model.vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00002.safetensors",
777
+ "vision_model.vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
778
+ "vision_model.vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
779
+ "vision_model.vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
780
+ "vision_model.vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
781
+ "vision_model.vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
782
+ "vision_model.vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
783
+ "vision_model.vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
784
+ "vision_model.vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
785
+ "vision_model.vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00002.safetensors",
786
+ "vision_model.vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00002.safetensors",
787
+ "vision_model.vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00002.safetensors",
788
+ "vision_model.vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00002.safetensors",
789
+ "vision_model.vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00002.safetensors",
790
+ "vision_model.vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00002.safetensors",
791
+ "vision_model.vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00002.safetensors",
792
+ "vision_model.vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00002.safetensors",
793
+ "vision_model.vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
794
+ "vision_model.vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
795
+ "vision_model.vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
796
+ "vision_model.vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
797
+ "vision_model.vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
798
+ "vision_model.vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
799
+ "vision_model.vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
800
+ "vision_model.vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
801
+ "vision_model.vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00002.safetensors",
802
+ "vision_model.vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00002.safetensors",
803
+ "vision_model.vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00002.safetensors",
804
+ "vision_model.vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00002.safetensors",
805
+ "vision_model.vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00002.safetensors",
806
+ "vision_model.vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00002.safetensors",
807
+ "vision_model.vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00002.safetensors",
808
+ "vision_model.vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00002.safetensors",
809
+ "vision_model.vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
810
+ "vision_model.vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
811
+ "vision_model.vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
812
+ "vision_model.vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
813
+ "vision_model.vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
814
+ "vision_model.vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
815
+ "vision_model.vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
816
+ "vision_model.vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
817
+ "vision_model.vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00002.safetensors",
818
+ "vision_model.vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00002.safetensors",
819
+ "vision_model.vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00002.safetensors",
820
+ "vision_model.vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00002.safetensors",
821
+ "vision_model.vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00002.safetensors",
822
+ "vision_model.vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00002.safetensors",
823
+ "vision_model.vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00002.safetensors",
824
+ "vision_model.vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00002.safetensors",
825
+ "vision_model.vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
826
+ "vision_model.vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
827
+ "vision_model.vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
828
+ "vision_model.vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
829
+ "vision_model.vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
830
+ "vision_model.vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
831
+ "vision_model.vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
832
+ "vision_model.vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
833
+ "vision_model.vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00002.safetensors",
834
+ "vision_model.vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00002.safetensors",
835
+ "vision_model.vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00002.safetensors",
836
+ "vision_model.vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00002.safetensors",
837
+ "vision_model.vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00002.safetensors",
838
+ "vision_model.vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00002.safetensors",
839
+ "vision_model.vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00002.safetensors",
840
+ "vision_model.vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00002.safetensors",
841
+ "vision_model.vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
842
+ "vision_model.vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
843
+ "vision_model.vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
844
+ "vision_model.vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
845
+ "vision_model.vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
846
+ "vision_model.vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
847
+ "vision_model.vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
848
+ "vision_model.vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
849
+ "vision_model.vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00002.safetensors",
850
+ "vision_model.vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00002.safetensors",
851
+ "vision_model.vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00002.safetensors",
852
+ "vision_model.vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00002.safetensors",
853
+ "vision_model.vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00002.safetensors",
854
+ "vision_model.vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00002.safetensors",
855
+ "vision_model.vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00002.safetensors",
856
+ "vision_model.vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00002.safetensors",
857
+ "vision_model.vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
858
+ "vision_model.vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
859
+ "vision_model.vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
860
+ "vision_model.vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
861
+ "vision_model.vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
862
+ "vision_model.vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
863
+ "vision_model.vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
864
+ "vision_model.vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
865
+ "vision_model.vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00002.safetensors",
866
+ "vision_model.vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00002.safetensors",
867
+ "vision_model.vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00002.safetensors",
868
+ "vision_model.vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00002.safetensors",
869
+ "vision_model.vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00002.safetensors",
870
+ "vision_model.vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00002.safetensors",
871
+ "vision_model.vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00002.safetensors",
872
+ "vision_model.vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00002.safetensors",
873
+ "vision_model.vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
874
+ "vision_model.vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
875
+ "vision_model.vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
876
+ "vision_model.vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
877
+ "vision_model.vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
878
+ "vision_model.vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
879
+ "vision_model.vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
880
+ "vision_model.vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
881
+ "vision_model.vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00002.safetensors",
882
+ "vision_model.vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00002.safetensors",
883
+ "vision_model.vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00002.safetensors",
884
+ "vision_model.vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00002.safetensors",
885
+ "vision_model.vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00002.safetensors",
886
+ "vision_model.vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00002.safetensors",
887
+ "vision_model.vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00002.safetensors",
888
+ "vision_model.vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00002.safetensors",
889
+ "vision_model.vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
890
+ "vision_model.vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
891
+ "vision_model.vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
892
+ "vision_model.vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
893
+ "vision_model.vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
894
+ "vision_model.vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
895
+ "vision_model.vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
896
+ "vision_model.vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
897
+ "vision_model.vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00002.safetensors",
898
+ "vision_model.vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00002.safetensors",
899
+ "vision_model.vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00002.safetensors",
900
+ "vision_model.vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00002.safetensors",
901
+ "vision_model.vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00002.safetensors",
902
+ "vision_model.vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00002.safetensors",
903
+ "vision_model.vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00002.safetensors",
904
+ "vision_model.vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00002.safetensors",
905
+ "vision_model.vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
906
+ "vision_model.vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
907
+ "vision_model.vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
908
+ "vision_model.vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
909
+ "vision_model.vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
910
+ "vision_model.vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
911
+ "vision_model.vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
912
+ "vision_model.vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
913
+ "vision_model.vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00002.safetensors",
914
+ "vision_model.vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00002.safetensors",
915
+ "vision_model.vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00002.safetensors",
916
+ "vision_model.vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00002.safetensors",
917
+ "vision_model.vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00002.safetensors",
918
+ "vision_model.vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00002.safetensors",
919
+ "vision_model.vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00002.safetensors",
920
+ "vision_model.vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00002.safetensors",
921
+ "vision_model.vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
922
+ "vision_model.vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
923
+ "vision_model.vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
924
+ "vision_model.vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
925
+ "vision_model.vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
926
+ "vision_model.vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
927
+ "vision_model.vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
928
+ "vision_model.vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
929
+ "vision_model.vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00002.safetensors",
930
+ "vision_model.vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00002.safetensors",
931
+ "vision_model.vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00002.safetensors",
932
+ "vision_model.vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00002.safetensors",
933
+ "vision_model.vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00002.safetensors",
934
+ "vision_model.vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00002.safetensors",
935
+ "vision_model.vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00002.safetensors",
936
+ "vision_model.vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00002.safetensors",
937
+ "vision_model.vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
938
+ "vision_model.vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
939
+ "vision_model.vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
940
+ "vision_model.vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
941
+ "vision_model.vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
942
+ "vision_model.vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
943
+ "vision_model.vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
944
+ "vision_model.vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
945
+ "vision_model.vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00002.safetensors",
946
+ "vision_model.vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00002.safetensors",
947
+ "vision_model.vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00002.safetensors",
948
+ "vision_model.vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00002.safetensors",
949
+ "vision_model.vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00002.safetensors",
950
+ "vision_model.vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00002.safetensors",
951
+ "vision_model.vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00002.safetensors",
952
+ "vision_model.vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00002.safetensors",
953
+ "vision_model.vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
954
+ "vision_model.vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
955
+ "vision_model.vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
956
+ "vision_model.vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
957
+ "vision_model.vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
958
+ "vision_model.vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
959
+ "vision_model.vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
960
+ "vision_model.vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
961
+ "vision_model.vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00002.safetensors",
962
+ "vision_model.vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00002.safetensors",
963
+ "vision_model.vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00002.safetensors",
964
+ "vision_model.vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00002.safetensors",
965
+ "vision_model.vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00002.safetensors",
966
+ "vision_model.vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00002.safetensors",
967
+ "vision_model.vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00002.safetensors",
968
+ "vision_model.vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00002.safetensors",
969
+ "vision_model.vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
970
+ "vision_model.vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
971
+ "vision_model.vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
972
+ "vision_model.vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
973
+ "vision_model.vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
974
+ "vision_model.vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
975
+ "vision_model.vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
976
+ "vision_model.vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
977
+ "vision_model.vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00002.safetensors",
978
+ "vision_model.vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00002.safetensors",
979
+ "vision_model.vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00002.safetensors",
980
+ "vision_model.vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00002.safetensors",
981
+ "vision_model.vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00002.safetensors",
982
+ "vision_model.vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00002.safetensors",
983
+ "vision_model.vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00002.safetensors",
984
+ "vision_model.vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00002.safetensors",
985
+ "vision_model.vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
986
+ "vision_model.vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
987
+ "vision_model.vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
988
+ "vision_model.vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
989
+ "vision_model.vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
990
+ "vision_model.vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
991
+ "vision_model.vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
992
+ "vision_model.vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
993
+ "vision_model.vision_model.post_layernorm.bias": "model-00001-of-00002.safetensors",
994
+ "vision_model.vision_model.post_layernorm.weight": "model-00001-of-00002.safetensors"
995
+ }
996
+ }
training_states.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76b560f3d0f446afbaae3d97a7eba5e4ca4d785fdd391afa8f7fe72ab45760b4
3
+ size 8672098091