Update config.json
Browse files- config.json +2 -2
config.json
CHANGED
|
@@ -9,7 +9,7 @@
|
|
| 9 |
"attention_bias": false,
|
| 10 |
"attention_dropout": 0.0,
|
| 11 |
"attn_output_gate": true,
|
| 12 |
-
"dtype": "
|
| 13 |
"eos_token_id": 248044,
|
| 14 |
"full_attention_interval": 4,
|
| 15 |
"head_dim": 256,
|
|
@@ -116,6 +116,7 @@
|
|
| 116 |
"transformers_version": "4.57.0.dev0",
|
| 117 |
"video_token_id": 248057,
|
| 118 |
"vision_config": {
|
|
|
|
| 119 |
"deepstack_visual_indexes": [],
|
| 120 |
"depth": 27,
|
| 121 |
"hidden_act": "gelu_pytorch_tanh",
|
|
@@ -133,7 +134,6 @@
|
|
| 133 |
},
|
| 134 |
"vision_end_token_id": 248054,
|
| 135 |
"vision_start_token_id": 248053,
|
| 136 |
-
"torch_dtype": "bfloat16",
|
| 137 |
"quantization_config": {
|
| 138 |
"quant_method": "awq",
|
| 139 |
"bits": 4,
|
|
|
|
| 9 |
"attention_bias": false,
|
| 10 |
"attention_dropout": 0.0,
|
| 11 |
"attn_output_gate": true,
|
| 12 |
+
"dtype": "float16",
|
| 13 |
"eos_token_id": 248044,
|
| 14 |
"full_attention_interval": 4,
|
| 15 |
"head_dim": 256,
|
|
|
|
| 116 |
"transformers_version": "4.57.0.dev0",
|
| 117 |
"video_token_id": 248057,
|
| 118 |
"vision_config": {
|
| 119 |
+
"torch_dtype": "float16",
|
| 120 |
"deepstack_visual_indexes": [],
|
| 121 |
"depth": 27,
|
| 122 |
"hidden_act": "gelu_pytorch_tanh",
|
|
|
|
| 134 |
},
|
| 135 |
"vision_end_token_id": 248054,
|
| 136 |
"vision_start_token_id": 248053,
|
|
|
|
| 137 |
"quantization_config": {
|
| 138 |
"quant_method": "awq",
|
| 139 |
"bits": 4,
|