Upload Blt model converted
Browse files- config.json +2 -2
config.json
CHANGED
|
@@ -42,7 +42,7 @@
|
|
| 42 |
"vocab_size": 260,
|
| 43 |
"cross_attn_all_layers": false,
|
| 44 |
"cross_attn_k": 4,
|
| 45 |
-
"hidden_size_global":
|
| 46 |
"pm_size": 0,
|
| 47 |
"hidden_size": 1280,
|
| 48 |
"num_attention_heads": 20,
|
|
@@ -63,7 +63,7 @@
|
|
| 63 |
"vocab_size": 260,
|
| 64 |
"cross_attn_all_layers": true,
|
| 65 |
"cross_attn_k": 4,
|
| 66 |
-
"hidden_size_global":
|
| 67 |
"hidden_size": 1280,
|
| 68 |
"num_attention_heads": 20,
|
| 69 |
"num_key_value_heads": null,
|
|
|
|
| 42 |
"vocab_size": 260,
|
| 43 |
"cross_attn_all_layers": false,
|
| 44 |
"cross_attn_k": 4,
|
| 45 |
+
"hidden_size_global": 4096,
|
| 46 |
"pm_size": 0,
|
| 47 |
"hidden_size": 1280,
|
| 48 |
"num_attention_heads": 20,
|
|
|
|
| 63 |
"vocab_size": 260,
|
| 64 |
"cross_attn_all_layers": true,
|
| 65 |
"cross_attn_k": 4,
|
| 66 |
+
"hidden_size_global": 4096,
|
| 67 |
"hidden_size": 1280,
|
| 68 |
"num_attention_heads": 20,
|
| 69 |
"num_key_value_heads": null,
|