Update
Browse files- config.json +2 -2
config.json
CHANGED
|
@@ -205,7 +205,7 @@
|
|
| 205 |
"num_hidden_layers": 6,
|
| 206 |
"num_attention_heads": 16,
|
| 207 |
"hidden_act": "gelu",
|
| 208 |
-
"intermediate_size":
|
| 209 |
"hidden_dropout_prob": 0.1,
|
| 210 |
"attention_probs_dropout_prob": 0.1,
|
| 211 |
"max_position_embeddings": 512,
|
|
@@ -250,4 +250,4 @@
|
|
| 250 |
"initializer_factor": 1.0,
|
| 251 |
"initializer_range": 0.02,
|
| 252 |
"model_type": "mplug-owl"
|
| 253 |
-
}
|
|
|
|
| 205 |
"num_hidden_layers": 6,
|
| 206 |
"num_attention_heads": 16,
|
| 207 |
"hidden_act": "gelu",
|
| 208 |
+
"intermediate_size": 2816,
|
| 209 |
"hidden_dropout_prob": 0.1,
|
| 210 |
"attention_probs_dropout_prob": 0.1,
|
| 211 |
"max_position_embeddings": 512,
|
|
|
|
| 250 |
"initializer_factor": 1.0,
|
| 251 |
"initializer_range": 0.02,
|
| 252 |
"model_type": "mplug-owl"
|
| 253 |
+
}
|