Upload config.json with huggingface_hub
Browse files- config.json +4 -4
config.json
CHANGED
|
@@ -2,12 +2,12 @@
|
|
| 2 |
"hidden_size": 512,
|
| 3 |
"num_attention_heads": 4,
|
| 4 |
"attention_head_dim": 128,
|
| 5 |
-
"
|
| 6 |
-
"num_double_layers": 15,
|
| 7 |
-
"mlp_ratio": 4.0,
|
| 8 |
"joint_attention_dim": 768,
|
| 9 |
"pooled_projection_dim": 768,
|
| 10 |
-
"
|
|
|
|
|
|
|
| 11 |
"axes_dims_rope": [
|
| 12 |
16,
|
| 13 |
56,
|
|
|
|
| 2 |
"hidden_size": 512,
|
| 3 |
"num_attention_heads": 4,
|
| 4 |
"attention_head_dim": 128,
|
| 5 |
+
"in_channels": 16,
|
|
|
|
|
|
|
| 6 |
"joint_attention_dim": 768,
|
| 7 |
"pooled_projection_dim": 768,
|
| 8 |
+
"num_double_layers": 15,
|
| 9 |
+
"num_single_layers": 25,
|
| 10 |
+
"mlp_ratio": 4.0,
|
| 11 |
"axes_dims_rope": [
|
| 12 |
16,
|
| 13 |
56,
|