Commit
·
e89ddfe
1
Parent(s):
48184f0
Update transformer/config.json
Browse files- transformer/config.json +7 -14
transformer/config.json
CHANGED
|
@@ -2,10 +2,9 @@
|
|
| 2 |
"_class_name": "Transformer2DModel",
|
| 3 |
"_diffusers_version": "0.7.0.dev0",
|
| 4 |
"attention_bias": true,
|
| 5 |
-
"
|
| 6 |
-
"
|
| 7 |
-
"
|
| 8 |
-
"discrete": true,
|
| 9 |
"dropout": 0.0,
|
| 10 |
"ff_layers": [
|
| 11 |
"Linear",
|
|
@@ -13,16 +12,10 @@
|
|
| 13 |
"Linear",
|
| 14 |
"Dropout"
|
| 15 |
],
|
| 16 |
-
"height": 32,
|
| 17 |
"in_channels": null,
|
| 18 |
-
"
|
| 19 |
-
"
|
| 20 |
-
"AdaLayerNorm",
|
| 21 |
-
"AdaLayerNorm",
|
| 22 |
-
"LayerNorm"
|
| 23 |
-
],
|
| 24 |
-
"num_embed": 4097,
|
| 25 |
"num_embeds_ada_norm": 100,
|
| 26 |
-
"
|
| 27 |
-
"
|
| 28 |
}
|
|
|
|
| 2 |
"_class_name": "Transformer2DModel",
|
| 3 |
"_diffusers_version": "0.7.0.dev0",
|
| 4 |
"attention_bias": true,
|
| 5 |
+
"cross_attention_dim": 512,
|
| 6 |
+
"attention_head_dim": 88,
|
| 7 |
+
"num_layers": 36,
|
|
|
|
| 8 |
"dropout": 0.0,
|
| 9 |
"ff_layers": [
|
| 10 |
"Linear",
|
|
|
|
| 12 |
"Linear",
|
| 13 |
"Dropout"
|
| 14 |
],
|
|
|
|
| 15 |
"in_channels": null,
|
| 16 |
+
"num_attentinon_heads": 16,
|
| 17 |
+
"num_vector_embeds": 4097,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
"num_embeds_ada_norm": 100,
|
| 19 |
+
"norm_num_groups": 32,
|
| 20 |
+
"sample_size": 32
|
| 21 |
}
|