Update config.json
Browse files- config.json +15 -19
config.json
CHANGED
|
@@ -1,19 +1,15 @@
|
|
| 1 |
-
|
| 2 |
-
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
| 16 |
-
buf = io.BytesIO()
|
| 17 |
-
image.save(buf, format='PNG')
|
| 18 |
-
byte_im = buf.getvalue()
|
| 19 |
-
return {"image_bytes": byte_im}
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": ["UNet2DConditionModel"],
|
| 3 |
+
"_class_name": "StableDiffusionPipeline",
|
| 4 |
+
"_diffusers_version": "0.21.4",
|
| 5 |
+
"feature_extractor": [
|
| 6 |
+
"CLIPFeatureExtractor",
|
| 7 |
+
"openai/clip-vit-base-patch32"
|
| 8 |
+
],
|
| 9 |
+
"safety_checker": null,
|
| 10 |
+
"scheduler": ["PNDMScheduler"],
|
| 11 |
+
"text_encoder": ["CLIPTextModel", "openai/clip-vit-base-patch32"],
|
| 12 |
+
"tokenizer": ["CLIPTokenizer", "openai/clip-vit-base-patch32"],
|
| 13 |
+
"unet": ["UNet2DConditionModel"],
|
| 14 |
+
"vae": ["AutoencoderKL"]
|
| 15 |
+
}
|
|
|
|
|
|
|
|
|
|
|
|