Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,10 +5,11 @@ import spaces
|
|
| 5 |
import torch
|
| 6 |
from diffusers import DiffusionPipeline, AutoencoderKL, UNet2DConditionModel, FlowMatchEulerDiscreteScheduler
|
| 7 |
from transformers import AutoTokenizer, AutoModel
|
|
|
|
| 8 |
|
| 9 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 10 |
dtype = torch.float16 if torch.cuda.is_available() else torch.float32
|
| 11 |
-
model_repo_id = "AiArtLab/
|
| 12 |
max_length = 150
|
| 13 |
|
| 14 |
class SimpleDiffusionPipeline(DiffusionPipeline):
|
|
@@ -94,18 +95,25 @@ class SimpleDiffusionPipeline(DiffusionPipeline):
|
|
| 94 |
|
| 95 |
return images
|
| 96 |
|
| 97 |
-
vae = AutoencoderKL.from_pretrained(model_repo_id, subfolder="vae", torch_dtype=dtype).to(device)
|
| 98 |
-
unet = UNet2DConditionModel.from_pretrained(model_repo_id, subfolder="unet", torch_dtype=dtype).to(device)
|
| 99 |
-
tokenizer = AutoTokenizer.from_pretrained(model_repo_id, subfolder="tokenizer")
|
| 100 |
-
text_encoder = AutoModel.from_pretrained(model_repo_id, subfolder="text_encoder", torch_dtype=dtype).to(device)
|
| 101 |
-
scheduler = FlowMatchEulerDiscreteScheduler.from_pretrained(model_repo_id, subfolder="scheduler")
|
| 102 |
-
|
| 103 |
-
pipe = SimpleDiffusionPipeline(
|
| 104 |
-
vae=vae,
|
| 105 |
-
text_encoder=text_encoder,
|
| 106 |
-
tokenizer=tokenizer,
|
| 107 |
-
unet=unet,
|
| 108 |
-
scheduler=scheduler,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 109 |
).to(device)
|
| 110 |
|
| 111 |
MAX_SEED = np.iinfo(np.int32).max
|
|
|
|
| 5 |
import torch
|
| 6 |
from diffusers import DiffusionPipeline, AutoencoderKL, UNet2DConditionModel, FlowMatchEulerDiscreteScheduler
|
| 7 |
from transformers import AutoTokenizer, AutoModel
|
| 8 |
+
from diffusers import DiffusionPipeline
|
| 9 |
|
| 10 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 11 |
dtype = torch.float16 if torch.cuda.is_available() else torch.float32
|
| 12 |
+
model_repo_id = "AiArtLab/sdxs"
|
| 13 |
max_length = 150
|
| 14 |
|
| 15 |
class SimpleDiffusionPipeline(DiffusionPipeline):
|
|
|
|
| 95 |
|
| 96 |
return images
|
| 97 |
|
| 98 |
+
#vae = AutoencoderKL.from_pretrained(model_repo_id, subfolder="vae", torch_dtype=dtype).to(device)
|
| 99 |
+
#unet = UNet2DConditionModel.from_pretrained(model_repo_id, subfolder="unet", torch_dtype=dtype).to(device)
|
| 100 |
+
#tokenizer = AutoTokenizer.from_pretrained(model_repo_id, subfolder="tokenizer")
|
| 101 |
+
#text_encoder = AutoModel.from_pretrained(model_repo_id, subfolder="text_encoder", torch_dtype=dtype).to(device)
|
| 102 |
+
#scheduler = FlowMatchEulerDiscreteScheduler.from_pretrained(model_repo_id, subfolder="scheduler")
|
| 103 |
+
|
| 104 |
+
#pipe = SimpleDiffusionPipeline(
|
| 105 |
+
# vae=vae,
|
| 106 |
+
# text_encoder=text_encoder,
|
| 107 |
+
# tokenizer=tokenizer,
|
| 108 |
+
# unet=unet,
|
| 109 |
+
# scheduler=scheduler,
|
| 110 |
+
#).to(device)
|
| 111 |
+
|
| 112 |
+
pipe_id = "AiArtLab/sdxs"
|
| 113 |
+
pipe = SdxsPipeline.from_pretrained(
|
| 114 |
+
pipe_id,
|
| 115 |
+
torch_dtype=dtype,
|
| 116 |
+
trust_remote_code=True
|
| 117 |
).to(device)
|
| 118 |
|
| 119 |
MAX_SEED = np.iinfo(np.int32).max
|