Spaces:
Sleeping
Sleeping
Upload inference.py
Browse files- inference.py +3 -8
inference.py
CHANGED
|
@@ -42,12 +42,6 @@ class UNetNoCondWrapper(nn.Module):
|
|
| 42 |
|
| 43 |
def inference(model_id,device, img1, img2):
|
| 44 |
|
| 45 |
-
|
| 46 |
-
# vae = AutoencoderKL.from_pretrained(f"{model_id}/vae").to(device)
|
| 47 |
-
# scheduler = DDPMScheduler.from_pretrained(f"{model_id}/scheduler")
|
| 48 |
-
# tokenizer = CLIPTokenizer.from_pretrained(f"{model_id}/tokenizer")
|
| 49 |
-
# text_encoder = CLIPTextModel.from_pretrained(f"{model_id}/text_encoder").to(device)
|
| 50 |
-
# feature_extractor = CLIPImageProcessor.from_pretrained(f"{model_id}/feature_extractor")
|
| 51 |
|
| 52 |
vae = AutoencoderKL.from_pretrained(model_id, subfolder="vae").to(device)
|
| 53 |
scheduler = DDPMScheduler.from_pretrained(model_id, subfolder="scheduler")
|
|
@@ -69,9 +63,10 @@ def inference(model_id,device, img1, img2):
|
|
| 69 |
safety_checker=None,
|
| 70 |
feature_extractor=feature_extractor,
|
| 71 |
)
|
| 72 |
-
pipe = pipe.to(torch.float16).to(device)
|
|
|
|
| 73 |
|
| 74 |
-
generator = torch.Generator("
|
| 75 |
|
| 76 |
|
| 77 |
img1 = img1.resize((512, 512))
|
|
|
|
| 42 |
|
| 43 |
def inference(model_id,device, img1, img2):
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
|
| 46 |
vae = AutoencoderKL.from_pretrained(model_id, subfolder="vae").to(device)
|
| 47 |
scheduler = DDPMScheduler.from_pretrained(model_id, subfolder="scheduler")
|
|
|
|
| 63 |
safety_checker=None,
|
| 64 |
feature_extractor=feature_extractor,
|
| 65 |
)
|
| 66 |
+
# pipe = pipe.to(torch.float16).to(device)
|
| 67 |
+
pipe = pipe.to(torch.float32).to(device)
|
| 68 |
|
| 69 |
+
generator = torch.Generator("cpu").manual_seed(0)
|
| 70 |
|
| 71 |
|
| 72 |
img1 = img1.resize((512, 512))
|