Upload src/pipeline.py with huggingface_hub
Browse files- src/pipeline.py +3 -3
src/pipeline.py
CHANGED
|
@@ -49,10 +49,10 @@ def load_pipeline() -> Pipeline:
|
|
| 49 |
|
| 50 |
pipeline.transformer.to(memory_format=torch.channels_last)
|
| 51 |
pipeline.transformer = torch.compile(pipeline.transformer, mode="max-autotune", fullgraph=True)
|
| 52 |
-
|
| 53 |
# quantize_(pipeline.vae, int8_weight_only())
|
| 54 |
-
|
| 55 |
-
|
| 56 |
PROMPT = 'semiconformity, peregrination, quip, twineless, emotionless, tawa, depickle'
|
| 57 |
with torch.inference_mode():
|
| 58 |
for _ in range(4):
|
|
|
|
| 49 |
|
| 50 |
pipeline.transformer.to(memory_format=torch.channels_last)
|
| 51 |
pipeline.transformer = torch.compile(pipeline.transformer, mode="max-autotune", fullgraph=True)
|
| 52 |
+
pipeline.vae.to(memory_format=torch.channels_last)
|
| 53 |
# quantize_(pipeline.vae, int8_weight_only())
|
| 54 |
+
pipeline.vae = torch.compile(pipeline.vae, fullgraph=True, mode="max-autotune")
|
| 55 |
+
|
| 56 |
PROMPT = 'semiconformity, peregrination, quip, twineless, emotionless, tawa, depickle'
|
| 57 |
with torch.inference_mode():
|
| 58 |
for _ in range(4):
|