import torch
from diffusers import DiffusionPipeline
# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Heasterian/AsymmetricAutoencoderKLUpscaler_v2", dtype=torch.bfloat16, device_map="cuda")
prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]Another take on upscaler using AsymmetricAutoencoderKL, in this case I did re-use decoder of ostris/vae-kl-f8-d16 as decoder and trained just encoder.
Oversharpening was reduced by usage of sobel in loss calculation. I guess that without high resolution image used as reference, it might look more blury than previous one.
- Downloads last month
- 10
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for Heasterian/AsymmetricAutoencoderKLUpscaler_v2
Base model
ostris/vae-kl-f8-d16