How to use from the
Use from the
Diffusers library
pip install -U diffusers transformers accelerate
import torch
from diffusers import DiffusionPipeline

# switch to "mps" for apple devices
pipe = DiffusionPipeline.from_pretrained("Gokuldaskumar/AAM_XL_AnimeMix", dtype=torch.bfloat16, device_map="cuda")

prompt = "Astronaut in a jungle, cold color palette, muted colors, detailed, 8k"
image = pipe(prompt).images[0]

AAM XL AnimeMix

Lykon/AAM_XL_AnimeMix is a Stable Diffusion model that has been fine-tuned on stabilityai/stable-diffusion-xl-base-1.0.

Please consider supporting me:

License: Fair AI Public License 1.0-SD

Diffusers

For more general information on how to run text-to-image models with 🧨 Diffusers, see the docs.

  1. Installation
pip install diffusers transformers accelerate
  1. Run
from diffusers import AutoPipelineForText2Image, DEISMultistepScheduler
import torch

pipe = AutoPipelineForText2Image.from_pretrained('Lykon/AAM_XL_AnimeMix', torch_dtype=torch.float16, variant="fp16")
pipe.scheduler = DEISMultistepScheduler.from_config(pipe.scheduler.config)
pipe = pipe.to("cuda")

prompt = "anime girl, night, blue light behind her, ((Galaxy, Lens flare)), short hair, flower field, night sky, cinematic shot. Wallpaper. (Blue color schema), detailed background, a city in the distance"

generator = torch.manual_seed(0)
image = pipe(prompt, num_inference_steps=25).images[0]  
image.save("./image.png")

Downloads last month
4
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support