| - sections: |
| - local: index |
| title: 🧨 Diffusers |
| - local: quicktour |
| title: Quicktour |
| - local: stable_diffusion |
| title: Effective and efficient diffusion |
| - local: installation |
| title: Installation |
| title: Get started |
| - sections: |
| - local: tutorials/tutorial_overview |
| title: Overview |
| - local: using-diffusers/write_own_pipeline |
| title: Understanding pipelines, models and schedulers |
| - local: tutorials/autopipeline |
| title: AutoPipeline |
| - local: tutorials/basic_training |
| title: Train a diffusion model |
| - local: tutorials/using_peft_for_inference |
| title: Load LoRAs for inference |
| - local: tutorials/fast_diffusion |
| title: Accelerate inference of text-to-image diffusion models |
| - local: tutorials/inference_with_big_models |
| title: Working with big models |
| title: Tutorials |
| - sections: |
| - local: using-diffusers/loading |
| title: Load pipelines |
| - local: using-diffusers/custom_pipeline_overview |
| title: Load community pipelines and components |
| - local: using-diffusers/schedulers |
| title: Load schedulers and models |
| - local: using-diffusers/other-formats |
| title: Model files and layouts |
| - local: using-diffusers/loading_adapters |
| title: Load adapters |
| - local: using-diffusers/push_to_hub |
| title: Push files to the Hub |
| title: Load pipelines and adapters |
| - sections: |
| - local: using-diffusers/unconditional_image_generation |
| title: Unconditional image generation |
| - local: using-diffusers/conditional_image_generation |
| title: Text-to-image |
| - local: using-diffusers/img2img |
| title: Image-to-image |
| - local: using-diffusers/inpaint |
| title: Inpainting |
| - local: using-diffusers/text-img2vid |
| title: Text or image-to-video |
| - local: using-diffusers/depth2img |
| title: Depth-to-image |
| title: Generative tasks |
| - sections: |
| - local: using-diffusers/overview_techniques |
| title: Overview |
| - local: training/distributed_inference |
| title: Distributed inference |
| - local: using-diffusers/merge_loras |
| title: Merge LoRAs |
| - local: using-diffusers/scheduler_features |
| title: Scheduler features |
| - local: using-diffusers/callback |
| title: Pipeline callbacks |
| - local: using-diffusers/reusing_seeds |
| title: Reproducible pipelines |
| - local: using-diffusers/image_quality |
| title: Controlling image quality |
| - local: using-diffusers/weighted_prompts |
| title: Prompt techniques |
| title: Inference techniques |
| - sections: |
| - local: advanced_inference/outpaint |
| title: Outpainting |
| title: Advanced inference |
| - sections: |
| - local: using-diffusers/sdxl |
| title: Stable Diffusion XL |
| - local: using-diffusers/sdxl_turbo |
| title: SDXL Turbo |
| - local: using-diffusers/kandinsky |
| title: Kandinsky |
| - local: using-diffusers/ip_adapter |
| title: IP-Adapter |
| - local: using-diffusers/pag |
| title: PAG |
| - local: using-diffusers/controlnet |
| title: ControlNet |
| - local: using-diffusers/t2i_adapter |
| title: T2I-Adapter |
| - local: using-diffusers/inference_with_lcm |
| title: Latent Consistency Model |
| - local: using-diffusers/textual_inversion_inference |
| title: Textual inversion |
| - local: using-diffusers/shap-e |
| title: Shap-E |
| - local: using-diffusers/diffedit |
| title: DiffEdit |
| - local: using-diffusers/inference_with_tcd_lora |
| title: Trajectory Consistency Distillation-LoRA |
| - local: using-diffusers/svd |
| title: Stable Video Diffusion |
| - local: using-diffusers/marigold_usage |
| title: Marigold Computer Vision |
| title: Specific pipeline examples |
| - sections: |
| - local: training/overview |
| title: Overview |
| - local: training/create_dataset |
| title: Create a dataset for training |
| - local: training/adapt_a_model |
| title: Adapt a model to a new task |
| - isExpanded: false |
| sections: |
| - local: training/unconditional_training |
| title: Unconditional image generation |
| - local: training/text2image |
| title: Text-to-image |
| - local: training/sdxl |
| title: Stable Diffusion XL |
| - local: training/kandinsky |
| title: Kandinsky 2.2 |
| - local: training/wuerstchen |
| title: Wuerstchen |
| - local: training/controlnet |
| title: ControlNet |
| - local: training/t2i_adapters |
| title: T2I-Adapters |
| - local: training/instructpix2pix |
| title: InstructPix2Pix |
| title: Models |
| - isExpanded: false |
| sections: |
| - local: training/text_inversion |
| title: Textual Inversion |
| - local: training/dreambooth |
| title: DreamBooth |
| - local: training/lora |
| title: LoRA |
| - local: training/custom_diffusion |
| title: Custom Diffusion |
| - local: training/lcm_distill |
| title: Latent Consistency Distillation |
| - local: training/ddpo |
| title: Reinforcement learning training with DDPO |
| title: Methods |
| title: Training |
| - sections: |
| - local: optimization/fp16 |
| title: Speed up inference |
| - local: optimization/memory |
| title: Reduce memory usage |
| - local: optimization/torch2.0 |
| title: PyTorch 2.0 |
| - local: optimization/xformers |
| title: xFormers |
| - local: optimization/tome |
| title: Token merging |
| - local: optimization/deepcache |
| title: DeepCache |
| - local: optimization/tgate |
| title: TGATE |
| - local: optimization/xdit |
| title: xDiT |
| - sections: |
| - local: using-diffusers/stable_diffusion_jax_how_to |
| title: JAX/Flax |
| - local: optimization/onnx |
| title: ONNX |
| - local: optimization/open_vino |
| title: OpenVINO |
| - local: optimization/coreml |
| title: Core ML |
| title: Optimized model formats |
| - sections: |
| - local: optimization/mps |
| title: Metal Performance Shaders (MPS) |
| - local: optimization/habana |
| title: Habana Gaudi |
| title: Optimized hardware |
| title: Accelerate inference and reduce memory |
| - sections: |
| - local: conceptual/philosophy |
| title: Philosophy |
| - local: using-diffusers/controlling_generation |
| title: Controlled generation |
| - local: conceptual/contribution |
| title: How to contribute? |
| - local: conceptual/ethical_guidelines |
| title: Diffusers' Ethical Guidelines |
| - local: conceptual/evaluation |
| title: Evaluating Diffusion Models |
| title: Conceptual Guides |
| - sections: |
| - local: community_projects |
| title: Projects built with Diffusers |
| title: Community Projects |
| - sections: |
| - isExpanded: false |
| sections: |
| - local: api/configuration |
| title: Configuration |
| - local: api/logging |
| title: Logging |
| - local: api/outputs |
| title: Outputs |
| title: Main Classes |
| - isExpanded: false |
| sections: |
| - local: api/loaders/ip_adapter |
| title: IP-Adapter |
| - local: api/loaders/lora |
| title: LoRA |
| - local: api/loaders/single_file |
| title: Single files |
| - local: api/loaders/textual_inversion |
| title: Textual Inversion |
| - local: api/loaders/unet |
| title: UNet |
| - local: api/loaders/peft |
| title: PEFT |
| title: Loaders |
| - isExpanded: false |
| sections: |
| - local: api/models/overview |
| title: Overview |
| - sections: |
| - local: api/models/controlnet |
| title: ControlNetModel |
| - local: api/models/controlnet_flux |
| title: FluxControlNetModel |
| - local: api/models/controlnet_hunyuandit |
| title: HunyuanDiT2DControlNetModel |
| - local: api/models/controlnet_sd3 |
| title: SD3ControlNetModel |
| - local: api/models/controlnet_sparsectrl |
| title: SparseControlNetModel |
| title: ControlNets |
| - sections: |
| - local: api/models/aura_flow_transformer2d |
| title: AuraFlowTransformer2DModel |
| - local: api/models/cogvideox_transformer3d |
| title: CogVideoXTransformer3DModel |
| - local: api/models/dit_transformer2d |
| title: DiTTransformer2DModel |
| - local: api/models/flux_transformer |
| title: FluxTransformer2DModel |
| - local: api/models/hunyuan_transformer2d |
| title: HunyuanDiT2DModel |
| - local: api/models/latte_transformer3d |
| title: LatteTransformer3DModel |
| - local: api/models/lumina_nextdit2d |
| title: LuminaNextDiT2DModel |
| - local: api/models/pixart_transformer2d |
| title: PixArtTransformer2DModel |
| - local: api/models/prior_transformer |
| title: PriorTransformer |
| - local: api/models/sd3_transformer2d |
| title: SD3Transformer2DModel |
| - local: api/models/stable_audio_transformer |
| title: StableAudioDiTModel |
| - local: api/models/transformer2d |
| title: Transformer2DModel |
| - local: api/models/transformer_temporal |
| title: TransformerTemporalModel |
| title: Transformers |
| - sections: |
| - local: api/models/stable_cascade_unet |
| title: StableCascadeUNet |
| - local: api/models/unet |
| title: UNet1DModel |
| - local: api/models/unet2d |
| title: UNet2DModel |
| - local: api/models/unet2d-cond |
| title: UNet2DConditionModel |
| - local: api/models/unet3d-cond |
| title: UNet3DConditionModel |
| - local: api/models/unet-motion |
| title: UNetMotionModel |
| - local: api/models/uvit2d |
| title: UViT2DModel |
| title: UNets |
| - sections: |
| - local: api/models/autoencoderkl |
| title: AutoencoderKL |
| - local: api/models/autoencoderkl_cogvideox |
| title: AutoencoderKLCogVideoX |
| - local: api/models/asymmetricautoencoderkl |
| title: AsymmetricAutoencoderKL |
| - local: api/models/consistency_decoder_vae |
| title: ConsistencyDecoderVAE |
| - local: api/models/autoencoder_oobleck |
| title: Oobleck AutoEncoder |
| - local: api/models/autoencoder_tiny |
| title: Tiny AutoEncoder |
| - local: api/models/vq |
| title: VQModel |
| title: VAEs |
| title: Models |
| - isExpanded: false |
| sections: |
| - local: api/pipelines/overview |
| title: Overview |
| - local: api/pipelines/amused |
| title: aMUSEd |
| - local: api/pipelines/animatediff |
| title: AnimateDiff |
| - local: api/pipelines/attend_and_excite |
| title: Attend-and-Excite |
| - local: api/pipelines/audioldm |
| title: AudioLDM |
| - local: api/pipelines/audioldm2 |
| title: AudioLDM 2 |
| - local: api/pipelines/aura_flow |
| title: AuraFlow |
| - local: api/pipelines/auto_pipeline |
| title: AutoPipeline |
| - local: api/pipelines/blip_diffusion |
| title: BLIP-Diffusion |
| - local: api/pipelines/cogvideox |
| title: CogVideoX |
| - local: api/pipelines/consistency_models |
| title: Consistency Models |
| - local: api/pipelines/controlnet |
| title: ControlNet |
| - local: api/pipelines/controlnet_flux |
| title: ControlNet with Flux.1 |
| - local: api/pipelines/controlnet_hunyuandit |
| title: ControlNet with Hunyuan-DiT |
| - local: api/pipelines/controlnet_sd3 |
| title: ControlNet with Stable Diffusion 3 |
| - local: api/pipelines/controlnet_sdxl |
| title: ControlNet with Stable Diffusion XL |
| - local: api/pipelines/controlnetxs |
| title: ControlNet-XS |
| - local: api/pipelines/controlnetxs_sdxl |
| title: ControlNet-XS with Stable Diffusion XL |
| - local: api/pipelines/dance_diffusion |
| title: Dance Diffusion |
| - local: api/pipelines/ddim |
| title: DDIM |
| - local: api/pipelines/ddpm |
| title: DDPM |
| - local: api/pipelines/deepfloyd_if |
| title: DeepFloyd IF |
| - local: api/pipelines/diffedit |
| title: DiffEdit |
| - local: api/pipelines/dit |
| title: DiT |
| - local: api/pipelines/flux |
| title: Flux |
| - local: api/pipelines/hunyuandit |
| title: Hunyuan-DiT |
| - local: api/pipelines/i2vgenxl |
| title: I2VGen-XL |
| - local: api/pipelines/pix2pix |
| title: InstructPix2Pix |
| - local: api/pipelines/kandinsky |
| title: Kandinsky 2.1 |
| - local: api/pipelines/kandinsky_v22 |
| title: Kandinsky 2.2 |
| - local: api/pipelines/kandinsky3 |
| title: Kandinsky 3 |
| - local: api/pipelines/kolors |
| title: Kolors |
| - local: api/pipelines/latent_consistency_models |
| title: Latent Consistency Models |
| - local: api/pipelines/latent_diffusion |
| title: Latent Diffusion |
| - local: api/pipelines/latte |
| title: Latte |
| - local: api/pipelines/ledits_pp |
| title: LEDITS++ |
| - local: api/pipelines/lumina |
| title: Lumina-T2X |
| - local: api/pipelines/marigold |
| title: Marigold |
| - local: api/pipelines/panorama |
| title: MultiDiffusion |
| - local: api/pipelines/musicldm |
| title: MusicLDM |
| - local: api/pipelines/pag |
| title: PAG |
| - local: api/pipelines/paint_by_example |
| title: Paint by Example |
| - local: api/pipelines/pia |
| title: Personalized Image Animator (PIA) |
| - local: api/pipelines/pixart |
| title: PixArt-α |
| - local: api/pipelines/pixart_sigma |
| title: PixArt-Σ |
| - local: api/pipelines/self_attention_guidance |
| title: Self-Attention Guidance |
| - local: api/pipelines/semantic_stable_diffusion |
| title: Semantic Guidance |
| - local: api/pipelines/shap_e |
| title: Shap-E |
| - local: api/pipelines/stable_audio |
| title: Stable Audio |
| - local: api/pipelines/stable_cascade |
| title: Stable Cascade |
| - sections: |
| - local: api/pipelines/stable_diffusion/overview |
| title: Overview |
| - local: api/pipelines/stable_diffusion/text2img |
| title: Text-to-image |
| - local: api/pipelines/stable_diffusion/img2img |
| title: Image-to-image |
| - local: api/pipelines/stable_diffusion/svd |
| title: Image-to-video |
| - local: api/pipelines/stable_diffusion/inpaint |
| title: Inpainting |
| - local: api/pipelines/stable_diffusion/depth2img |
| title: Depth-to-image |
| - local: api/pipelines/stable_diffusion/image_variation |
| title: Image variation |
| - local: api/pipelines/stable_diffusion/stable_diffusion_safe |
| title: Safe Stable Diffusion |
| - local: api/pipelines/stable_diffusion/stable_diffusion_2 |
| title: Stable Diffusion 2 |
| - local: api/pipelines/stable_diffusion/stable_diffusion_3 |
| title: Stable Diffusion 3 |
| - local: api/pipelines/stable_diffusion/stable_diffusion_xl |
| title: Stable Diffusion XL |
| - local: api/pipelines/stable_diffusion/sdxl_turbo |
| title: SDXL Turbo |
| - local: api/pipelines/stable_diffusion/latent_upscale |
| title: Latent upscaler |
| - local: api/pipelines/stable_diffusion/upscale |
| title: Super-resolution |
| - local: api/pipelines/stable_diffusion/k_diffusion |
| title: K-Diffusion |
| - local: api/pipelines/stable_diffusion/ldm3d_diffusion |
| title: LDM3D Text-to-(RGB, Depth), Text-to-(RGB-pano, Depth-pano), LDM3D Upscaler |
| - local: api/pipelines/stable_diffusion/adapter |
| title: T2I-Adapter |
| - local: api/pipelines/stable_diffusion/gligen |
| title: GLIGEN (Grounded Language-to-Image Generation) |
| title: Stable Diffusion |
| - local: api/pipelines/stable_unclip |
| title: Stable unCLIP |
| - local: api/pipelines/text_to_video |
| title: Text-to-video |
| - local: api/pipelines/text_to_video_zero |
| title: Text2Video-Zero |
| - local: api/pipelines/unclip |
| title: unCLIP |
| - local: api/pipelines/unidiffuser |
| title: UniDiffuser |
| - local: api/pipelines/value_guided_sampling |
| title: Value-guided sampling |
| - local: api/pipelines/wuerstchen |
| title: Wuerstchen |
| title: Pipelines |
| - isExpanded: false |
| sections: |
| - local: api/schedulers/overview |
| title: Overview |
| - local: api/schedulers/cm_stochastic_iterative |
| title: CMStochasticIterativeScheduler |
| - local: api/schedulers/consistency_decoder |
| title: ConsistencyDecoderScheduler |
| - local: api/schedulers/cosine_dpm |
| title: CosineDPMSolverMultistepScheduler |
| - local: api/schedulers/ddim_inverse |
| title: DDIMInverseScheduler |
| - local: api/schedulers/ddim |
| title: DDIMScheduler |
| - local: api/schedulers/ddpm |
| title: DDPMScheduler |
| - local: api/schedulers/deis |
| title: DEISMultistepScheduler |
| - local: api/schedulers/multistep_dpm_solver_inverse |
| title: DPMSolverMultistepInverse |
| - local: api/schedulers/multistep_dpm_solver |
| title: DPMSolverMultistepScheduler |
| - local: api/schedulers/dpm_sde |
| title: DPMSolverSDEScheduler |
| - local: api/schedulers/singlestep_dpm_solver |
| title: DPMSolverSinglestepScheduler |
| - local: api/schedulers/edm_multistep_dpm_solver |
| title: EDMDPMSolverMultistepScheduler |
| - local: api/schedulers/edm_euler |
| title: EDMEulerScheduler |
| - local: api/schedulers/euler_ancestral |
| title: EulerAncestralDiscreteScheduler |
| - local: api/schedulers/euler |
| title: EulerDiscreteScheduler |
| - local: api/schedulers/flow_match_euler_discrete |
| title: FlowMatchEulerDiscreteScheduler |
| - local: api/schedulers/flow_match_heun_discrete |
| title: FlowMatchHeunDiscreteScheduler |
| - local: api/schedulers/heun |
| title: HeunDiscreteScheduler |
| - local: api/schedulers/ipndm |
| title: IPNDMScheduler |
| - local: api/schedulers/stochastic_karras_ve |
| title: KarrasVeScheduler |
| - local: api/schedulers/dpm_discrete_ancestral |
| title: KDPM2AncestralDiscreteScheduler |
| - local: api/schedulers/dpm_discrete |
| title: KDPM2DiscreteScheduler |
| - local: api/schedulers/lcm |
| title: LCMScheduler |
| - local: api/schedulers/lms_discrete |
| title: LMSDiscreteScheduler |
| - local: api/schedulers/pndm |
| title: PNDMScheduler |
| - local: api/schedulers/repaint |
| title: RePaintScheduler |
| - local: api/schedulers/score_sde_ve |
| title: ScoreSdeVeScheduler |
| - local: api/schedulers/score_sde_vp |
| title: ScoreSdeVpScheduler |
| - local: api/schedulers/tcd |
| title: TCDScheduler |
| - local: api/schedulers/unipc |
| title: UniPCMultistepScheduler |
| - local: api/schedulers/vq_diffusion |
| title: VQDiffusionScheduler |
| title: Schedulers |
| - isExpanded: false |
| sections: |
| - local: api/internal_classes_overview |
| title: Overview |
| - local: api/attnprocessor |
| title: Attention Processor |
| - local: api/activations |
| title: Custom activation functions |
| - local: api/normalization |
| title: Custom normalization layers |
| - local: api/utilities |
| title: Utilities |
| - local: api/image_processor |
| title: VAE Image Processor |
| - local: api/video_processor |
| title: Video Processor |
| title: Internal classes |
| title: API |
|
|