|
|
""" |
|
|
LCM + LTX scheduler combining Latent Consistency Model with RectifiedFlow (LTX). |
|
|
Optimized for Lightning LoRA compatibility and ultra-fast inference. |
|
|
""" |
|
|
|
|
|
import torch |
|
|
import math |
|
|
from diffusers.schedulers.scheduling_utils import SchedulerMixin, SchedulerOutput |
|
|
|
|
|
|
|
|
class LCMScheduler(SchedulerMixin): |
|
|
""" |
|
|
LCM + LTX scheduler combining Latent Consistency Model with RectifiedFlow. |
|
|
- LCM: Enables 2-8 step inference with consistency models |
|
|
- LTX: Uses RectifiedFlow for better flow matching dynamics |
|
|
Optimized for Lightning LoRAs and ultra-fast, high-quality generation. |
|
|
""" |
|
|
|
|
|
def __init__(self, num_train_timesteps: int = 1000, num_inference_steps: int = 4, shift: float = 1.0): |
|
|
self.num_train_timesteps = num_train_timesteps |
|
|
self.num_inference_steps = num_inference_steps |
|
|
self.shift = shift |
|
|
self._step_index = None |
|
|
|
|
|
def set_timesteps(self, num_inference_steps: int, device=None, shift: float = None, **kwargs): |
|
|
"""Set timesteps for LCM+LTX inference using RectifiedFlow approach""" |
|
|
self.num_inference_steps = min(num_inference_steps, 8) |
|
|
|
|
|
if shift is None: |
|
|
shift = self.shift |
|
|
|
|
|
|
|
|
|
|
|
t = torch.linspace(0, 1, self.num_inference_steps + 1, dtype=torch.float32) |
|
|
|
|
|
|
|
|
|
|
|
sigma_max = 1.0 |
|
|
sigma_min = 0.003 / 1.002 |
|
|
|
|
|
|
|
|
|
|
|
sigmas = sigma_min + (sigma_max - sigma_min) * (1 - t) |
|
|
|
|
|
|
|
|
sigmas = shift * sigmas / (1 + (shift - 1) * sigmas) |
|
|
|
|
|
self.sigmas = sigmas |
|
|
self.timesteps = self.sigmas[:-1] * self.num_train_timesteps |
|
|
|
|
|
if device is not None: |
|
|
self.timesteps = self.timesteps.to(device) |
|
|
self.sigmas = self.sigmas.to(device) |
|
|
self._step_index = None |
|
|
|
|
|
def step(self, model_output: torch.Tensor, timestep: torch.Tensor, sample: torch.Tensor, **kwargs) -> SchedulerOutput: |
|
|
""" |
|
|
Perform LCM + LTX step combining consistency model with rectified flow. |
|
|
- LCM: Direct consistency model prediction for fast inference |
|
|
- LTX: RectifiedFlow dynamics for optimal probability flow path |
|
|
""" |
|
|
if self._step_index is None: |
|
|
self._init_step_index(timestep) |
|
|
|
|
|
|
|
|
sigma = self.sigmas[self._step_index] |
|
|
if self._step_index + 1 < len(self.sigmas): |
|
|
sigma_next = self.sigmas[self._step_index + 1] |
|
|
else: |
|
|
sigma_next = torch.zeros_like(sigma) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
sigma_diff = (sigma_next - sigma) |
|
|
while len(sigma_diff.shape) < len(sample.shape): |
|
|
sigma_diff = sigma_diff.unsqueeze(-1) |
|
|
|
|
|
|
|
|
|
|
|
prev_sample = sample + model_output * sigma_diff |
|
|
self._step_index += 1 |
|
|
|
|
|
return SchedulerOutput(prev_sample=prev_sample) |
|
|
|
|
|
def _init_step_index(self, timestep): |
|
|
"""Initialize step index based on current timestep""" |
|
|
if isinstance(timestep, torch.Tensor): |
|
|
timestep = timestep.to(self.timesteps.device) |
|
|
indices = (self.timesteps == timestep).nonzero() |
|
|
if len(indices) > 0: |
|
|
self._step_index = indices[0].item() |
|
|
else: |
|
|
|
|
|
diffs = torch.abs(self.timesteps - timestep) |
|
|
self._step_index = torch.argmin(diffs).item() |
|
|
|