12G vram failed to run

#2
by ZKong - opened

takes too long, can we set max vram use to not overflow to the shared vram
2026-01-09_114924

always compile even torch.compile(pipe.transformer) is disabled.
enable_model_cpu_offload still can over 12G, i have to use enable_sequential_cpu_offload but slower, maxvram is 3G, very small vram use!

ZKong changed discussion status to closed

Sign up or log in to comment