It's too Large

#12
by qpqpqpqpqpqp - opened

Please consider distilling, pruning or converting to ControlLora, it should save memory without great quality degradation

Alibaba-PAI org

Do you mean the model weights are too large?

Do you mean the model weights are too large?

Yes, 3 GB fp16

Yes, 3 GB fp16

Try to optimize your flow with RAM memory, ComfyUI-MultiGPU custom node
also sage attention + torch compile = must have

Try to optimize your flow with RAM memory, ComfyUI-MultiGPU custom node
also sage attention + torch compile = must have

Valid, but the file size is 3 GB anyway, it's possible to make a decent ControlLora or another adapter to save memory

Sign up or log in to comment