VRAM required to run inference
#7
by
niktheod - opened
Hi. I have 16GBs of VRAM available but when I try to run the inference code you provide on hugging face I run out of memory. Any idea why this could be? Usually for models of similar size 16GBs of VRAM are more than enough.
niktheod changed discussion status to
closed