Where can I find the inference.py mentioned in the CLI Inference section?
#29 opened 2 days ago
by
Zebin01
How to run multi-GPU inference on 2x NVIDIA T4?
#28 opened 4 days ago
by
Marshalldom
There will be model AI video Generator on 14B, 20B, 27B, 32B MoE to active 7B, 8B, 9B? And.... support model AI video Generator type Reasoning/thinking?
π 7
3
#26 opened 12 days ago
by
aomsan461
Can MoE be used to get this to 4B and still fit into 5090 VRAM?
1
#11 opened 18 days ago
by
usernameSRSalreadyexists
Is the QK^T result the VRAM bottleneck for video models?
1
#10 opened 19 days ago
by
yunming181920
GGUF versions
ππ 5
2
#4 opened 22 days ago
by
maroo87
Struggling to get this to run on a 24gb gpu
17
#3 opened 24 days ago
by
CodeExplode
Distilled from VJEPA feature space?
1
#1 opened 25 days ago
by
eggsbenedicto