Load LoRA via PeftModel on top of standard base models (fixes r=16 vs r=8 mismatch) eb30a86 verified DanielRegaladoCardoso commited on 18 days ago
ZeroGPU best practice: load models at module level (cuda), inference only inside @spaces.GPU a57eca6 verified DanielRegaladoCardoso commited on 18 days ago
Initial deploy: Apple/Claude design, DuckDB, 3 trained LoRAs a067ada verified DanielRegaladoCardoso commited on 18 days ago