Switch to full model with fp16/bf16 inference for better performance 39fa408 Txu647 commited on Jan 28
UI improvements: move status bar to right side, simplify layout, update defaults to Wang Xizhi 89e2699 TSXu commited on Jan 27
perf: enable FlashAttention/MemEfficient SDPA backends instead of torch.compile 9de4f7d Txu647 commited on Jan 27
fix: remove flash-attn (requires source build), add batch generation support 5a3fec3 Txu647 commited on Jan 27
fix: use assign=True in load_state_dict to preserve checkpoint dtype c6a1e05 Txu647 commited on Jan 27
fix: import spaces first and lazy load inference to avoid CUDA init 9d5c8cc Txu647 commited on Jan 27
fix: pin huggingface-hub and transformers versions for compatibility 583c05d Txu647 commited on Jan 27
Duplicate from gradio-templates/text-to-image-gradio-template c2ad4cd verified TSXu fffiloni commited on Jan 27