Add GGUF quantized models (BF16, Q8_0, Q4_K_M) and update README with GGUF usage section 60a6bce root commited on 3 days ago