Fix: wrap as Qwen3_5ForConditionalGeneration + model.language_model.* prefix (drop-in for vLLM --language-model-only) 88b0c4a verified HeyMiku commited on 4 days ago
Replace MLX 4-bit with BF16 merged model (SFT best@6k, val loss 0.015). Compatible with vLLM/transformers. 7588fa9 verified HeyMiku commited on 4 days ago
Intent Classifier 4B — SFT best@6k (val loss 0.015, ~99% accuracy). Qwen3.5-4B base, 13 intents, 5 languages. 9c90cb4 verified HeyMiku commited on 4 days ago