data: add raw benchmark results JSON (KMMLU + HAE-RAE, 3-model comparison) de3e104 verified intrect commited on 2 days ago
docs: add Korean LLM benchmark results (KMMLU + HAE-RAE, 3-model comparison) 4043357 verified intrect commited on 2 days ago
fix: replace Qwen default system prompt with VELA identity in chat templates 8fa0fb0 verified intrect commited on 3 days ago
docs: add llama-cpp-python, vllm, ollama tags for library discovery 6eb484c verified intrect commited on 3 days ago
docs: add MLX 4-bit format, update quant links in model card 8bdb2c7 verified intrect commited on 3 days ago
feat: add MLX 4-bit quantized model (Apple Silicon optimized) d7f1985 verified intrect commited on 3 days ago
feat: update to DPO v6 merged model (BF16 safetensors) ffbdde5 verified intrect commited on 3 days ago
fix: generation_config rep_penalty=1.05, top_k=20, top_p=0.8 (vela-v5-merged) a0a3973 verified intrect commited on Feb 18
docs: add recommended inference settings and backend configuration guide 3e8a713 verified intrect commited on Feb 17
fix: match llama-cpp-python defaults (top_k=40, top_p=0.95, rep_penalty=1.0) dea4e49 verified intrect commited on Feb 17
fix: rollback generation_config to safe Qwen2.5 defaults (rep_penalty 1.3→1.1) bfc3043 verified intrect commited on Feb 16
fix: update generation params — repetition_penalty 1.05→1.3, top_k 20→50, top_p 0.8→0.92 d4a1479 verified intrect commited on Feb 16
docs: add real output example (RT + Quick Assessment + Analysis Report) c66f3a2 verified intrect commited on Feb 16
docs: update model card with v3 training data (58K SFT, 26K DPO), MLX benchmark, Markdown RT format 7f641b5 verified intrect commited on Feb 15
feat: update to DPO v4 merged model (SFT + DPO v4 language leak fix) c34c4f4 verified intrect commited on Feb 15
fix: change tokenizer_class to Qwen2TokenizerFast for vLLM compatibility ecf8e6b verified intrect commited on Feb 13
docs: update training data distribution with accurate numbers (SFT 36,713 + DPO 24,779) d35ad96 verified intrect commited on Feb 12
docs: update model card with GGUF formats, benchmarks, usage examples 2448e8a verified intrect commited on Feb 12
Fix tokenizer_config.json - remove extra_special_tokens list causing vLLM error f650ef7 verified intrect commited on Jan 28
Fix config.json for vLLM compatibility (remove layer_types, fix rope_parameters) 7dc1232 verified intrect commited on Jan 28