{ "model_type": "ValueModel", "base_model": "Qwen/Qwen2.5-Math-1.5B-Instruct", "value_head": { "file": "value_head.safetensors", "shape": [ 1, 1536 ], "dtype": "torch.bfloat16" }, "inference_note": "Load base model via transformers; load value_head.safetensors separately and apply on last_hidden_state." }