license: apache-2.0
base_model:
- Qwen/Qwen3.5-4B
pipeline_tag: image-text-to-text
tags:
- reasoning
- vision
- multimodal
- instruct
- chat
- coding
- math
- science
GRM-2.5
1. Introduction
GRM-2.5 is a 4B-parameter reasoning model built for general-purpose local AI. It is designed to deliver strong performance across a wide range of tasks while remaining efficient and accessible for local inference.
The model is optimized for structured reasoning, helping it produce more accurate, coherent, and reliable responses on complex problems. GRM-2.5 aims to combine strong reasoning ability, practical usability, and efficient deployment in a compact form factor.
2. Key Capabilities
- Strong Reasoning for Everyday and Advanced Tasks: GRM-2.5 is built to handle both daily conversations and more demanding reasoning workloads with clarity and consistency.
- Efficient Local Coding and Agentic Use: Despite its compact size, the model is well suited for code generation, structured problem-solving, and local agent-style workflows.
- Optimized for Local Deployment: GRM-2.5 is designed for accessible inference across a broad range of hardware, making it a practical choice for users who want capable AI running locally.
3. Performance
GRM-2.5 is designed to be a highly capable option for local AI use across many scenarios. It performs well in complex reasoning tasks, everyday chat, coding, and agentic workflows, while maintaining the efficiency expected from a compact 4B model.
Its focus is not only raw capability, but also practical intelligence: strong reasoning, stable long-context behavior, and usability on consumer hardware.