---
license: apache-2.0
base_model:
- Qwen/Qwen3.5-4B
pipeline_tag: image-text-to-text
tags:
- reasoning
- vision
- multimodal
- instruct
- chat
- coding
- math
- science
---
# GRM-2.5
## 1. Introduction
GRM-2.5 is a **4B-parameter reasoning model** built for **general-purpose local AI**. It is designed to deliver strong performance across a wide range of tasks while remaining efficient and accessible for local inference.
The model is optimized for **structured reasoning**, helping it produce more accurate, coherent, and reliable responses on complex problems. GRM-2.5 aims to combine strong reasoning ability, practical usability, and efficient deployment in a compact form factor.
## 2. Key Capabilities
- **Strong Reasoning for Everyday and Advanced Tasks:** GRM-2.5 is built to handle both daily conversations and more demanding reasoning workloads with clarity and consistency.
- **Efficient Local Coding and Agentic Use:** Despite its compact size, the model is well suited for code generation, structured problem-solving, and local agent-style workflows.
- **Optimized for Local Deployment:** GRM-2.5 is designed for accessible inference across a broad range of hardware, making it a practical choice for users who want capable AI running locally.
## 3. Performance
GRM-2.5 is designed to be a highly capable option for **local AI use** across many scenarios. It performs well in **complex reasoning tasks, everyday chat, coding, and agentic workflows**, while maintaining the efficiency expected from a compact 4B model.
Its focus is not only raw capability, but also **practical intelligence**: strong reasoning, stable long-context behavior, and usability on consumer hardware.