DedeProGames commited on
Commit
7e47eb9
·
verified ·
1 Parent(s): 88550ab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -1
README.md CHANGED
@@ -37,7 +37,6 @@ The model is optimized for **structured reasoning**, helping it produce more acc
37
  - **Strong Reasoning for Everyday and Advanced Tasks:** GRM-2.5 is built to handle both daily conversations and more demanding reasoning workloads with clarity and consistency.
38
  - **Efficient Local Coding and Agentic Use:** Despite its compact size, the model is well suited for code generation, structured problem-solving, and local agent-style workflows.
39
  - **Optimized for Local Deployment:** GRM-2.5 is designed for accessible inference across a broad range of hardware, making it a practical choice for users who want capable AI running locally.
40
- - **Long Context and Multimodal Support:** This private mirror inherits long-context and multimodal capabilities from the upstream `Qwen/Qwen3.5-4B` release.
41
 
42
  ## 3. Performance
43
  GRM-2.5 is designed to be a highly capable option for **local AI use** across many scenarios. It performs well in **complex reasoning tasks, everyday chat, coding, and agentic workflows**, while maintaining the efficiency expected from a compact 4B model.
 
37
  - **Strong Reasoning for Everyday and Advanced Tasks:** GRM-2.5 is built to handle both daily conversations and more demanding reasoning workloads with clarity and consistency.
38
  - **Efficient Local Coding and Agentic Use:** Despite its compact size, the model is well suited for code generation, structured problem-solving, and local agent-style workflows.
39
  - **Optimized for Local Deployment:** GRM-2.5 is designed for accessible inference across a broad range of hardware, making it a practical choice for users who want capable AI running locally.
 
40
 
41
  ## 3. Performance
42
  GRM-2.5 is designed to be a highly capable option for **local AI use** across many scenarios. It performs well in **complex reasoning tasks, everyday chat, coding, and agentic workflows**, while maintaining the efficiency expected from a compact 4B model.