Alex Hant
hardhant
AI & ML interests
None yet
Recent Activity
reacted to DedeProGames's post with ๐ฅ about 1 hour ago
๐ Introducing the GRM-2.6 Family
The GRM-2.6 family is a new generation of reasoning-focused models from Orion LLM Labs, built for difficult tasks, coding, STEM, terminal agents, and advanced local AI workflows.
GRM-2.6-Plus is the main high-capability model in the family: a 27B-class reasoning model based on Qwen3.6, designed for strong structured reasoning, coding, agentic use, and practical local deployment.
GRM-2.6-Opus builds on GRM-2.6-Plus as a merge with an Opus-style reasoning distilled model, improving structured reasoning behavior, terminal-agent workflows, coding ability, and complex problem solving.
Both models are designed for users who want powerful reasoning models that remain practical for research, local inference, coding, and agent experiments.
Models:
GRM-2.6-Plus: https://huggingface.co/OrionLLM/GRM-2.6-Plus
GRM-2.6-Opus: https://huggingface.co/OrionLLM/GRM-2.6-Opus
Organization:
https://huggingface.co/OrionLLM reacted to danielhanchen's post with ๐ฅ 3 days ago
We made a guide on how to run open LLMs in Claude Code, Codex and OpenClaw.
Use Gemma 4 and Qwen3.6 GGUFs for local agentic coding on 24GB RAM
Run with self-healing tool calls, code execution, web search via the Unsloth API endpoint and llama.cpp
Guide: https://unsloth.ai/docs/basics/api reacted to DedeProGames's post with ๐ฅ 3 days ago
GRaPE 2 Pro is now available.
https://huggingface.co/SL-AI/GRaPE-2-Pro
This is the flagship model of the GRaPE 2 family and the largest model I have trained to date, sitting at 27B parameters. It is built on Qwen3.5-27B and trained on a closed-source proprietary dataset, with roughly half of post-training focused on code and the rest split between STEAM subjects and structured logical reasoning. It punches seriously above its weight class.
GRaPE 2 Pro supports multimodal input (image + text) and features 6 thinking modes via the `<thinking_mode>` tag. This gives you real control over how hard the model thinks, from skipping the reasoning phase entirely with `minimal`, all the way up to `xtra-Hi` for deep, extended thought on hard problems. For most agentic use, `auto` or `low` is the move to keep things snappy.
It also runs on consumer hardware. You can get it going with as low as 12GB of VRAM on a quantized build.
If you want to try it out and give feedback, that would be really appreciated. Email us at `contact@skinnertopia.com`Organizations
None yet