Join the conversation

Join the community of Machine Learners and AI enthusiasts.

Sign Up
DavidAU 
posted an update Dec 19, 2025
Post
7146
SAVANT COMMANDER: 48B-A4B , 256k Context, GATED MOE.

I am going to showcase some other people's tuning work, that I have put into a GATED Distill MOE (Qwen3) ; 256 K context. Special thanks to all the tuners (listed in the model tree and repo page with special shoutout to "TeichAI" - using Unsloth for a lot of the Distills in this model):

Savant Commander is a specialized MOE model that allows you to control which expert(s) (of 12) are assigned to your use case(s) / prompt(s) ... directly (by name(s)), as opposed to having the "choices" made for you.

The model is composed of 12 DISTILLS (compressed 12x4B MOE) of top closed (GPT5.1, OpenAI 120 GPT Oss, Gemini (3), Claude (2) ) and open source models (Kimi, GLM, Deepseek, Command-A, JanV1 ) all in one.

256k Context, 2 experts activated.

PS: There is also a "heretic" / "decensored" version too ; listed on this model page.

DavidAU/Qwen3-48B-A4B-Savant-Commander-GATED-12x-Closed-Open-Source-Distill-GGUF

Hello, I'm a person who converts models into MLX format so people with apple silicon can run them efficiently. I can only convert the full model to MLX, I cannot convert GGUF models. So if you only upload the GGUF version of a model, then I cannot convert them to MLX and you will not have any apple silicon users of that model

·

Hey;

I am currently restricting access to the source presently due to past issues with abuse of the source (of my models), which lead to community issues due to non-disclosure of tech details of the model as well as issues related to non-attribution of multiple parties.

I may release it in a few weeks.