LocoOperator-4B โ€” MLX 6-bit Quantized

This is a 6-bit quantized MLX version of LocoreMind/LocoOperator-4B, converted for efficient inference on Apple Silicon using MLX.

Model Overview

Attribute Value
Original Model LocoreMind/LocoOperator-4B
Architecture Qwen3 (4B parameters)
Quantization 6-bit (MLX)
Base Model Qwen3-4B-Instruct-2507
Teacher Model Qwen3-Coder-Next
Training Method Full-parameter SFT (distillation from 170K samples)
Max Sequence Length 16,384 tokens
License MIT

About LocoOperator-4B

LocoOperator-4B is a 4B-parameter tool-calling agent model trained via knowledge distillation from Qwen3-Coder-Next inference traces. It specializes in multi-turn codebase exploration โ€” reading files, searching code, and navigating project structures within a Claude Code-style agent loop.

Key Features

  • Tool-Calling Agent: Generates structured <tool_call> JSON for Read, Grep, Glob, Bash, Write, Edit, and Task (subagent delegation)
  • 100% JSON Validity: Every tool call is valid JSON with all required arguments โ€” outperforming the teacher model (87.6%)
  • Multi-Turn: Handles conversation depths of 3โ€“33 messages with consistent tool-calling behavior

Performance

Metric Score
Tool Call Presence Alignment 100% (65/65)
First Tool Type Match 65.6% (40/61)
JSON Validity 100% (76/76)
Argument Syntax Correctness 100% (76/76)

Usage with MLX

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("DJLougen/LocoOperator-4B-MLX-6bit")

messages = [
    {
        "role": "system",
        "content": "You are a read-only codebase search specialist."
    },
    {
        "role": "user",
        "content": "Analyze the project structure at /workspace/myproject and explain the architecture."
    }
]

prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
response = generate(model, tokenizer, prompt=prompt, max_tokens=512)
print(response)

Other Quantizations

Acknowledgments

Downloads last month
27
Safetensors
Model size
0.9B params
Tensor type
BF16
ยท
U32
ยท
MLX
Hardware compatibility
Log In to add your hardware

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for DJLougen/LocoOperator-4B-MLX-6bit

Quantized
(18)
this model