Qwenjamin Franklin
Collection
Qwen 3.5 9b better at false smoke and workbench • 2 items • Updated
How to use stamsam/Qwenjamin_Franklin with MLX:
# Make sure mlx-lm is installed
# pip install --upgrade mlx-lm
# Generate text with mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("stamsam/Qwenjamin_Franklin")
prompt = "Write a story about Einstein"
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
text = generate(model, tokenizer, prompt=prompt, verbose=True)How to use stamsam/Qwenjamin_Franklin with Pi:
# Install MLX LM: uv tool install mlx-lm # Start a local OpenAI-compatible server: mlx_lm.server --model "stamsam/Qwenjamin_Franklin"
# Install Pi:
npm install -g @mariozechner/pi-coding-agent
# Add to ~/.pi/agent/models.json:
{
"providers": {
"mlx-lm": {
"baseUrl": "http://localhost:8080/v1",
"api": "openai-completions",
"apiKey": "none",
"models": [
{
"id": "stamsam/Qwenjamin_Franklin"
}
]
}
}
}# Start Pi in your project directory: pi
How to use stamsam/Qwenjamin_Franklin with Hermes Agent:
# Install MLX LM: uv tool install mlx-lm # Start a local OpenAI-compatible server: mlx_lm.server --model "stamsam/Qwenjamin_Franklin"
# Install Hermes: curl -fsSL https://hermes-agent.nousresearch.com/install.sh | bash hermes setup # Point Hermes at the local server: hermes config set model.provider custom hermes config set model.base_url http://127.0.0.1:8080/v1 hermes config set model.default stamsam/Qwenjamin_Franklin
hermes
How to use stamsam/Qwenjamin_Franklin with MLX LM:
# Install MLX LM uv tool install mlx-lm # Interactive chat REPL mlx_lm.chat --model "stamsam/Qwenjamin_Franklin"
# Install MLX LM
uv tool install mlx-lm
# Start the server
mlx_lm.server --model "stamsam/Qwenjamin_Franklin"
# Calling the OpenAI-compatible server with curl
curl -X POST "http://localhost:8000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "stamsam/Qwenjamin_Franklin",
"messages": [
{"role": "user", "content": "Hello"}
]
}'Qwenjamin Franklin is the dequantized raw MLX export of the same tuned v14 branch used for the 4-bit release.
This repo is for people who want the larger raw MLX artifact from the same local workshop line instead of the compact 4-bit build.
Qwen/Qwen3.5-9Bv14 broad-benchmark daily-driverInternal workshop evals. These scores are project-specific and directional, not public leaderboard claims.
| Eval | Base Qwen3.5-9B-MLX-4bit |
Qwenjamin Franklin |
|---|---|---|
workbench_local_agent_smoke |
63/100 | 72/100 |
full40 |
309/400 | 325/400 |
json_hard |
15/30 | 30/30 |
parser_gate |
2/3, 1/3, 1/3 | 3/3, 3/3, 3/3 |
code_smoke |
95/120 | 95/120 |
false_smoke |
102/110 | 110/110 |
tool_schema_canary |
50/175 | 106/175 |
no_tool_leakage |
99/100 | 100/100 |
python -m mlx_lm generate \
--model stamsam/Qwenjamin_Franklin \
--prompt "Write one complete TypeScript module and return only code." \
--max-tokens 512 \
--temp 0.0
Quantized