qwen25-3b-tealkit

โš ๏ธ This model is purpose-built for the TealKit agentic AI app. It is optimised for MCP tool-call generation inside TealKit's server mode. It will not perform well as a general-purpose assistant.

GGUF model fine-tuned for structured MCP tool-calling, ready for local inference via Ollama.

Model Details

Property Value
Base model (training) mlx-community/Qwen2.5-3B-Instruct-4bit
Base model (fused export) Qwen/Qwen2.5-3B-Instruct
Fine-tune method QLoRA / LoRA adapter fusion
Quantization Q5_K_M (5-bit, recommended)
GGUF file qwen25-3b-tealkit-q5_k_m.gguf
Preset qwen2_5_3b

Intended Use

Only intended for use within the TealKit AI mobile app.

TealKit uses this model in server mode to generate structured JSON MCP tool calls. The model was trained on a custom MCP tool-call JSONL dataset and is not suited for general chat.

Quick Start (Ollama)

ollama pull qwen25-3b-tealkit   # if published to Ollama registry
# or after local registration:
ollama run qwen25-3b-tealkit

Project

Files

  • qwen25-3b-tealkit-q5_k_m.gguf
  • Modelfile โ€” Ollama model definition with system prompt

Notes

Produced via LoRA fine-tuning on Mac Apple Silicon (MLX), adapter fusion, and llama.cpp GGUF conversion. See the TealKit training guide for full pipeline details.

Downloads last month
42
GGUF
Model size
3B params
Architecture
qwen2
Hardware compatibility
Log In to add your hardware

5-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for lschaffer/qwen25-3b-tealkit

Base model

Qwen/Qwen2.5-3B
Adapter
(1262)
this model