CPP-qwen3-4B-Instruct-2507
Features
- This model is based on qwen3-4B-Instruct-2507
- Fine Tuned on MCES10-Software/CPP-Code-Solutions Dataset
- 4 Billion Parameters
- Finetuned with MLX
- model.safetensors
Benchmark
MAX TOKENS = 500 Apple Silicon Macbook Pro 18,3 M1 PRO 16 GB RAM
Write a function that checks if a number is prime.
bool isPrime(int n) {
if (n <= 1) return false;
for (int i = 2; i * i <= n; ++i) {
if (n % i == 0) return false;
}
return true;
}
49.27 tok/sec 63 tokens 0.24s to first token Credits
MCES10 Software
Thanks to:
QWEN for the model
- Downloads last month
- 30
Model size
4B params
Tensor type
BF16
·
Hardware compatibility
Log In to add your hardware
Quantized
# Make sure mlx-lm is installed # pip install --upgrade mlx-lm # Generate text with mlx-lm from mlx_lm import load, generate model, tokenizer = load("MCES10-Software/cpp-qwen3-4B-Instruct-2507") prompt = "Write a story about Einstein" messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, add_generation_prompt=True ) text = generate(model, tokenizer, prompt=prompt, verbose=True)