SpermAI's picture
Create README.md
fab3d40 verified
metadata
license: apache-2.0
language:
  - en
base_model:
  - mistralai/Ministral-3-3B-Instruct-2512

SpermLLM Ministral3 3B

A fine-tuned version of Ministral-3-3B-Instruct-2512 trained with Unsloth on a carefully curated mix of SOTA reasoning and instruction-following datasets.

Optimized for math, code, science, and general reasoning — competitive with models 2-3x its size.


Model Details

Property Value
Base Model mistralai/Ministral-3-3B-Instruct-2512
Model Type Causal Language Model (Decoder-only)
Parameters 3.84B
Trainable Parameters 135M (3.39% via LoRA)
Architecture Mistral with Sliding Window Attention
Context Length 8,192 tokens
Training Hardware NVIDIA B200 (180GB VRAM)
Training Framework Unsloth + TRL SFTTrainer
Precision BFloat16
Quantization 4-bit QLoRA during training
License Apache 2.0

This AI is (sorta SOTA) for it's size, as it can create multiple stuff without errors, this is our latest model