Mistral Nemo 12B - Humanities Distilled (GGUF)

This model is a distilled version of Mistral Nemo 12B, fine-tuned on humanities content including:

  • MMLU Humanities datasets (history, philosophy, religion, ethics)
  • Synthetic data generated from GPT-OSS-120B

Training Details

  • Base Model: mistralai/Mistral-Nemo-Instruct-2407
  • Teacher Model: openai/gpt-oss-120b
  • Training Method: LoRA with 8-bit quantization
  • Dataset: MMLU humanities + synthetic data
  • Specialization: History, philosophy, and humanities understanding

Available Quantizations

  • FP16 (model-f16.gguf): Full precision, ~24GB
  • Q8_0 (model-Q8_0.gguf): 8-bit quantization, ~13GB, high quality
  • Q4_K_M (model-Q4_K_M.gguf): 4-bit quantization, ~7GB, recommended for most users

Usage with llama.cpp

# Download model
huggingface-cli download ambrosfitz/mistral-nemo-gguf model-Q4_K_M.gguf --local-dir ./models

# Run inference
./llama.cpp/main -m ./models/model-Q4_K_M.gguf -p "Question: What was the Renaissance?\n\nAnswer:" -n 256

Usage with Ollama

# Create Modelfile
cat > Modelfile <<EOF
FROM ./model-Q4_K_M.gguf
PARAMETER temperature 0.7
PARAMETER top_p 0.9
EOF

# Create model
ollama create mistral-humanities -f Modelfile

# Run
ollama run mistral-humanities "What was the Renaissance?"

Example Prompts

Question: What were the main causes of World War I?

Answer:

Question: Explain the philosophical ideas of the Enlightenment.

Answer:

Question: Who wrote 'The Republic' and what is it about?

Answer:

Limitations

This model is specialized for humanities topics and may not perform as well on:

  • Technical/scientific questions
  • Mathematics
  • Coding
  • Current events (knowledge cutoff applies)

License

Apache 2.0 (same as base model)

Downloads last month
23
GGUF
Model size
12B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ambrosfitz/mistral-nemo-humanities-gguf