🚀 DeepSeek Math 7B - Full Fine-tune

GSM8K Open LLM Leaderboard

Model ID: sid172002/deepseek-math-7b-3epoch-678k-fullft

Fully fine-tuned DeepSeek Math 7B on 678K high-quality math problems. +17.8% improvement over base model on GSM8K (64.2% → 82.0%).

📊 Quick Stats

Metric Value
Parameters 7 Billion
Training Steps 63,609
Epochs 3.0
Dataset Size 678,494 samples
GSM8K Score 82.0%
Training Loss 0.6394
Eval Loss 0.6411

🏆 Benchmarks

Benchmark Score Base Improvement
GSM8K 82.0% 64.2% +17.8%
MathBench 92.0% ~70% +22%
MMLU Pending - Leaderboard Eval
MATH TBD 33.2% In Progress

💻 Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "sid172002/deepseek-math-7b-3epoch-678k-fullft",
    torch_dtype=torch.bfloat16,
    device_map="auto",
    trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained(
    "sid172002/deepseek-math-7b-3epoch-678k-fullft",
    trust_remote_code=True
)

# Solve math problem
problem = "What is 2x + 5 = 13?"
prompt = f"Solve step by step:\n\n{problem}\n\nSolution:\n"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True))
Downloads last month
5
Safetensors
Model size
7B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sid172002/deepseek-math-7b-3epoch-678k-fullft

Finetuned
(24)
this model

Datasets used to train sid172002/deepseek-math-7b-3epoch-678k-fullft

Evaluation results