Pythagoras-Prover-SFT-32B

Pythagoras-Prover-SFT-32B is a Qwen3-32B based causal language model fine-tuned for formal mathematical reasoning and theorem proving.

This repository contains the merged model weights. The LoRA adapter from checkpoint-6242 was merged into the Qwen3-32B base model with LLaMA-Factory.

Model Details

  • Base model: Qwen/Qwen3-32B
  • Fine-tuning method: LoRA SFT, merged into the base weights for this release
  • Training framework: LLaMA-Factory
  • Source checkpoint: saves/qwen3-32b/lora_v2/sft_data_updated_adaptive_clean_lora/checkpoint-6242
  • Export precision: follows the source Qwen3-32B checkpoint dtype

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model_id = "Pythagoras-LM/Pythagoras-Prover-SFT-32B"

tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
    model_id,
    torch_dtype="auto",
    device_map="auto",
    trust_remote_code=True,
)

Intended Use

This model is intended for research and development workflows involving formal reasoning, mathematical problem solving, and Lean-style theorem proving.

Downloads last month
46
Safetensors
Model size
33B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Pythagoras-LM/Pythagoras-Prover-SFT-32B

Base model

Qwen/Qwen3-32B
Finetuned
(508)
this model