NPC Fin 32B SFT β€” Financial Reasoning LLM

A domain-specific financial reasoning model fine-tuned from Qwen2.5-32B-Instruct using QLoRA, focused on crypto market analysis, macro reasoning, and multi-step financial logic.

Model Details

Parameter Value
Base Model Qwen/Qwen2.5-32B-Instruct
Method QLoRA (4-bit NF4 quantization)
LoRA Rank 64
LoRA Alpha 128
LoRA Dropout 0.05
Target Modules q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
Training Framework Unsloth + trl SFTTrainer
Precision float16
Max Sequence Length 4,096 tokens

Training Data

  • 32,496 SFT examples (59.7M tokens)
  • 5 domain-specific tags:
    • crypto_signal β€” real-time market signal analysis and trade reasoning
    • crypto_general β€” broad crypto ecosystem knowledge
    • logic_tree β€” multi-path reasoning with correct and incorrect branches
    • stocks_macro β€” equities and macroeconomic analysis
    • cross_market β€” cross-asset correlation and regime detection
  • Synthetic data generated via HF Inference API (Qwen2.5-72B-Instruct) at zero incremental cost
  • Source signals exported from production MongoDB (btunified database)
  • MinHash deduplication applied, quality-filtered with automated scoring

Training Configuration

Parameter Value
Optimizer AdamW 8-bit
Learning Rate 2e-4
LR Schedule Cosine decay
Warmup Ratio 0.05
Weight Decay 0.01
Batch Size 4
Gradient Accumulation 8 (effective batch size 32)
Epochs 3
Gradient Checkpointing Enabled
Hardware NVIDIA A40 48GB (RunPod)

Evaluation

Benchmark Score
CryptoQA (custom, 500 questions) 93.6%

CryptoQA covers: token fundamentals, DeFi mechanics, on-chain analytics interpretation, market regime identification, and risk assessment.

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel

# Load base model
base_model = AutoModelForCausalLM.from_pretrained(
    "Qwen/Qwen2.5-32B-Instruct",
    torch_dtype="auto",
    device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-32B-Instruct")

# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "ramankrishna10/npc-fin-32b-sft")

messages = [
    {"role": "system", "content": "You are a financial reasoning assistant."},
    {"role": "user", "content": "Analyze the risk/reward of entering a long ETH position given declining on-chain activity but increasing institutional inflows."}
]

text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Intended Use

  • Financial market analysis and reasoning
  • Crypto signal interpretation and trade logic
  • Multi-step reasoning over market scenarios
  • Research and educational purposes

Limitations

  • This is the SFT base model only β€” it does not include tool-use or identity fine-tuning
  • Trained primarily on crypto/DeFi data; performance on traditional equities may be lower
  • Not intended as financial advice β€” outputs are AI-generated analysis
  • Single-GPU training (A40) β€” not trained at cluster scale
  • May hallucinate token names or market data not present in training set

Related Models

  • npc-fin-prm-7b β€” Process Reward Model for step-level reasoning verification

Citation

@misc{bachu2026npcfin,
  title={NPC Fin 32B SFT: Domain-Specific Financial Reasoning via QLoRA},
  author={Ramakrishna Bachu},
  year={2026},
  url={https://huggingface.co/ramankrishna10/npc-fin-32b-sft}
}

Author

Ramakrishna Bachu β€” GitHub | LinkedIn

Part of the NPC Model Family by Bottensor.

Downloads last month
33
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for ramankrishna10/npc-fin-32b-sft

Base model

Qwen/Qwen2.5-32B
Adapter
(87)
this model