NPC Fin 32B SFT β Financial Reasoning LLM
A domain-specific financial reasoning model fine-tuned from Qwen2.5-32B-Instruct using QLoRA, focused on crypto market analysis, macro reasoning, and multi-step financial logic.
Model Details
| Parameter | Value |
|---|---|
| Base Model | Qwen/Qwen2.5-32B-Instruct |
| Method | QLoRA (4-bit NF4 quantization) |
| LoRA Rank | 64 |
| LoRA Alpha | 128 |
| LoRA Dropout | 0.05 |
| Target Modules | q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj |
| Training Framework | Unsloth + trl SFTTrainer |
| Precision | float16 |
| Max Sequence Length | 4,096 tokens |
Training Data
- 32,496 SFT examples (59.7M tokens)
- 5 domain-specific tags:
crypto_signalβ real-time market signal analysis and trade reasoningcrypto_generalβ broad crypto ecosystem knowledgelogic_treeβ multi-path reasoning with correct and incorrect branchesstocks_macroβ equities and macroeconomic analysiscross_marketβ cross-asset correlation and regime detection
- Synthetic data generated via HF Inference API (Qwen2.5-72B-Instruct) at zero incremental cost
- Source signals exported from production MongoDB (btunified database)
- MinHash deduplication applied, quality-filtered with automated scoring
Training Configuration
| Parameter | Value |
|---|---|
| Optimizer | AdamW 8-bit |
| Learning Rate | 2e-4 |
| LR Schedule | Cosine decay |
| Warmup Ratio | 0.05 |
| Weight Decay | 0.01 |
| Batch Size | 4 |
| Gradient Accumulation | 8 (effective batch size 32) |
| Epochs | 3 |
| Gradient Checkpointing | Enabled |
| Hardware | NVIDIA A40 48GB (RunPod) |
Evaluation
| Benchmark | Score |
|---|---|
| CryptoQA (custom, 500 questions) | 93.6% |
CryptoQA covers: token fundamentals, DeFi mechanics, on-chain analytics interpretation, market regime identification, and risk assessment.
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
# Load base model
base_model = AutoModelForCausalLM.from_pretrained(
"Qwen/Qwen2.5-32B-Instruct",
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-32B-Instruct")
# Load LoRA adapter
model = PeftModel.from_pretrained(base_model, "ramankrishna10/npc-fin-32b-sft")
messages = [
{"role": "system", "content": "You are a financial reasoning assistant."},
{"role": "user", "content": "Analyze the risk/reward of entering a long ETH position given declining on-chain activity but increasing institutional inflows."}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=512, temperature=0.7)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Intended Use
- Financial market analysis and reasoning
- Crypto signal interpretation and trade logic
- Multi-step reasoning over market scenarios
- Research and educational purposes
Limitations
- This is the SFT base model only β it does not include tool-use or identity fine-tuning
- Trained primarily on crypto/DeFi data; performance on traditional equities may be lower
- Not intended as financial advice β outputs are AI-generated analysis
- Single-GPU training (A40) β not trained at cluster scale
- May hallucinate token names or market data not present in training set
Related Models
- npc-fin-prm-7b β Process Reward Model for step-level reasoning verification
Citation
@misc{bachu2026npcfin,
title={NPC Fin 32B SFT: Domain-Specific Financial Reasoning via QLoRA},
author={Ramakrishna Bachu},
year={2026},
url={https://huggingface.co/ramankrishna10/npc-fin-32b-sft}
}
Author
Ramakrishna Bachu β GitHub | LinkedIn
Part of the NPC Model Family by Bottensor.
- Downloads last month
- 33