πŸ’° Qwen2.5 Financial Reasoning Model (LoRA Fine-Tuned)

A parameter-efficient fine-tuned version of Qwen2.5-3B-Instruct for step-by-step financial reasoning and numerical problem solving.


πŸ“‹ Model Details

Model Description

This model is a fine-tuned version of Qwen2.5-3B-Instruct, trained to perform step-by-step financial reasoning and calculations using Low-Rank Adaptation (LoRA).

It is designed to:

  • πŸ”’ Solve finance-related numerical problems
  • 🧠 Provide structured, step-by-step reasoning explanations
  • πŸ“š Support educational use cases around core financial concepts
Field Details
Developed by Parth Kadoo
Model type Causal Language Model (LLM)
Base model unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit
Fine-tuning method LoRA (PEFT)
Task Financial reasoning & text generation
Language(s) English
License MIT
PEFT Version 0.18.1

Model Sources


🎯 Uses

βœ… Direct Use

This model can be used directly for:

  • Financial question answering (e.g., interest calculations, ratio analysis)
  • Step-by-step numerical problem solving
  • Educational purposes β€” learning and understanding finance concepts
  • Prototyping finance-related NLP applications

πŸ”„ Downstream Use

This model can be plugged into larger pipelines for:

  • Finance-focused chatbots or tutoring assistants
  • Automated financial report summarization tools
  • RAG (Retrieval-Augmented Generation) systems combined with financial knowledge bases

β›” Out-of-Scope Use

This model is not intended for:

  • Real-world financial advice or investment decisions
  • High-stakes trading or portfolio management systems
  • Legal, regulatory, or compliance use cases
  • Any production application without human oversight and validation

⚠️ Bias, Risks, and Limitations

  • Trained on a relatively small dataset (~5,500 samples) β€” may not generalize to all financial domains
  • May produce confident but incorrect answers, especially on complex or unseen problems
  • Limited to English language only
  • Can hallucinate steps or results in edge cases
  • May reflect biases present in the FinQA training dataset
  • Not a substitute for professional financial advice

Recommendations

  • Use for learning and experimentation only
  • Always verify numerical outputs independently before acting on them
  • Combine with external validation or rule-based checks in real applications
  • Users should be made aware of the model's limitations before deployment

πŸš€ How to Get Started with the Model

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "Parth7007/qwen2.5-finance-model"  

tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

prompt = "If you invest β‚Ή10,000 at 10% annual interest for 2 years, what is the final amount?"

inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))

πŸ’‘ Example

Input:

If you invest β‚Ή10,000 at 10% annual interest for 2 years, what is the final amount?

Output:

Step 1: Use the compound interest formula: A = P(1 + r)^t
Step 2: A = 10,000 Γ— (1 + 0.10)^2
Step 3: A = 10,000 Γ— 1.21

Final Answer: β‚Ή12,100

πŸ‹οΈ Training Details

Training Data

  • Dataset: dreamerdeo/finqa
  • Description: FinQA is a financial reasoning dataset containing numerical and conceptual questions derived from earnings reports and financial documents. It requires multi-step reasoning over structured and unstructured financial data.
  • Size: ~5,500 training samples
  • Preprocessing: Prompts were formatted as instruction-response pairs for supervised fine-tuning (SFT).

Training Procedure

Training Hyperparameters

Parameter Value
Training regime 4-bit quantization (NF4) with bf16 mixed precision
Fine-tuning method LoRA (Low-Rank Adaptation)
Trainable parameters ~0.48% of total parameters
Epochs 2
Final training loss ~0.82
Optimizer AdamW (via Unsloth)

Speeds, Sizes, Times

Detail Value
Hardware Google Colab GPU (T4)
Cloud Provider Google Colab
Framework Transformers + PEFT (Unsloth)
Precision 4-bit quantization (bitsandbytes)
Training time ~1–2 hours (estimated on Colab T4)

πŸ“Š Evaluation

Testing Data

Evaluation was performed on held-out examples from the dreamerdeo/finqa dataset.

Factors

  • Standard financial calculations (interest, percentages, ratios)
  • Multi-step numerical reasoning
  • Structured explanation quality

Metrics

⚠️ No formal benchmark evaluation has been conducted yet.
Qualitative observations were used to assess model performance.

Results

Qualitative Observations

  • βœ… Produces correct results for standard financial calculations
  • βœ… Demonstrates clear step-by-step reasoning capability
  • βœ… Outperforms the base model on structured finance questions
  • ⚠️ May struggle with highly complex or domain-specific edge cases

Summary

The model shows meaningful improvement over the base Qwen2.5-3B-Instruct on financial reasoning tasks, particularly for multi-step numerical problems. Formal benchmarking (e.g., on FinQA test split) is planned for future iterations.


πŸ—οΈ Technical Specifications

Model Architecture and Objective

  • Architecture: Decoder-only Transformer (Qwen2.5 series)
  • Objective: Causal language modeling with SFT on financial instruction-response pairs
  • Adaptation: LoRA adapters injected into attention layers for parameter-efficient fine-tuning

Compute Infrastructure

Hardware

  • Google Colab T4 GPU (15GB VRAM)

Software

Library Version
transformers Latest stable
peft 0.18.1
trl Latest stable
unsloth Latest stable
bitsandbytes Latest stable

🌍 Environmental Impact

Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).

Field Details
Hardware Type NVIDIA T4 GPU (Google Colab)
Hours used ~1–2 hours (estimated)
Cloud Provider Google (Colab)
Compute Region US (estimated)
Carbon Emitted Minimal (short training run on shared cloud GPU)

πŸ“œ Citation

If you use this model or find it helpful, please consider citing the original FinQA dataset:

BibTeX:

@inproceedings{chen2021finqa,
  title     = {FinQA: A Dataset of Numerical Reasoning over Financial Data},
  author    = {Chen, Zhiyu and Chen, Wenhu and Sha, Charese and Wang, Jianshu and Wang, William Yang},
  booktitle = {Proceedings of EMNLP 2021},
  year      = {2021}
}

APA: Chen, Z., Chen, W., Sha, C., Wang, J., & Wang, W. Y. (2021). FinQA: A Dataset of Numerical Reasoning over Financial Data. EMNLP 2021.


πŸ“– Glossary

Term Definition
LoRA Low-Rank Adaptation β€” a PEFT method that trains small adapter matrices instead of full model weights
PEFT Parameter-Efficient Fine-Tuning β€” techniques to fine-tune large models with minimal trainable parameters
SFT Supervised Fine-Tuning β€” training on labeled instruction-response pairs
4-bit quantization Reducing model weights to 4-bit precision to lower memory usage
FinQA A financial reasoning benchmark requiring numerical reasoning over earnings reports

πŸ‘€ Model Card Authors

Parth Kadoo

πŸ“¬ Model Card Contact

For questions, feedback, or collaboration, feel free to reach out via Hugging Face.


Framework Versions

  • peft == 0.18.1
Downloads last month
44
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Dataset used to train Parth7007/qwen2.5-finance-model

Paper for Parth7007/qwen2.5-finance-model