Vishva007's picture
Update README.md
9c9c715 verified
---
language:
- en
license: apache-2.0
base_model: Qwen/Qwen2.5-3B-Instruct
tags:
- finance
- banking
- rbi
- regulation
- india
- qwen2.5
- lora
datasets:
- Vishva007/RBI-Circular-QA-Dataset
metrics:
- accuracy
---
# Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A
A specialized model for answering questions about **Reserve Bank of India (RBI) regulations and banking policies**.
**Performance**: 57.6% accuracy (8.2x improvement over base model's 7%)
## Quick Facts
- 🎯 **Accuracy**: 57.6% on 1000-sample evaluation set
- πŸ“š **Coverage**: 100+ regulation areas (Basel III, FEMA, AML, PSL, etc.)
- πŸš€ **Training**: 47K QA pairs with rephrased variants
- ⚑ **Efficient**: 3B parameters, optimized for deployment
## Performance Highlights
| Category | Base Model | Fine-tuned | Improvement |
|----------|-----------|-----------|-------------|
| Overall | 7.0% | **57.6%** | +50.6% |
| Fact-based | 6.8% | **57.6%** | +50.8% |
| Reasoning | 37.5% | **62.5%** | +25.0% |
**Top Categories** (70%+ accuracy): Anti-Money Laundering, Digital Payments, Government Banking, MSME Finance, Currency Management
## Training Details
- **Method**: LoRA fine-tuning (r=16, alpha=32)
- **Dataset**: [RBI-Circular-QA-Dataset](https://huggingface.co/datasets/Vishva007/RBI-Circular-QA-Dataset) (47K samples)
- **Training**: 1 epoch, 2 hours on NVIDIA L40S
- **Loss**: 0.79 β†’ 0.57 (train), 0.58 (eval)
## Code & Resources
- πŸ’» **Training Code**: [GitHub Repository](https://github.com/vishvaRam/Unsloth-FineTuning.git)
- πŸ“Š **Dataset**: [RBI-Circular-QA-Dataset](https://huggingface.co/datasets/Vishva007/RBI-Circular-QA-Dataset)
- πŸ€— **Base Model**: [Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct)
## πŸ“ Technical Deep Dive
Want to understand the theory and methodology behind this model?
**Read the full article:** [Fine-tuning Qwen 2.5 3B for RBI Regulations: Achieving 8x Performance with Smart Data Augmentation](https://dev.to/vishva_ram/fine-tuning-qwen-25-3b-for-rbi-regulations-achieving-8x-performance-with-smart-data-augmentation-5e38)
The article covers:
- 🧠 **LoRA Theory**: Why and how Low-Rank Adaptation works
- ⚑ **Unsloth Deep Dive**: Technical advantages and performance optimizations
- πŸ“Š **Data Augmentation Strategy**: The rephrasing technique that delivered 40% of improvement
- πŸŽ“ **Hyperparameter Analysis**: Detailed explanation of every training choice
- πŸ“ˆ **Evaluation Methodology**: Stratified sampling and LLM-as-judge approach
- πŸ”¬ **Ablation Studies**: What really mattered for the 8x improvement
Perfect for ML engineers who want to replicate or adapt this approach for their own domain-specific fine-tuning projects.
## Use Cases
βœ… Banking compliance chatbots
βœ… Regulatory Q&A systems
βœ… Training tools for banking professionals
βœ… RBI circular analysis
⚠️ Not for: Legal compliance decisions (requires expert review), real-time updates
## Limitations
- Knowledge cutoff: 2024 regulations
- 57.6% accuracy means ~42% of complex queries need verification
- Optimized for English only
## Citation
```
@misc{
qwen25-3b-rbi-qa,
author = {Vishva007},
title = {Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A},
year = {2025},
publisher = {HuggingFace},
howpublished = {\url{https://huggingface.co/Vishva007/Qwen2.5-3B-Instruct-RBI-QA}},
}
```
## License
Apache 2.0 (inherits from base model)
---
**Author**: [@Vishva007](https://huggingface.co/Vishva007) | **Updated**: November 2025