|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: apache-2.0 |
|
|
base_model: Qwen/Qwen2.5-3B-Instruct |
|
|
tags: |
|
|
- finance |
|
|
- banking |
|
|
- rbi |
|
|
- regulation |
|
|
- india |
|
|
- qwen2.5 |
|
|
- lora |
|
|
datasets: |
|
|
- Vishva007/RBI-Circular-QA-Dataset |
|
|
metrics: |
|
|
- accuracy |
|
|
--- |
|
|
|
|
|
# Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A |
|
|
|
|
|
A specialized model for answering questions about **Reserve Bank of India (RBI) regulations and banking policies**. |
|
|
|
|
|
**Performance**: 57.6% accuracy (8.2x improvement over base model's 7%) |
|
|
|
|
|
## Quick Facts |
|
|
|
|
|
- π― **Accuracy**: 57.6% on 1000-sample evaluation set |
|
|
- π **Coverage**: 100+ regulation areas (Basel III, FEMA, AML, PSL, etc.) |
|
|
- π **Training**: 47K QA pairs with rephrased variants |
|
|
- β‘ **Efficient**: 3B parameters, optimized for deployment |
|
|
|
|
|
|
|
|
## Performance Highlights |
|
|
|
|
|
| Category | Base Model | Fine-tuned | Improvement | |
|
|
|----------|-----------|-----------|-------------| |
|
|
| Overall | 7.0% | **57.6%** | +50.6% | |
|
|
| Fact-based | 6.8% | **57.6%** | +50.8% | |
|
|
| Reasoning | 37.5% | **62.5%** | +25.0% | |
|
|
|
|
|
**Top Categories** (70%+ accuracy): Anti-Money Laundering, Digital Payments, Government Banking, MSME Finance, Currency Management |
|
|
|
|
|
## Training Details |
|
|
|
|
|
- **Method**: LoRA fine-tuning (r=16, alpha=32) |
|
|
- **Dataset**: [RBI-Circular-QA-Dataset](https://huggingface.co/datasets/Vishva007/RBI-Circular-QA-Dataset) (47K samples) |
|
|
- **Training**: 1 epoch, 2 hours on NVIDIA L40S |
|
|
- **Loss**: 0.79 β 0.57 (train), 0.58 (eval) |
|
|
|
|
|
## Code & Resources |
|
|
|
|
|
- π» **Training Code**: [GitHub Repository](https://github.com/vishvaRam/Unsloth-FineTuning.git) |
|
|
- π **Dataset**: [RBI-Circular-QA-Dataset](https://huggingface.co/datasets/Vishva007/RBI-Circular-QA-Dataset) |
|
|
- π€ **Base Model**: [Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) |
|
|
|
|
|
## π Technical Deep Dive |
|
|
|
|
|
Want to understand the theory and methodology behind this model? |
|
|
|
|
|
**Read the full article:** [Fine-tuning Qwen 2.5 3B for RBI Regulations: Achieving 8x Performance with Smart Data Augmentation](https://dev.to/vishva_ram/fine-tuning-qwen-25-3b-for-rbi-regulations-achieving-8x-performance-with-smart-data-augmentation-5e38) |
|
|
|
|
|
The article covers: |
|
|
- π§ **LoRA Theory**: Why and how Low-Rank Adaptation works |
|
|
- β‘ **Unsloth Deep Dive**: Technical advantages and performance optimizations |
|
|
- π **Data Augmentation Strategy**: The rephrasing technique that delivered 40% of improvement |
|
|
- π **Hyperparameter Analysis**: Detailed explanation of every training choice |
|
|
- π **Evaluation Methodology**: Stratified sampling and LLM-as-judge approach |
|
|
- π¬ **Ablation Studies**: What really mattered for the 8x improvement |
|
|
|
|
|
Perfect for ML engineers who want to replicate or adapt this approach for their own domain-specific fine-tuning projects. |
|
|
|
|
|
## Use Cases |
|
|
|
|
|
β
Banking compliance chatbots |
|
|
β
Regulatory Q&A systems |
|
|
β
Training tools for banking professionals |
|
|
β
RBI circular analysis |
|
|
|
|
|
β οΈ Not for: Legal compliance decisions (requires expert review), real-time updates |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Knowledge cutoff: 2024 regulations |
|
|
- 57.6% accuracy means ~42% of complex queries need verification |
|
|
- Optimized for English only |
|
|
|
|
|
## Citation |
|
|
|
|
|
``` |
|
|
@misc{ |
|
|
qwen25-3b-rbi-qa, |
|
|
author = {Vishva007}, |
|
|
title = {Qwen2.5-3B-Instruct Fine-tuned for RBI Regulations Q&A}, |
|
|
year = {2025}, |
|
|
publisher = {HuggingFace}, |
|
|
howpublished = {\url{https://huggingface.co/Vishva007/Qwen2.5-3B-Instruct-RBI-QA}}, |
|
|
} |
|
|
``` |
|
|
|
|
|
|
|
|
## License |
|
|
|
|
|
Apache 2.0 (inherits from base model) |
|
|
|
|
|
--- |
|
|
|
|
|
**Author**: [@Vishva007](https://huggingface.co/Vishva007) | **Updated**: November 2025 |
|
|
|
|
|
|
|
|
|