--- license: mit tags: - finance - financial-qa - finetuned-llm - phi3 - phi-3-mini - fintech - accounting - banking - investment - risk-analysis - lora - merged-model language: - en pipeline_tag: text-generation base_model: microsoft/phi-3-mini-4k-instruct model_name: finance_llm_full model_creator: devAnurag pretty_name: Finance LLM Full --- # ๐Ÿ’Ž Finance LLM Full โ€” Next-Gen Financial Intelligence Model

Finance LLM Logo

๐Ÿ’ผ Finance LLM Full

A next-generation Financial Intelligence Model
Fine-Tuned, Merged & Optimized for Real-World Finance

--- **Finance LLM Full** is a high-performance, fully merged financial Large Language Model (LLM) designed to deliver **crystal-clear, accurate, and structured financial reasoning**. It is trained using **LoRA fine-tuning** on top of **Phi-3 Mini 4K Instruct**, and later **merged into a single standalone model** for seamless deployment. This model specializes in **Finance, Accounting, Banking, Investment, Stock Markets, and Business Analysis** โ€” making it ideal for **FinTech products, AI advisors, investment copilots, and enterprise bots**. --- # โšก Why Finance LLM Full is Special ### ๐Ÿ”น 1. Purpose-Built For Finance Unlike general LLMs, this model deeply understands: - Balance Sheet Interpretation - Profit & Loss Breakdown - Cashflow Logic - EBITDA / EPS / ROE / DCF - Risk & Return Analysis - Banking, Loans, Limits, Credit Rules - Valuation Basics - Investment & Portfolio Concepts ### ๐Ÿ”น 2. Merged Model โ†’ One File, Zero Hassle โœ” No LoRA needed โœ” No adapter loading โœ” Direct plug-and-play โœ” Works on CPU / GPU / Colab / Docker ### ๐Ÿ”น 3. Small Model โ†’ Big Capability Powered by **Phi-3 Mini**, optimized for: - Low latency - Low VRAM/RAM usage - Clean, structured answers - High domain accuracy --- # ๐Ÿงช Quick Start (Copy & Run) ```python from transformers import AutoTokenizer, AutoModelForCausalLM import torch model_id = "devAnurag/finance_llm_full" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16) prompt = "Explain the difference between EBITDA and Net Profit." inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=150) print(tokenizer.decode(outputs[0], skip_special_tokens=True))