FinBloom_7B / README.md
Chaitanya14's picture
Update README.md
2d6ace3 verified
---
base_model: bigscience/bloom-7b1
library_name: peft
---
FinBloom 7B is a specialized Large Language Model engineered for high-performance financial NLP applications. Built upon the BLOOM-7B architecture, the model has undergone extensive fine-tuning using a robust corpus of historical financial data, including comprehensive news archives from Reuters and DPA, as well as several years of official SEC filings. This targeted training enables the model to excel at complex downstream financial analysis and domain-specific reasoning.
## Breakup of datasets used for finetuning
| Dataset | Documents | Mean Words | Mean Tokens | Time Period
|:---|:---|:---|:---|:---|
| Reuters news | 14,574,641 | 369.23 | 459.09 | 1st Jan 2003-31st Dec 2012 |
| DPA news | 387,187 | 286.20 | 390.37| 1st Jun 2001-31st May 2011
| SEC filings | 12,238,570 | 379.96 | 536.56 | 31st Mar 2009-31st Oct 2023
## How to Get Started with the Model
Use the code below to get started with the model.
```python
from peft import PeftModel, PeftConfig
from transformers import AutoModelForCausalLM, AutoTokenizer
peft_model_id = "Chaitanya14/FinBloom_7B"
config = PeftConfig.from_pretrained(peft_model_id)
tokenizer = AutoTokenizer.from_pretrained(config.base_model_name_or_path)
model = AutoModelForCausalLM.from_pretrained(config.base_model_name_or_path)
model = PeftModel.from_pretrained(model, peft_model_id)
```
## Citation
If you use the FinBloom 7B LLM, please cite with the following BibTex entry:
```
@misc{sinha2025finbloomknowledgegroundinglarge,
title={FinBloom: Knowledge Grounding Large Language Model with Real-time Financial Data},
author={Ankur Sinha and Chaitanya Agarwal and Pekka Malo},
year={2025},
eprint={2502.18471},
archivePrefix={arXiv},
primaryClass={cs.IR},
url={https://arxiv.org/abs/2502.18471},
}
```