--- language: en license: mit tags: - financial-qa - distilgpt2 - fine-tuned datasets: - financial-qa metrics: - perplexity --- # Financial QA Fine-Tuned Model This model is a fine-tuned version of `distilgpt2` on financial question-answering data from Allstate's financial reports. ## Model description The model was fine-tuned to answer questions about Allstate's financial reports and performance. ## Intended uses & limitations This model is intended to be used for answering factual questions about Allstate's financial reports for 2022-2023. It should not be used for financial advice or decision-making without verification from original sources. ## Training data The model was trained on a custom dataset of financial QA pairs derived from Allstate's 10-K reports. ## Training procedure The model was fine-tuned using the `Trainer` class from Hugging Face's Transformers library with the following parameters: - Learning rate: default - Batch size: 2 - Number of epochs: 3 ## Evaluation results The model achieved a final training loss of 0.44 and validation loss of 0.43. ## Limitations and bias This model has limited knowledge only of Allstate's financial data and cannot answer questions about other companies or financial topics outside its training data.