keesephillips commited on
Commit
2f754cc
·
verified ·
1 Parent(s): 6d9f2c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -28
README.md CHANGED
@@ -1,29 +1,41 @@
1
- # AIPI 590 Large Language Models
2
- ## Project 1 - Fine Tuning LLM
3
-
4
- ### Files:
5
- - model.ipynb
6
- - notebook containing the code for fine tuning the Llama 3 model using QLoRa
7
- - data/train.json
8
- - json file containing the training set provided in the FINQA paper
9
- - data/test.json
10
- - json file containing the validation set provided in the FINQA paper
11
-
12
- ### Process:
13
- The focal property of interest is analysis financial documents for numerical reasoning. Specifically numerical reasoning over quarterly financial filings with the SEC. The Llama-3-8B model was chosen to fine tune using the QLoRa approach. This approach was chosen due to the paper's findings of a performance increase while utilizing minimal memory and hardware. The aggressive quantization seemed to significantly decreased training time while offering increased performance on financial analysis.
14
-
15
- ### Evaluation:
16
- #### Rogue Score
17
- | ROUGE Score | Base Model | QLoRa Fine Tuned Model |
18
- | ------------- | ------------- | ------------- |
19
- | ROUGE-1 | 0.05104785 | 0.25257307 |
20
- | ROUGE-2 | 0.01158752 | 0.10479990 |
21
- | ROUGE-L | 0.05104785 | 0.25175429 |
22
-
23
- ### Collaborators:
24
- - Keese Phillips
25
-
26
- ### Attribution:
27
- - [FINQA: A Dataset of Numerical Reasoning over Financial Data](https://arxiv.org/pdf/2109.00122v3)
28
- - [LORA: Low-Rank Adaptation of Large Language Models](https://arxiv.org/pdf/2106.09685)
 
 
 
 
 
 
 
 
 
 
 
 
29
  - [QLORA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/pdf/2305.14314)
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - en
5
+ metrics:
6
+ - rouge
7
+ base_model:
8
+ - meta-llama/Meta-Llama-3-8B
9
+ pipeline_tag: text-generation
10
+ tags:
11
+ - finance
12
+ ---
13
+ # AIPI 590 Large Language Models
14
+ ## Project 1 - Fine Tuning LLM
15
+
16
+ ### Files:
17
+ - model.ipynb
18
+ - notebook containing the code for fine tuning the Llama 3 model using QLoRa
19
+ - data/train.json
20
+ - json file containing the training set provided in the FINQA paper
21
+ - data/test.json
22
+ - json file containing the validation set provided in the FINQA paper
23
+
24
+ ### Process:
25
+ The focal property of interest is analysis financial documents for numerical reasoning. Specifically numerical reasoning over quarterly financial filings with the SEC. The Llama-3-8B model was chosen to fine tune using the QLoRa approach. This approach was chosen due to the paper's findings of a performance increase while utilizing minimal memory and hardware. The aggressive quantization seemed to significantly decreased training time while offering increased performance on financial analysis.
26
+
27
+ ### Evaluation:
28
+ #### Rouge Score
29
+ | ROUGE Score | Base Model | QLoRa Fine Tuned Model |
30
+ | ------------- | ------------- | ------------- |
31
+ | ROUGE-1 | 0.05104785 | 0.25257307 |
32
+ | ROUGE-2 | 0.01158752 | 0.10479990 |
33
+ | ROUGE-L | 0.05104785 | 0.25175429 |
34
+
35
+ ### Collaborators:
36
+ - Keese Phillips
37
+
38
+ ### Attribution:
39
+ - [FINQA: A Dataset of Numerical Reasoning over Financial Data](https://arxiv.org/pdf/2109.00122v3)
40
+ - [LORA: Low-Rank Adaptation of Large Language Models](https://arxiv.org/pdf/2106.09685)
41
  - [QLORA: Efficient Finetuning of Quantized LLMs](https://arxiv.org/pdf/2305.14314)