PyTorch
English
llama
ZixuanKe commited on
Commit
491d4c7
ยท
verified ยท
1 Parent(s): bc698fe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -3
README.md CHANGED
@@ -1,3 +1,43 @@
1
- ---
2
- license: cc-by-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ datasets:
4
+ - Salesforce/FinTrain
5
+ - Salesforce/FinEval
6
+ language:
7
+ - en
8
+ base_model:
9
+ - meta-llama/Meta-Llama-3-8B-Instruct
10
+ ---
11
+
12
+ # ๐Ÿ’ฐ Demystifying Domain-adaptive Post-training for Financial LLMs
13
+
14
+ This is the **finance-specific large language model** trained using the recipe described in our paper:
15
+ ๐Ÿ“„ [*Demystifying Domain-adaptive Post-training for Financial LLMs*](https://arxiv.org/abs/2501.04961)
16
+
17
+ For more details, please check the following resources:
18
+
19
+ - ๐ŸŒ **Project Page:** [https://vincent950129.github.io/adapt-llm/](https://vincent950129.github.io/adapt-llm/)
20
+ - ๐Ÿ“š **Training Data:** [https://huggingface.co/datasets/Salesforce/FinTrain](https://huggingface.co/datasets/Salesforce/FinTrain)
21
+ - ๐Ÿง  **Evaluation Data:** [https://huggingface.co/datasets/Salesforce/FinEval](https://huggingface.co/datasets/Salesforce/FinEval)
22
+ - ๐Ÿ’ป **Code Repository:** [https://github.com/SalesforceAIResearch/FinDAP](https://github.com/SalesforceAIResearch/FinDAP)
23
+
24
+
25
+ ### Ethical Considerations
26
+ Users need to make their own assessment regarding any obligations or responsibilities under the corresponding licenses or terms and conditions pertaining to the original datasets and data. This release is for research purposes only in support of an academic paper.
27
+
28
+
29
+ ## Citation
30
+
31
+ If you find our project helpful, please consider citing our paper ๐Ÿ˜Š
32
+
33
+ ```
34
+ @misc{ke2025demystifyingdomainadaptiveposttrainingfinancial,
35
+ title={Demystifying Domain-adaptive Post-training for Financial LLMs},
36
+ author={Zixuan Ke and Yifei Ming and Xuan-Phi Nguyen and Caiming Xiong and Shafiq Joty},
37
+ year={2025},
38
+ eprint={2501.04961},
39
+ archivePrefix={arXiv},
40
+ primaryClass={cs.CL},
41
+ url={https://arxiv.org/abs/2501.04961},
42
+ }
43
+ ```