Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -7,9 +7,28 @@ sdk: static
|
|
| 7 |
pinned: false
|
| 8 |
---
|
| 9 |
|
| 10 |
-
|
|
|
|
| 11 |
|
| 12 |
-
<div style="
|
| 13 |
-
|
| 14 |
-
|
| 15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
pinned: false
|
| 8 |
---
|
| 9 |
|
| 10 |
+
# README π»
|
| 11 |
+
#### A repository of Financial NLP Models and Benchmarks
|
| 12 |
|
| 13 |
+
<div style="background: linear-gradient(to right, red, blue); padding: 10px; border-radius: 10px;">
|
| 14 |
+
|
| 15 |
+
### π **Stage 1 Release** π
|
| 16 |
+
|
| 17 |
+
We are excited to announce the release of a specialised suite of **LLMs** designed specifically for accounting and finance. FinText models have been **pre-trained** on domain-specific historical data to overcome common issues such as **look-ahead bias** and **information leakage**. These models are tailored to enhance the performance of financial studies and analyses.
|
| 18 |
+
|
| 19 |
+
π‘ **Key Features:**
|
| 20 |
+
- **Domain-Specific Training:** FinText utilises diverse financial datasets such as news articles, regulatory filings, IP records, corporate speeches (ECB, FED), and more.
|
| 21 |
+
- **Time-Period Specific Models:** Separate models are pre-trained for each year from **2007 to 2023**, ensuring the utmost precision and historical relevance.
|
| 22 |
+
- **RoBERTa Architecture:** The suite includes both a **base model** with **125 million parameters** and a **smaller variant** with **51 million parameters**βtotalling 34 pre-trained models. π―
|
| 23 |
+
|
| 24 |
+
π **Datasets Used:**
|
| 25 |
+
- News articles π
|
| 26 |
+
- Regulatory filings ποΈ
|
| 27 |
+
- IP records π§βπΌ
|
| 28 |
+
- ECB & FED speeches π£οΈ
|
| 29 |
+
- Corporate event transcripts π
|
| 30 |
+
- Wikipedia π§ (for general knowledge)
|
| 31 |
+
|
| 32 |
+
Stay tuned for further updates and additions to FinText, as we continue refining and expanding our offerings for the financial and academic communities! πβ¨
|
| 33 |
+
|
| 34 |
+
</div>
|