bilalzafar commited on
Commit
6e4b5f3
·
verified ·
1 Parent(s): 085f6f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -22,11 +22,11 @@ tags:
22
  - Machine-Learning
23
  ---
24
 
25
- # FinAI-BERT
26
- FinAI-BERT is a domain-specific BERT-based model fine-tuned for detecting AI-related disclosures in financial texts.
27
 
28
  ## Intended Use
29
- FinAI-BERT is designed to assist researchers, analysts, and regulators in identifying AI narratives in financial disclosures at the sentence level.
30
 
31
  ## Performance
32
  - Accuracy: 99.37%
@@ -35,7 +35,7 @@ FinAI-BERT is designed to assist researchers, analysts, and regulators in identi
35
  - Brier Score: 0.0000
36
 
37
  ## Training Data
38
- FinAI-BERT was fine-tuned on a manually annotated dataset comprising sentences from U.S. bank annual reports spanning 2015 to 2023. The final training set included a balanced sample of 1,586 sentences—793 labeled as AI-related and 793 as non-AI. The model was initialized using the bert-base-uncased architecture.
39
 
40
  ## Training
41
  | Setting | Value |
@@ -70,8 +70,8 @@ Please cite my paper if you use this model:
70
  ## Usage
71
  ```python
72
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
73
- tokenizer = AutoTokenizer.from_pretrained("bilalzafar/FinAI-BERT")
74
- model = AutoModelForSequenceClassification.from_pretrained("bilalzafar/FinAI-BERT")
75
 
76
  ## Inference Example
77
  from transformers import pipeline
 
22
  - Machine-Learning
23
  ---
24
 
25
+ # BankAI-BERT
26
+ BankAI-BERT is a domain-specific BERT-based model fine-tuned for detecting AI-related disclosures in banking texts.
27
 
28
  ## Intended Use
29
+ BankAI-BERT is designed to assist researchers, analysts, and regulators in identifying AI narratives in financial disclosures at the sentence level.
30
 
31
  ## Performance
32
  - Accuracy: 99.37%
 
35
  - Brier Score: 0.0000
36
 
37
  ## Training Data
38
+ BankAI-BERT was fine-tuned on a manually annotated dataset comprising sentences from U.S. bank annual reports spanning 2015 to 2023. The final training set included a balanced sample of 1,586 sentences—793 labeled as AI-related and 793 as non-AI. The model was initialized using the bert-base-uncased architecture.
39
 
40
  ## Training
41
  | Setting | Value |
 
70
  ## Usage
71
  ```python
72
  from transformers import AutoTokenizer, AutoModelForSequenceClassification
73
+ tokenizer = AutoTokenizer.from_pretrained("bilalzafar/BankAI-BERT")
74
+ model = AutoModelForSequenceClassification.from_pretrained("bilalzafar/BankAI-BERT")
75
 
76
  ## Inference Example
77
  from transformers import pipeline