Fill-Mask
Transformers
PyTorch
German
bert
scherrmann commited on
Commit
74100d7
·
1 Parent(s): f868c50

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -26
README.md CHANGED
@@ -5,38 +5,53 @@ language:
5
  ---
6
  # German FinBERT (Further Pre-trained Version)
7
 
8
- This model card details the further pre-trained version of German FinBERT, a language model focusing on the financial domain within the German language.
 
9
 
10
  ## Overview
11
- **Author:
12
- ** Moritz Scherrmann
13
- **Framework:
14
- ** BERT-base
15
- **Language:
16
- ** German
17
- **Specialization:
18
- ** Financial textual data
19
- **Original Model:
20
- ** gbert of deepset
21
-
22
- ## Pre-training Corpus
23
- The pre-training corpus consists of German financial textual data. It comprises a comprehensive collection that includes financial reports, ad-hoc announcements, and news related to German companies. The corpus size is on par with those used for standard BERT models, indicating substantial coverage and depth.
24
  ## Performance
25
  ### Fine-tune Datasets
26
- German FinBERT has been evaluated on three finance-specific tasks against generic German language models, showing improved performance in:
27
- - Sentiment prediction
28
- - Topic recognition
29
- - Question answering
30
- The model effectively captures domain-specific nuances, outperforming standard models on finance-related texts.
31
 
32
  ### Benchmark Results
33
- - *The precise benchmark results and comparisons are not provided in the accessed part of the document.*
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  ## Authors
36
- **Moritz Scherrmann
37
- ** Institute for Finance & Banking
38
- Ludwig-Maximilians-Universität München
39
- Ludwigstr. 28, RB 80539 Munich, Germany
40
- Email: scherrmann@lmu.de
41
 
42
- For additional details regarding the performance on fine-tune datasets and benchmark results, please refer to the full documentation provided in the study. German FinBERT represents an innovative development in the field of financial NLP, offering enhanced capabilities for analyzing German financial texts.
 
 
 
 
 
5
  ---
6
  # German FinBERT (Further Pre-trained Version)
7
 
8
+ German FinBERT is a BERT language model focusing on the financial domain within the German language. In my [paper](https://arxiv.org/pdf/2010.10906.pdf) (UPDATE!!), I describe in more detail the steps taken to train the model and show that it outperforms its generic benchmarks for finance specific downstream tasks.
9
+ This version of German FinBERT starts with the [gbert-base](https://huggingface.co/deepset/gbert-base) model and continues pre-training on finance specific textual data.
10
 
11
  ## Overview
12
+ **Author:** [here](https://arxiv.org/pdf/2010.10906.pdf) (UPDATE!)
13
+ **Archticture:** BERT base
14
+ **Language:** German
15
+ **Specialization:** Financial textual data
16
+ **Original Model:** [gbert-base (deepset)](https://huggingface.co/deepset/gbert-base)
17
+ **Framework:** [MosaicML](https://github.com/mosaicml/examples/tree/main/examples/benchmarks/bert)
18
+
19
+ ## Pre-training
20
+ German FinBERT's pre-training corpus includes a diverse range of financial documents, such as Bundesanzeiger reports, Handelsblatt articles, MarketScreener data, and additional sources including FAZ, ad-hoc announcements, LexisNexis & Event Registry content, Zeit Online articles, Wikipedia entries, and Gabler Wirtschaftslexikon. In total, the corpus spans from 1996 to 2023, consisting of 12.15 million documents with 10.12 billion tokens over 53.19 GB.
21
+
22
+ I further pre-train the model for 10,400 steps with a batch size of 4096, which is one epoch. I use an Adam optimizer with decoupled weight decay regularization, with Adam parameters 0.9, 0.98, 1e − 6,a weight
23
+ decay of 1e − 5 and a maximal learning of 1e − 4. . I train the model using a Nvidia DGX A100 node consisting of 8 A100 GPUs with 80 GB of memory each.
24
+
25
  ## Performance
26
  ### Fine-tune Datasets
27
+ To fine-tune the model, I use several datasets, including:
28
+ - A manually labeled [multi-label database of German ad-hoc announcements](https://arxiv.org/pdf/2010.10906.pdf) (UPDATE!!!) containing 31,771 sentences, each associated with up to 20 possible topics.
29
+ - An extractive question-answering dataset based on the SQuAD format, which was created using 3,044 ad-hoc announcements processed by OpenAI's ChatGPT to generate and answer questions.
30
+ - The [financial phrase bank](https://arxiv.org/abs/1307.5336) of Malo et al. (2013) for sentiment classification, translated to German using [DeepL](https://www.deepl.com/translator)
 
31
 
32
  ### Benchmark Results
33
+ The further pre-trained German FinBERT model demonstrated the following performances on finance-specific downstream tasks:
34
+
35
+ Ad-Hoc Multi-Label Database:
36
+ - Macro F1: 86.08%
37
+ - Micro F1: 85.65%
38
+
39
+ Ad-Hoc QuAD (Question Answering):
40
+ - Exact Match (EM): 52.50%
41
+ - F1 Score: 74.61%
42
+
43
+ Translated Financial Phrase Bank:
44
+ - Accuracy: 95.41%
45
+ - Macro F1: 91.49%
46
 
47
  ## Authors
48
+ Moritz Scherrmann: `scherrmann [at] lmu.de`
49
+
50
+
51
+ For additional details regarding the performance on fine-tune datasets and benchmark results, please refer to the full documentation provided in the study.
 
52
 
53
+ See also:
54
+ scherrmann/GermanFinBERT_SC
55
+ scherrmann/GermanFinBERT_FP_Topic
56
+ scherrmann/GermanFinBERT_FP_QuAD
57
+ scherrmann/GermanFinBERT_SC_Sentiment