File size: 1,228 Bytes
a422b7b
d74b14e
a422b7b
d74b14e
 
 
a422b7b
d74b14e
98b202f
d74b14e
 
 
3da9585
d74b14e
98b202f
d74b14e
 
 
3da9585
 
 
 
 
 
 
 
 
 
d74b14e
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
language: en
license: apache-2.0
tags:
- ESG
- environmental
---

# Model Card for EnvironmentalBERT-base

## Model Description

Based on [this paper](https://www.sciencedirect.com/science/article/pii/S1544612324000096), this is the EnvironmentalBERT-base language model. A language model that is trained to better understand environmental texts in the ESG domain.

Using the [DistilRoBERTa](https://huggingface.co/distilroberta-base) model as a starting point, the EnvironmentalBERT-base Language Model is additionally pre-trained on a text corpus comprising environmental-related annual reports, sustainability reports, and corporate and general news.

## More details can be found in the paper
```bibtex
@article{schimanski_ESGBERT_2024,
title = {Bridging the gap in ESG measurement: Using NLP to quantify environmental, social, and governance communication},
journal = {Finance Research Letters},
volume = {61},
pages = {104979},
year = {2024},
issn = {1544-6123},
doi = {https://doi.org/10.1016/j.frl.2024.104979},
url = {https://www.sciencedirect.com/science/article/pii/S1544612324000096},
author = {Tobias Schimanski and Andrin Reding and Nico Reding and Julia Bingler and Mathias Kraus and Markus Leippold},
}
```