| CLIMATE BERT: A Pretrained Language Model for Climate-Related Text |
| Nicolas Webersinke,1Mathias Kraus,1Julia Anna Bingler,2Markus Leippold3 |
| 1FAU Erlangen-Nuremberg, Germany |
| 2ETH Zurich, Switzerland |
| 3University of Zurich, Switzerland |
| nicolas.webersinke@fau.de, mathias.kraus@fau.de, binglerj@ethz.ch, markus.leippold@bf.uzh.ch |
| Abstract |
| Over the recent years, large pretrained language models (LM) |
| have revolutionized the field of natural language processing |
| (NLP). However, while pretraining on general language has |
| been shown to work very well for common language, it has |
| been observed that niche language poses problems. In par- |
| ticular, climate-related texts include specific language that |
| common LMs can not represent accurately. We argue that |
| this shortcoming of today’s LMs limits the applicability of |
| modern NLP to the broad field of text processing of climate- |
| related texts. As a remedy, we propose C LIMATE BERT, a |
| transformer-based language model that is further pretrained |
| on over 2 million paragraphs of climate-related texts, crawled |
| from various sources such as common news, research arti- |
| cles, and climate reporting of companies. We find that C LI- |
| MATE BERT leads to a 48% improvement on a masked lan- |
| guage model objective which, in turn, leads to lowering error |
| rates by 3.57% to 35.71% for various climate-related down- |
| stream tasks like text classification, sentiment analysis, and |
| fact-checking. |
| 1 Introduction |
| Researchers working on climate change-related topics in- |
| creasingly use natural language processing (NLP) to auto- |
| matically extract relevant information from textual data. Ex- |
| amples include the sentiment or specificity of language used |
| by companies when discussing climate risks and measuring |
| corporate climate change exposure, which increases trans- |
| parency to help the public know where we stand on climate |
| change (e.g., Callaghan et al. 2021; Bingler et al. 2022b). |
| Many studies in this domain apply traditional NLP meth- |
| ods, such as dictionaries, bag-of-words approaches or sim- |
| ple extensions thereof (e.g., Gr ¨uning 2011; Sautner et al. |
| 2022). However, such analyses face considerable limita- |
| tions, since climate-related wording could vary substan- |
| tially by source (Kim and Kang 2018). Deep learning tech- |
| niques that promise higher accuracy are gradually replacing |
| these approaches (e.g., K ¨olbel et al. 2020; Luccioni, Baylor, |
| and Duchene 2020; Bingler et al. 2022a; Callaghan et al. |
| 2021; Wang, Chillrud, and McKeown 2021; Friederich et al. |
| 2021). Indeed, it has been shown in related domains that |
| Copyright c
2022, Association for the Advancement of Artificial |
| Intelligence (www.aaai.org). All rights reserved.deep learning in NLP allows for impressive results, outper- |
| forming traditional methods by large margins (Varini et al. |
| 2020). |
| These deep learning-based approaches make use of lan- |
| guage models (LMs), which are trained on large amounts |
| of textual and unlabelled data. This training on unlabelled |
| data is called pretraining and leads to the model learning |
| representations of words and patterns of common language. |
| One of the most prominent language models is called BERT |
| (Bidirectional Encoder Representations from Transformers) |
| (Devlin et al. 2018) with its successors R OBERT A(Liu et al. |
| 2019), Transformer-XL (Dai et al. 2019) and ELECTRA |
| (Clark et al. 2020). These models have been trained on huge |
| amounts of text which was crawled from an unprecedented |
| amount of online resources. |
| After the pretraining phase, most LMs are trained on addi- |
| tional tasks, the downstream task . For the downstream tasks, |
| the LM builds on and benefits from the word representations |
| and language patterns learned in the pretraining phase. The |
| pre-training benefit is especially large on downstream tasks |
| for which the collection of samples is difficult and, thus, the |
| resulting training datasets are small (hundreds or few thou- |
| sands of samples). Furthermore, it has been shown that a |
| model that was pretrained on the downstream task-specific |
| text exhibits better performance, compared to a model that |
| has been pretrained solely on general text (Araci 2019; Lee |
| et al. 2020). |
| Hence, a straightforward extension to the standard com- |
| bination of pretraining is the so-called domain-adaptive pre- |
| training (Gururangan et al. 2020). This approach has re- |
| cently been studied for various tasks and basically comes in |
| the form of pretraining multiple times — in particular pre- |
| training in the language domain of the downstream task, i.e., |
| pretraining (general domain) |
| +domain-adaptive |
| pretraining (downstream domain) |
| +training (downstream task) : |
| To date, regardless of the increase in using NLP for cli- |
| mate change related research, a model with climate domain- |
| adaptive pretraining has not been publicly available, yet. |
| Research so far rather relied on models pretrained on gen- |
| eral language, and fine-tuned on the downstream task. To |
|
|
| fill this gap, our contribution is threefold. First, we in- |
| troduce C LIMATE BERT, a state-of-the-art language model |
| that is specifically pretrained on climate-related text cor- |
| pora of various sources, namely news, corporate disclosures, |
| and scientific articles. This language model is designed to |
| support researchers of various disciplines in obtaining bet- |
| ter performing NLP models for a manifold of downstream |
| tasks in the climate change domain. Second, to illustrate |
| the strength of C LIMATE BERT, we highlight the perfor- |
| mance improvements using C LIMATE BERT on three stan- |
| dard climate-related NLP downstream tasks. Third, to fur- |
| ther promote research at the intersection of climate change |
| and NLP, we make the training code and weights of all lan- |
| guage models publicly available at GitHub and Hugging |
| Face.12 |
| 2 Background |
| As illustrated in Figure 1, our LM training approach for C LI- |
| MATE BERTcomprises all three phases — using an LM pre- |
| trained on a general domain, the domain-adaptive pretrain- |
| ing on the climate domain, and the training phase on climate- |
| related downstream tasks. |
| Pretraining on General Domain |
| As of 2018, pretraining became the quasi-standard for learn- |
| ing NLP models. First, a neural language model, often with |
| millions of parameters, is trained on large unlabeled corpora |
| in a semi-supervised fashion. By learning on multiple levels |
| which words/word-sequences/sentences appear in the same |
| context, an LM can represent a semantically similar text by |
| similar vectors. Typical objectives for training LMs are the |
| prediction of masked words or the prediction of a label indi- |
| cating whether two sentences occurred consecutively in the |
| corpora (Devlin et al. 2018). |
| In the earlier NLP pretraining days, LMs tradition- |
| ally used regular or convolutional neural networks (Col- |
| lobert and Weston 2008), or later Long-Short-Term-Memory |
| (LSTM) networks to process text (Howard and Ruder 2018). |
| Todays LMs mostly build on transformer models (e.g., De- |
| vlin et al. 2018; Dai et al. 2019; Liu et al. 2019). One of |
| the latter is named R OBERT A(Liu et al. 2019) which was |
| trained on 160GB of various English-language corpora - |
| data from BOOKCORPUS (Zhu et al. 2015), WIKIPEDIA, |
| a portion of the CCNEWS dataset (Nagel 2016), OPEN- |
| WEBTEXT corpus of web content extracted from URLs |
| shared on Reddit (Gokaslan and Cohen 2019), and a sub- |
| set of CommonCrawl that is said to resemble the story-like |
| style of WINOGRAD schemas (Trinh and Le 2019). While |
| these sources are valuable to build a model working on gen- |
| eral language, it has been shown that domain-specific, niche |
| language (such as climate-related text) poses a problem to |
| current state-of-the-art language models (Araci 2019). |
| Domain-Specific Pretraining |
| As a remedy to inferior performance of general language |
| models when applied to niche topics, multiple language |
| 1www.github.com/climatebert/language-model |
| 2www.huggingface.co/climatebertmodels have been proposed which build on the pretrained |
| models but continue pretraining on their respective domains. |
| FinBERT, LegalBert, MedBert are just a few language mod- |
| els that have been further pretrained on the finance, legal, or |
| medical domain (Araci 2019; Chalkidis et al. 2020; Rasmy |
| et al. 2021). In general, this domain-adaptive pretraining |
| yields more accurate models on downstream tasks (Guru- |
| rangan et al. 2020). |
| Domain-specific pretraining requires a decision about |
| which samples to include in the training process. It is still an |
| open debate which sample strategy improves performance |
| best. Various strategies can be applied to extract the text |
| samples on which the LM is further pretrained. For exam- |
| ple, while traditional pretraining uses all samples from the |
| pretraining corpus, similar sample selection (S IM-SELECT ) |
| uses only a subset of the corpus, in which the samples are |
| similar to the samples in the downstream task (Ruder and |
| Plank 2017). In contrast, diverse sample selection (D IV- |
| SELECT ) uses a subset of the corpus, which includes dissim- |
| ilar samples compared to the downstream dataset (Ruder and |
| Plank 2017). Previous research has investigated the benefit |
| of these approaches, yet no final conclusion about the effi- |
| ciency has been obtained. Consequently, we compare these |
| approaches in our experiments. |
| NLP on Climate-Related Text |
| In the past, climate-related textual analysis often used pre- |
| defined dictionaries of presumably relevant words and then |
| simply searched for these words within the documents. |
| For example, Cody et al. (2015) use such an approach |
| for climate-related tweets. Similarly, Sautner et al. (2022) |
| use a keyword-based approach to capture firm-level climate |
| change exposure. However, these methods do not account |
| for context. The lack of context is a significant drawback, |
| given the ambiguity of many climate-related words such |
| as ”environment,” ”sustainable,” or ”climate” itself (Varini |
| et al. 2020). |
| Only recently, BERT has been used for NLP in climate- |
| related text. The transformers-based BERT models are ca- |
| pable of accounting for the context of words and have out- |
| performed traditional approaches by large margins across |
| various climate-related datasets (K ¨olbel et al. 2020; Luc- |
| cioni, Baylor, and Duchene 2020; Varini et al. 2020; Bin- |
| gler et al. 2022a; Callaghan et al. 2021; Wang, Chillrud, and |
| McKeown 2021; Friederich et al. 2021; Stammbach et al. |
| 2022). However, this research has also shown that extracting |
| climate-related information from textual sources is a chal- |
| lenge, as climate change is a complex, fast-moving, and of- |
| ten ambiguous topic with scarce resources for popular text- |
| based AI tasks. |
| While context-based algorithms like BERT can detect |
| a variety of complex and implicit topic patterns in addi- |
| tion to many trivial cases, there remains great potential |
| for improvement in several directions. To our knowledge, |
| none of the above cited work has examined the effects of |
| domain-adaptive pretraining on their specific downstream |
| tasks. Therefore, we investigate whether domain-adaptive |
| pretraining will improve performance for climate change- |
| related downstream tasks such as text classification, senti- |
|
|
| News Abstracts ReportsCommon |
| crawlPretraining (general domain)Domain-adaptive pretraining (climate |
| domain)Training (downstream tasks) |
| + +- Text classification |
| - Sentiment analysis |
| - Fact-checkingFigure 1: Sequence of training phases. Our main contribution is the continued pretraining of language models on the climate |
| domain. In addition, we evaluate the obtained climate domain-specific language models on various downstream tasks. |
| ment analysis, and fact-checking. |
| 3 C LIMATE BERT |
| In the following, we describe our approach to train C LI- |
| MATE BERT. We first list the underlying data sources before |
| describing our sample selection techniques and, finally, the |
| vocabulary augmentation we used for training the language |
| model. |
| Text Corpus |
| Our goal was to collect a large corpus of text, C ORP, that |
| included general and domain-specific climate-related lan- |
| guage. We decided to include the following three sources: |
| news articles, research abstracts, and corporate climate re- |
| ports. We decided not to include full research articles be- |
| cause this language is likely too specific and does not rep- |
| resent general climate language. We also did not include |
| Twitter data, as we assume that these texts are too noisy. In |
| total, we collected 2,046,523 paragraphs of climate-related |
| text (see Table 1). |
| The N EWS dataset is mainly retrieved from Refinitiv |
| Workspace and includes 135,391 articles tagged with cli- |
| mate change topics such as climate politics, climate actions, |
| and floods and droughts. In addition, we crawled climate- |
| related news articles from the web. |
| The A BSTRACTS dataset includes abstracts of climate- |
| related research articles crawled from the Web of Science, |
| primarily published between 2000 and 2019. |
| The R EPORTS dataset comprises corporate climate and |
| sustainability reports of more than 600 companies from the |
| years 2015-2020 retrieved from Refinitiv Workspace and the |
| respective company websites. |
| Given the nature of the datasets, we find a large het- |
| erogeneity between the paragraphs in terms of number |
| of words. Unsurprisingly, on average, the paragraphs with |
| the least words come from the N EWS and the R EPORTS |
| datasets. In contrast, A BSTRACTS includes paragraphs with |
| the most words. Table 1 lists these descriptives. |
| To estimate the benefit from domain-adaptive pretrain- |
| ing, we compare the similarity of our text corpus with the |
| one used for pretraining R OBERT A. Following Gururangan |
| et al. (2020), we consider the vocabulary overlap between |
| both corpora. The resulting overlap of 57.05% highlights the |
| dissimilarity between the two domains and the need to add |
| specific vocabularies. Therefore, we expect to see consid- |
| erable performance improvements of domain-adaptive pre- |
| training.Dataset Num. of Avg. num. of words |
| paragraphs Q1 Mean Q3 |
| News 1,025,412 34 56 65 |
| Abstracts 530,819 165 218 260 |
| Reports 490,292 34 65 79 |
| Total 2,046,523 36 107 168 |
| Table 1: Corpus C ORP used for pretraining C LIMATE BERT. |
| Q1 and Q3 stand for the 0.25 and 0.75 quantiles, respec- |
| tively. |
| Sample Selection |
| Prior work has shown that specific selections of the samples |
| used for pretraining can foster the performance of the LM. In |
| particular, incorporating information from the downstream |
| task by selecting similar or diverse samples has been shown |
| to yield favorable results compared to using all samples from |
| the dataset. We follow both approaches and select samples |
| that are similar or diverse to climate-text using our text clas- |
| sification task (see 5). We experiment with three different |
| strategies from Ruder and Plank (2017) for the selection of |
| samples from our corpus: |
| In the most traditional sample selection strategy, F ULL- |
| SELECT , we use all paragraphs from C ORP to train |
| CLIMATE BERTF. |
| In S IM-SELECT , we select the 70% of samples from |
| CORP, which are most similar to the samples of our text |
| classification task. We use a Euclidean similarity met- |
| ric for this sample selection strategy. We call this LM |
| CLIMATE BERTS. |
| In D IV-SELECT , we select the 70% of samples from |
| CORP, which are most diverse compared to the samples |
| from our text classification task. We use the sum be- |
| tween the type-token-ratio and the Shannon-entropy for |
| measuring diversity (Ruder and Plank 2017). This LM is |
| named C LIMATE BERTD. |
| In D IV-SELECT + SIM-SELECT , we use the same diver- |
| sity and similarity metrics as before. We then compute |
| a composite score by summing over their scaled values. |
| We keep the 70% of the samples with the highest com- |
| posite score to train C LIMATE BERTD+S. |
|
|
| Downstream domain- Downstream tasks |
| adaptive pretraining training |
| Hyperparameter Value |
| Batch size 2016 32 |
| Learning rate 5e-4 5e-5 |
| Number of epochs 12 1000 |
| Patience — 4 |
| Class weight — Balanced |
| Feedforward nonlinearity — tanh |
| Feedforward layer — 1 |
| Output neurons — Task dependent |
| Optimizer Adam |
| Adam epsilon 1e-6 |
| Adam beta weights (0.9, 0.999) |
| Learning rate scheduler Warmup linear |
| Weight decay 0.01 |
| Table 2: Hyperparameters used for the downstream domain- |
| adaptive pretraining and the downstream tasks training of |
| CLIMATE BERT. |
| Vocabulary Augmentation |
| We extend the existing vocabulary of the original model |
| to include domain-specific terminology. This allows C LI- |
| MATE BERT to explicitly learn representations of terminol- |
| ogy that frequently occur in a climate-related text but not in |
| the general domain. In particular, we add the 235 most com- |
| mon tokens as new tokens to the tokenizer, thereby extend- |
| ing the size of the vocabulary for our basis language model |
| (DistilR OBERT A) from 50,265 to 50,500. See Appendix C |
| for a list of all added tokens. We also experimented with |
| language models that do not use vocabulary augmentation |
| or add more tokens. However, overall we find improvements |
| using this technique and, thus, apply it to all language mod- |
| els which we pretrain on the climate domain. |
| Model Selection |
| For all our experiments, we use DistilR OBERT A, a distilled |
| version of R OBERT Afrom Huggingface,3as our starting |
| point for training (Sanh et al. 2019). All our language mod- |
| els are trained with a masked language modeling objective |
| (i.e., cross-entropy loss on predicting randomly masked to- |
| kens). We report all hyperparameters in Table 2. The large |
| batch size of 2016 for training the LM is achieved using gra- |
| dient accumulation. |
| Training on Downstream Task |
| After pretraining DistilR OBERT Aon C ORP, we follow |
| standard practice (Devlin et al. 2018) and pass the final layer |
| [CLS] token representation to a task-specific feedforward |
| layer for prediction. We report all hyperparameters of this |
| feedforward layer in Table 2. |
| 4 Performance Analysis of Language Model |
| Table 3 lists the results after pretraining DistilR OBERT Aon |
| CORP with various sample selection strategies. For evalu- |
| ation, we split C ORP randomly into 80% training data and |
| 20% validation data. The reported loss is the cross-entropy |
| 3www.huggingface.co/distilroberta-baseloss on predicting randomly masked tokens from the valida- |
| tion data. We find that C LIMATE BERTFleads to the lowest |
| validation loss. This performance is followed by the other |
| CLIMATE BERT LMs, which all show similar results. Over- |
| all, we find that our domain-adaptive pretraining decreases |
| the cross-entropy loss by 46–48% compared to the basis Dis- |
| tilR OBERT A, cutting the loss almost in half. |
| Model Val. loss |
| DistilR OBERT A 2.238 |
| CLIMATE BERTF 1.157 |
| CLIMATE BERTS 1.205 |
| CLIMATE BERTD 1.204 |
| CLIMATE BERTD+S 1.203 |
| Table 3: Loss of our language models on a validation set |
| from our text corpus C ORP. |
| 5 Performance Analysis for Climate-Related |
| Downstream Tasks |
| For our experiments, we used the following downstream |
| tasks: text classification, sentiment analysis, and fact- |
| checking. Table 4 lists basic statistics about the downstream |
| tasks. We repeated the training and evaluation phase 60 |
| times for each experiment, each time with a random 90% |
| set of samples for training and the remaining 10% set for |
| validation. |
| Downstream Num. of Labels Label |
| task samples distribution |
| Text classification 1220 climate-related: yes/no 1000/220 |
| Sentiment analysis 1000 opportunity/neutral/risk 250/408/342 |
| Fact-checking 2745 claim: support/refute 1943/802 |
| Table 4: Overview of our downstream tasks used for evalu- |
| ating C LIMATE BERT. |
| Text Classification |
| For our text classification experiment, we use a dataset con- |
| sisting of hand-selected paragraphs from companies’ annual |
| reports or sustainability reports. All paragraphs were anno- |
| tated as yes(climate-related) or no(not climate-related) by at |
| least four experts from the field using the software prodigy.4 |
| See Appendix B for our annotation guidelines. In case of a |
| close verdict or a tie between the annotators, the authors of |
| this paper discussed the paragraph in depth before reaching |
| an agreement. |
| In the following, Table 5 reports the results of the lan- |
| guage models when trained on our text classification task, |
| i.e., whether the text is climate-related or not. Overall, |
| we find that all C LIMATE BERT LMs outperform a non- |
| pre-trained DistilR OBERT Aacross both metrics for the |
| text classification task. Most notably, domain-adaptive pre- |
| training with similar samples to our downstream tasks |
| 4www.prodi.gy |
|
|
| (CLIMATE BERTS) leads to improvements of 32.64% in |
| terms of cross-entropy loss and a reduction in the error rate |
| of the F1 score by 35.71%. |
| Text classification |
| Model Loss F1 |
| DistilR OBERT A 0:242 0:171 0:986 0:010 |
| CLIMATE BERTF 0:191 0:136 0:989 0:010 |
| CLIMATE BERTS 0:163 0:132 0:991 0:008 |
| CLIMATE BERTD 0:197 0:153 0:988 0:009 |
| CLIMATE BERTD+S0:217 0:153 0:988 0:009 |
| Table 5: Results on our text classification task. Reported are |
| the average cross-entropy loss and the average weighted F1 |
| score on the validation sets across 60 evaluation runs. Value |
| subscripts report the standard deviations. |
| Sentiment Analysis |
| Our next task studies the sentiment behind the climate- |
| related paragraphs, using the same dataset as in the previ- |
| ous section. In our context, we use the term ‘sentiment’ to |
| distinguish whether an entity reports on climate-related de- |
| velopments as negative risk, as positive opportunity , or as |
| neutral . |
| Therefore, we created a second labeled dataset on climate- |
| related sentiment, for which we asked the annotators to label |
| the paragraphs by one of the three categories — risk,neutral , |
| oropportunity . See Appendix B for our annotation guide- |
| lines. Similarly, as before, in case of a close verdict or a tie |
| between the annotators, the authors of this paper discussed |
| the paragraph in depth before reaching an agreement. |
| Table 6 shows the performance of our models in senti- |
| ment prediction. Again, all C LIMATE BERTLMs outperform |
| the DistilR OBERT Abaseline model in terms of F1 score and |
| average cross-entropy loss. The largest improvements can be |
| observed with C LIMATE BERTF, which amount to a 7.33% |
| lower cross-entropy loss and a 7.42% lower error rate in |
| terms of average F1 score compared to the DistilR OBERT A |
| baseline LM. |
| Sentiment analysis |
| Model Loss F1 |
| DistilR OBERT A 0:150 0:069 0:825 0:046 |
| CLIMATE BERTF 0:139 0:042 0:838 0:036 |
| CLIMATE BERTS 0:140 0:057 0:836 0:033 |
| CLIMATE BERTD 0:138 0:043 0:835 0:040 |
| CLIMATE BERTD+S0:139 0:043 0:834 0:036 |
| Table 6: Results on our sentiment analysis task in terms |
| of average validation loss and average weighted F1 score |
| across 60 evaluation runs. Subscripts report the standard de- |
| viations.Fact-Checking |
| We now turn to the fact-checking downstream task. We ap- |
| ply our model to a dataset that was proposed by Diggelmann |
| et al. (2020) and comprises 1.5k sentences that make a claim |
| about climate-related topics. This CLIMATE -FEVER dataset |
| is to the best of our knowledge to date the only dataset |
| that focuses on climate change fact-checking. CLIMATE - |
| FEVER adapts the methodology of FEVER , the largest dataset |
| of artificially designed claims, to real-life claims on cli- |
| mate change collected online. The authors of CLIMATE - |
| FEVER find that the surprising, subtle complexity of mod- |
| eling real-world climate-related claims provides a valuable |
| challenge for general natural language understanding. Work- |
| ing with this dataset, Wang, Chillrud, and McKeown (2021) |
| recently introduced a novel semi-supervised training method |
| to achieve a state-of-the-art (SotA) F1 score of 0.7182 on the |
| fact-checking dataset CLIMATE -FEVER . |
| Claim : 97% consensus on human-caused |
| global warming has been disproven. |
| Evidence |
| REFUTE: In a 2019 CBS poll, 64% of the US |
| population said that climate change |
| is a ””crisis”” or a ””serious prob- |
| lem””, with 44% saying human ac- |
| tivity was a significant contributor. |
| Claim : The melting Greenland ice sheet is |
| already a major contributor to ris- |
| ing sea level and if it was eventu- |
| ally lost entirely, the oceans would |
| rise by six metres around the world, |
| flooding many of the world’s largest |
| cities. |
| Evidence |
| SUPPORT: The Greenland ice sheet occupies |
| about 82% of the surface of Green- |
| land, and if melted would cause sea |
| levels to rise by 7.2 metres. |
| Table 7: Examples taken from CLIMATE -FEVER . |
| Each claim in CLIMATE -FEVER is supported or refuted by |
| evidence sentences (see Table 7), and an evidence sentence |
| can also be classified as giving not enough information. The |
| objective of the model is to classify an evidence sentence to |
| support orrefute a claim. To feed this combination of claim |
| and evidence into the model, we concatenate the claims with |
| the related evidence sentences, with a [SEP] token sepa- |
| rating them. As in Wang, Chillrud, and McKeown (2021), |
| and for comparison with their results, we filter out all evi- |
| dence sentences with the label NOT ENOUGH INFO in the |
| CLIMATE -FEVER dataset. |
| Table 8 lists the results of our experiments on the |
| CLIMATE -FEVER dataset. In line with our previous exper- |
| iments, we find similar or better results for all C LIMATE - |
| BERTLMs across all metrics. Our C LIMATE BERTD+SLM |
| achieves similar cross-entropy loss compared to the basis |
| DistilR OBERT Amodel, yet pushes the average F1 score |
| from 0.748 to 0.757, which outperforms Wang, Chillrud, and |
| McKeown (2021)’s previous SotA F1 score of 0.7182, and |
|
|
| is hence, to the best of our knowledge, the new SotA on this |
| dataset. |
| Fact-checking |
| Model Loss F1 |
| DistilR OBERT A 0:135 0:017 0:748 0:036 |
| CLIMATE BERTF 0:134 0:020 0:755 0:037 |
| CLIMATE BERTS 0:133 0:017 0:753 0:042 |
| CLIMATE BERTD 0:135 0:016 0:752 0:042 |
| CLIMATE BERTD+S0:135 0:018 0:757 0:044 |
| Table 8: Results on our fact-checking task on CLIMATE - |
| FEVER in terms of average validation loss and average |
| weighted F1 score across 60 evaluation runs. Subscripts re- |
| port the standard deviations. |
| 6 Carbon Footprint |
| Training deep neural networks in general and large lan- |
| guage models in particular, has a significant carbon footprint |
| already today. If the LM research trends continue, this detri- |
| mental climate impact will increase considerably. The topic |
| of efficient NLP was also discussed by a working group |
| appointed by the ACL Executive Committee to promote |
| ways that the ACL community can reduce the computational |
| costs of model training (https://public.ukp.informatik.tu- |
| darmstadt.de/enlp/Efficient-NLP-policy-document.pdf). |
| We acknowledge that our work is part of this trend. In |
| total, training C LIMATE BERTcaused 115.15 kg CO2 emis- |
| sions. We use two energy efficient NVIDIA RTX A5000 |
| GPUs: 0.7 kW (power consumption of GPU server) x 350 |
| hours (combined training time of all experiments) x 470 |
| gCO2e/kWh (emission factor in Germany in 2018 according |
| to www.umweltbundesamt.de/publikationen/entwicklung- |
| der-spezifischen-kohlendioxid-7) = 115,149 gCO2e. We |
| list all details about our climate impact in Table 9 in |
| Appendix A. Nevertheless, we decided to carry out this |
| project, as we see the high potential of NLP to support |
| action against climate change. Given our awareness of the |
| carbon footprint of our research, we address this sensitive |
| topic as follows: |
| 1. We specifically decided to focus on DistilR OBERT A, |
| which is a considerably smaller model in terms of num- |
| ber of parameters compared to the non-distilled version |
| and, thus, requires less energy to train. Moreover, we do |
| not crawl huge amounts of data without considering the |
| quality. This way, we try to take into account the issues |
| mentioned by Bender et al. (2021). |
| 2. Hyperparameter tuning yields considerably higher CO2 |
| emissions in the training stage due to tens or hundreds |
| of different training runs. Note that our multiple train- |
| ing runs on the downstream task are not causing long |
| training times as the downstream datasets are very small |
| compared to the dataset used for training the language |
| model. We therefore refrain from exhaustive hyperpa- |
| rameter tuning. Rather, we build on previous findings.We systematically experimented with a few hyperparam- |
| eter combinations and found that the hyperparameters |
| proposed by Gururangan et al. (2020) lead to the best |
| results. |
| 3. We would have liked to train and run our model on |
| servers powered by renewable energy. This first best op- |
| tion was unfortunately not available. In order to speed |
| up the energy system transformation required to achieve |
| the global climate targets, we contribute our part by do- |
| nating Euro 100 to atmosfair. atmosfair was founded in |
| 2005 and is supported by the German Federal Environ- |
| ment Agency. atmosfair offsets carbon dioxide in more |
| than 20 locations: from efficient cookstoves in Nigeria, |
| Ethiopia and India to biogas plants in Nepal and Thai- |
| land to solar energy in Senegal and Brazil and renewable |
| energies in Tansania and Indonesia. See www.atmosfair. |
| de/en/offset/fix/. We explicitly refrain from calling this |
| donation a CO2 compensation, and we refrain from a so- |
| lution that is based on afforestation. |
| 7 Conclusion |
| We propose C LIMATE BERT, the first language model that |
| was pretrained on a large scale dataset of over 2 mil- |
| lion climate-related paragraphs. We study various selec- |
| tion strategies to find samples from our corpus which are |
| most helpful for later tasks. Our experiments reveal that |
| our domain-adaptive pretraining leads to considerably lower |
| masked language modeling loss on our climate corpus. We |
| further find that this improvement is also reflected in predic- |
| tive performance across three essential downstream climate- |
| related NLP tasks: text classification, the analysis of risk and |
| opportunity statements by corporations, and fact-checking |
| climate-related claims. |
| Acknowledgments |
| We are very thankful to Jan Minx and Max Callaghan from |
| the Mercator Research Institute on Global Commons and |
| Climate Change (MCC) Berlin for providing us with the |
| data, which is a subset of the data they used in Berrang-Ford |
| et al. (2021) and Callaghan et al. (2021). |
| References |
| Araci, D. 2019. Finbert: Financial sentiment analy- |
| sis with pre-trained language models. arXiv preprint |
| arXiv:1908.10063 . |
| Bender, E. M.; Gebru, T.; McMillan-Major, A.; and |
| Shmitchell, S. 2021. On the Dangers of Stochastic Par- |
| rots: Can Language Models Be Too Big? In Proceedings |
| of the 2021 ACM Conference on Fairness, Accountability, |
| and Transparency , 610–623. |
| Berrang-Ford, L.; Sietsma, A. J.; Callaghan, M.; Minx, J. C.; |
| Scheelbeek, P. F.; Haddaway, N. R.; Haines, A.; and Dan- |
| gour, A. D. 2021. Systematic mapping of global research on |
| climate and health: a machine learning review. The Lancet |
| Planetary Health , 5(8): e514–e525. |
|
|
| Bingler, J. A.; Kraus, M.; Leippold, M.; and Webersinke, |
| N. 2022a. Cheap talk and cherry-picking: What C LIMATE - |
| BERT has to say on corporate climate risk disclosures. Fi- |
| nance Research Letters , 102776. |
| Bingler, J. A.; Kraus, M.; Leippold, M.; and Webersinke, N. |
| 2022b. Cheap talk in corporate climate commitments: The |
| role of active institutional ownership, signaling, materiality, |
| and sentiment. Swiss Finance Institute Research Paper . |
| Callaghan, M.; Schleussner, C.-F.; Nath, S.; Lejeune, Q.; |
| Knutson, T. R.; Reichstein, M.; Hansen, G.; Theokritoff, E.; |
| Andrijevic, M.; Brecha, R. J.; et al. 2021. Machine-learning- |
| based evidence and attribution mapping of 100,000 climate |
| impact studies. Nature climate change , 11(11): 966–972. |
| Chalkidis, I.; Fergadiotis, M.; Malakasiotis, P.; Aletras, N.; |
| and Androutsopoulos, I. 2020. LEGAL -BERT : The muppets |
| straight out of law school. arXiv preprint arXiv:2010.02559 . |
| Clark, K.; Luong, M.-T.; Le, Q. V .; and Manning, C. D. |
| 2020. ELECTRA: Pre-training Text Encoders as Discrimi- |
| nators Rather Than Generators. In International Conference |
| on Learning Representations . |
| Cody, E. M.; Reagan, A. J.; Mitchell, L.; Dodds, P. S.; and |
| Danforth, C. M. 2015. Climate Change Sentiment on Twit- |
| ter: An Unsolicited Public Opinion Poll. PLOS ONE , 10(8): |
| e0136092. |
| Collobert, R.; and Weston, J. 2008. A Unified Architecture |
| for Natural Language Processing: Deep Neural Networks |
| with Multitask Learning. In Proceedings of the 25th Inter- |
| national Conference on Machine Learning , 160–167. |
| Dai, Z.; Yang, Z.; Yang, Y .; Carbonell, J. G.; Le, Q.; and |
| Salakhutdinov, R. 2019. Transformer-XL: Attentive Lan- |
| guage Models beyond a Fixed-Length Context. In Proceed- |
| ings of the 57th Annual Meeting of the Association for Com- |
| putational Linguistics , 2978–2988. |
| Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2018. |
| Bert: Pre-training of deep bidirectional transformers for lan- |
| guage understanding. arXiv preprint arXiv:1810.04805 . |
| Diggelmann, T.; Boyd-Graber, J.; Bulian, J.; Ciaramita, M.; |
| and Leippold, M. 2020. CLIMATE-FEVER: A Dataset for |
| Verification of Real-World Climate Claims. arXiv preprint |
| arXiv:2012.00614 . |
| Friederich, D.; Kaack, L. H.; Luccioni, A.; and Steffen, |
| B. 2021. Automated Identification of Climate Risk Dis- |
| closures in Annual Corporate Reports. arXiv preprint |
| arXiv:2108.01415 . |
| Gokaslan, A.; and Cohen, V . 2019. OpenWebText Corpus. |
| Gr¨uning, M. 2011. Artificial intelligence measurement of |
| disclosure (AIMD). European Accounting Review , 20(3): |
| 485–519. |
| Gururangan, S.; Marasovi ´c, A.; Swayamdipta, S.; Lo, K.; |
| Beltagy, I.; Downey, D.; and Smith, N. A. 2020. Don’t Stop |
| Pretraining: Adapt Language Models to Domains and Tasks. |
| InProceedings of the 58th Annual Meeting of the Associa- |
| tion for Computational Linguistics , 8342–8360. |
| Hershcovich, D.; Webersinke, N.; Kraus, M.; Bingler, J. A.; |
| and Leippold, M. 2022. Towards Climate Awareness in NLP |
| Research. arXiv preprint arXiv:2205.05071 .Howard, J.; and Ruder, S. 2018. Universal Language Model |
| Fine-tuning for Text Classification. In Proceedings of the |
| 58th Annual Meeting of the Association for Computational |
| Linguistics , 328–339. |
| Kim, D.-Y .; and Kang, S.-W. 2018. Analysis of Recognition |
| of Climate Changes using Word2Vec. International Journal |
| of Pure and Applied Mathematics , 120(6): 5793–5807. |
| K¨olbel, J. F.; Leippold, M.; Rillaerts, J.; and Wang, Q. 2020. |
| Ask BERT: How regulatory disclosure of transition and |
| physical climate risks affects the CDS term structure. Avail- |
| able at SSRN 3616324 . |
| Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C. H.; and |
| Kang, J. 2020. BioBERT: a pre-trained biomedical language |
| representation model for biomedical text mining. Bioinfor- |
| matics , 36(4): 1234–1240. |
| Liu, Y .; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, |
| O.; Lewis, M.; Zettlemoyer, L.; and Stoyanov, V . 2019. |
| ROBERT A: A robustly optimized bert pretraining approach. |
| arXiv preprint arXiv:1907.11692 . |
| Luccioni, A.; Baylor, E.; and Duchene, N. 2020. Analyzing |
| Sustainability Reports Using Natural Language Processing. |
| arXiv preprint arXiv:2011.08073 . |
| Nagel, S. 2016. CC-NEWS. |
| Rasmy, L.; Xiang, Y .; Xie, Z.; Tao, C.; and Zhi, D. 2021. |
| Med-BERT: pretrained contextualized embeddings on large- |
| scale structured electronic health records for disease predic- |
| tion. NPJ digital medicine , 4(1): 1–13. |
| Ruder, S.; and Plank, B. 2017. Learning to select data for |
| transfer learning with Bayesian Optimization. In Proceed- |
| ings of the 2017 Conference on Empirical Methods in Natu- |
| ral Language Processing , 372–382. |
| Sanh, V .; Debut, L.; Chaumond, J.; and Wolf, T. 2019. Dis- |
| tilBERT, a distilled version of BERT: smaller, faster, cheaper |
| and lighter. arXiv preprint arXiv:1910.01108 . |
| Sautner, Z.; van Lent, L.; Vilkov, G.; and Zhang, R. 2022. |
| Firm-level Climate Change Exposure. Journal of Finance , |
| forthcoming. |
| Stammbach, D.; Webersinke, N.; Bingler, J. A.; Kraus, M.; |
| and Leippold, M. 2022. A Dataset for Detecting Real-World |
| Environmental Claims. arXiv preprint arXiv:2209.00507 . |
| Trinh, T. H.; and Le, Q. V . 2019. A Simple Method for Com- |
| monsense Reasoning. arXiv preprint arXiv:1806.02847 . |
| Varini, F. S.; Boyd-Graber, J.; Ciaramita, M.; and Leippold, |
| M. 2020. ClimaText: A dataset for climate change topic de- |
| tection. In Tackling Climate Change with Machine Learning |
| (Climate Change AI) workshop at NeurIPS . |
| Wang, G.; Chillrud, L.; and McKeown, K. 2021. Evidence |
| based Automatic Fact-Checking for Climate Change Misin- |
| formation. International Workshop on Social Sensing on The |
| International AAAI Conference on Web and Social Media . |
| Zhu, Y .; Kiros, R.; Zemel, R.; Salakhutdinov, R.; Urtasun, |
| R.; Torralba, A.; and Fidler, S. 2015. Aligning books and |
| movies: Towards story-like visual explanations by watching |
| movies and reading books. In Proceedings of the IEEE in- |
| ternational conference on computer vision , 19–27. |
|
|
| Appendix |
| A Climate Performance Model Card |
| Table 9 shows our climate performance model card, follow- |
| ing Hershcovich et al. (2022). |
| ClimateBert |
| 1. Model publicly available? Yes |
| 2. Time to train final model 48 hours |
| 3. Time for all experiments 350 hours |
| 4. Power of GPU and CPU 0.7 kW |
| 5. Location for computations Germany |
| 6. Energy mix at location 470 gCO2eq/kWh |
| 7. CO2eq for final model 15.79 kg |
| 8. CO2eq for all experiments 115.15 kg |
| 9. Average CO2eq for inference per sample 0.62 mg |
| Table 9: Climate performance model card for ClimateBert. |
| B Annotation Guidelines |
| For our annotation procedure, we implemented the fol- |
| lowing general rules. The annotators had to label climate- |
| relevant paragraphs. If the paragraph was climate-relevant, |
| then they had to attach a sentiment to a paragraph. Annota- |
| tors were asked to apply common sense, e.g., when a given |
| paragraph might not provide all the context, but the context |
| might seem obvious. Moreover, annotators were informed |
| that each annotation should be a 0-1 decision. Hence, if |
| an annotator was 70% certain, then this was rounded up to |
| 100%. We asked, on average, five researchers to annotate the |
| same tasks to obtain some measure of dispersion. In case of |
| a close verdict or a tie between the annotators, the authors of |
| this paper discussed the paragraph in depth before reaching |
| an agreement. |
| Text classification |
| The first task was to label climate-relevant paragraphs. The |
| labels are YesorNo. As a general rule, we determined that |
| just discussing nature/environment can be sufficient, and |
| mentioning clean energy, emissions, fossil fuels, etc., can |
| also be sufficient. It is a Yes, if the paragraph includes some |
| wording on a climate change or environment related topic |
| (including transition and litigation risks, i.e., emission mit- |
| igation measures, energy consumption and energy sources |
| etc.; and physical risks, i.e., increase in risk of floods, coastal |
| area exposure, storms etc.). It is a No, if the paragraph is not |
| related to climate policy, climate change or an environmen- |
| tal topic at all. For some examples, see Table 10. |
| Sentiment Analysis |
| For the sentiment analysis, annotators had to provide la- |
| bels as to whether a (climate change-related) paragraph talks |
| about a Risk or threat that negatively impacts an entity of in- |
| terest, i.e. a company (negative sentiment), or whether an en- |
| tity is referring to some Opportunity arising due to climate |
| change (positive sentiment). The paragraph can also make |
| just a Neutral statement.Label Examples |
| Yes Sustainability: The Group is subject |
| to stringent and evolving laws, reg- |
| ulations, standards and best prac- |
| tices in the area of sustainabil- |
| ity (comprising corporate gover- |
| nance, environmental management |
| and climate change (specifically |
| capping of emissions), health and |
| safety management and social per- |
| formance) which may give rise |
| to increased ongoing remediation |
| and/or other compliance costs and |
| may adversely affect the Group’s |
| business, results of operations, fi- |
| nancial condition and/or prospects. |
| Yes Scope 3: Optional scope that in- |
| cludes indirect emissions associ- |
| ated with the goods and services |
| supply chain produced outside the |
| organization. Included are emis- |
| sions from the transport of products |
| from our logistics centres to stores |
| (downstream) performed by exter- |
| nal logistics operators (air, land |
| and sea transport) as well as the |
| emissions associated with electric- |
| ity consumption in franchise stores. |
| No Risk and risk management Opera- |
| tional risk and compliance risk Op- |
| erational risk is the risk of loss re- |
| sulting from inadequate or failed |
| internal processes, people and sys- |
| tems, or from external events in- |
| cluding legal risk but excluding |
| strategic and reputation risk. It also |
| includes, among other things, tech- |
| nology risk, model risk and out- |
| sourcing risk. |
| Table 10: Examples for the annotation task climate |
| (Yes/No). |
| To be more precise, we consider a paragraph relating to |
| risk, if the paragraph mainly talks about 1) business down- |
| side risks, potential losses and adverse developments detri- |
| mental to the entity 2) and/or about negative impact of an |
| entity’s activities on the society/environment 3) and/or asso- |
| ciates specific negative adjectives to the anticipated, past or |
| present developments and topics covered. |
| We consider a paragraph relating to opportunities, if the |
| paragraph mainly talks about 1) business opportunities aris- |
| ing from mitigating climate change, from adapting to cli- |
| mate change etc. which might be beneficial for a specific |
| entity 2) and/or about positive impact of an entity’s activi- |
| ties on the society/environment 3) and/or associates specific |
|
|
| positive adjectives to the anticipated, past or present devel- |
| opments and topics covered. |
| Lastly, we consider a paragraph as neutral if it mainly |
| states facts and developments 1) without putting them into |
| positive or negative perspective for a specific entity and/or |
| the society and/or the environment, 2) and/or does not as- |
| sociate specific positive or negative adjectives to the antic- |
| ipated, past or present facts stated and topics covered. For |
| some examples, see Table 11. |
| C Added Tokens |
| ’CO2’, ’emissions’, ”’, ’temperature’, ’environmental’, |
| ’soil’, ’increase’, ’conditions’, ’potential’, ’increased’, ’ar- |
| eas’, ’degrees’, ’across’, ’systems’, ’emission’, ’precipi- |
| tation’, ’impacts’, ’compared’, ’countries’, ’sustainable’, |
| ’provide’, ’reduction’, ’annual’, ’reduce’, ’greenhouse’, |
| ’approach’, ’processes’, ’factors’, ’observed’, ’renewable’, |
| ’temperatures’, ’distribution’, ’studies’, ’variability’, ’sig- |
| nificantly’, ’–’, ’further’, ’regions’, ’addition’, ’showed’, |
| ’“’, ’industry’, ’consumption’, ’regional’, ’risks’, ’atmo- |
| spheric’, ’supply’, ’companies’, ’plants’, ’biomass’, ’elec- |
| tricity’, ’respectively’, ’activities’, ’communities’, ’cli- |
| matic’, ’solar’, ’investment’, ’spatial’, ’rainfall’, ’ ’, ’sus- |
| tainability’, ’costs’, ’reduced’, ’2021’, ’influence’, ’vegeta- |
| tion’, ’sources’, ’possible’, ’ecosystem’, ’scenarios’, ’sum- |
| mer’, ’drought’, ’structure’, ’economy’, ’considered’, ’var- |
| ious’, ’atmosphere’, ’several’, ’technologies’, ’transition’, |
| ’assessment’, ’dioxide’, ’ocean’, ’fossil’, ’patterns’, ’waste’, |
| ’solutions’, ’transport’, ’strategy’, ’CH4’, ’policies’, ’un- |
| derstanding’, ’concentration’, ’customers’, ’methane’, ’ap- |
| plied’, ’increases’, ’estimated’, ’flood’, ’measured’, ’ther- |
| mal’, ’concentrations’, ’decrease’, ’greater’, ’following’, |
| ’proposed’, ’trends’, ’basis’, ’provides’, ’operations’, ’dif- |
| ferences’, ’hydrogen’, ’adaptation’, ’methods’, ’capture’, |
| ’variation’, ’reducing’, ’N2O’, ’parameters’, ’ecosystems’, |
| ’investigated’, ’yield’, ’strategies’, ’indicate’, ’caused’, ’dy- |
| namics’, ’obtained’, ’efforts’, ’coastal’, ’become’, ’agri- |
| cultural’, ’decreased’, ’GHG’, ’materials’, ’mainly’, ’rela- |
| tionship’, ’ecological’, ’benefits’, ’+/-’, ’challenges’, ’nitro- |
| gen’, ’forests’, ’trend’, ’estimates’, ’towards’, ’Committee’, |
| ’seasonal’, ’developing’, ’particular’, ’importance’, ’tropi- |
| cal’, ’ratio’, ’2030’, ’composition’, ’employees’, ’charac- |
| teristics’, ’scenario’, ’measurements’, ’plans’, ’fuels’, ’in- |
| frastructure’, ’overall’, ’responses’, ’presented’, ’least’, ’as- |
| sess’, ’diversity’, ’periods’, ’delta’, ’included’, ’already’, |
| ’targets’, ’achieve’, ’affect’, ’conducted’, ’operating’, ’pop- |
| ulations’, ’variations’, ’studied’, ’additional’, ’construction’, |
| ’northern’, ’variables’, ’soils’, ’ensure’, ’recovery’, ’com- |
| bined’, ’decision’, ’practices’, ’however’, ’determined’, ’re- |
| sulting’, ’mitigation’, ’conservation’, ’estimate’, ’identify’, |
| ’observations’, ’losses’, ’productivity’, ’agreement’, ’mon- |
| itoring’, ’investments’, ’pollution’, ’contribution’, ’oppor- |
| tunities’, ’simulations’, ’gases’, ’statements’, ’planning’, |
| ’shares’, ’sediment’, ’flux’, ’requirements’, ’trees’, ’tempo- |
| ral’, ’determine’, ’southern’, ’previous’, ’integrated’, ’rel- |
| atively’, ’analyses’, ’means’, ’2050’, ’”’, ’uncertainty’, |
| ’pandemic’, ’fluxes’, ’findings’, ’moisture’, ’consistent’, |
| ’decades’, ’snow’, ’performed’, ’contribute’, ’crisis’Label Examples |
| Opportunity Grid & Infrastructure and Retail – today represent |
| the energy world of tomorrow. We rank among Eu- |
| rope‘s market leaders in the grid and retail busi- |
| ness and have leading positions in renewables. We |
| intend to spend a total of between Euro 6.5 bil- |
| lion and Euro 7.0 billion in capital throughout the |
| Group from 2017 to 2019. |
| Opportunity We want to contribute to the transition to a circu- |
| lar economy. The linear economy is not sustain- |
| able. We discard a great deal (waste and there- |
| fore raw materials, experience, social capital and |
| knowledge) and are squandering value as a result. |
| This is not tenable from an economic and ecolog- |
| ical perspective. As investor we can ‘direct’ com- |
| panies and with our network, our scale and our in- |
| fluence we can help the movement towards a cir- |
| cular future (creating a sustainable society) further |
| along. |
| Neutral A similar approach could be used for allocating |
| emissions in the fossil fuel electricity supply chain |
| between coal miners, transporters and generators. |
| We don’t invest in fossil fuel companies, but those |
| investors who do should account properly for their |
| role in the production of dangerous emissions from |
| burning fossil fuels. |
| Neutral Omissions: Emissions associated with joint ven- |
| tures and investments are not included in the emis- |
| sions disclosure as they fall outside the scope of our |
| operational boundary. We do not have any emis- |
| sions associated with heat, steam or cooling. We |
| are not aware of any other material sources of omis- |
| sions from our emissions reporting. |
| Risk We estimated that between 36.5 and 52.9 per cent |
| of loans granted to our clients are exposed to tran- |
| sition risks. If the regulator decides to pass am- |
| bitious laws to accelerate the transition towards a |
| low-carbon economy, carbon-intensive companies |
| would incur in higher costs, which may prevent |
| them from repaying their debt. In turn, this would |
| weaken our bank’s balance sheets. . |
| Risk American National Insurance Company recognizes |
| that increased claims activity resulting from catas- |
| trophic events, whether natural or man-made, may |
| result in significant losses, and that climate change |
| may also affect the affordability and availability of |
| property and casualty insurance and the pricing for |
| such products. |
| Table 11: Examples for the annotation task sentiment (Op- |
| portunity/Neutral/Risk). |
|
|
|
|