File size: 961 Bytes
cb5178e
 
 
 
fe19ebb
 
 
d144f0b
fe19ebb
5d0c894
f04dc8c
 
 
 
46ee9f3
 
fe19ebb
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
---
license: mit
language:
- en
---
# TaxBERT

This repository accompanies the paper: Hechtner, F., Schmidt, L., Seebeck, A., & Weiß, M. (2025). How to design and employ specialized large language models for accounting and tax research: The example of TaxBERT.
TaxBERT is a domain-adapated RoBERTa model, specifically designed to analyze qualitative corporate tax disclosures.

In the future, we will add the following features:
- Tax Sentence Recognition
- Tax Risk Sentiment

**SSRN**: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5146523
          The paper provides an ‘A-to-Z’ description of how to design and employ specialized Bidirectional Encoder Representation of Transformers (BERT) models that are environmentally sustainable and practically feasible for accounting and tax researchers.

**GitHub**: https://github.com/TaxBERT/TaxBERT

If the following Guide/Repository is used for academic or scientific purposes, please cite the paper.