| | --- |
| | library_name: transformers |
| | license: apache-2.0 |
| | language: |
| | - de |
| | base_model: |
| | - google-bert/bert-base-german-cased |
| | pipeline_tag: token-classification |
| | --- |
| | |
| | # C-BERT |
| |
|
| | CausalBERT (C-BERT) is a multi-task fine-tuned German BERT that extracts causal attributions. |
| |
|
| | ## Model details |
| | - **Model architecture**: BERT-base-German-cased + token & relation heads |
| | - **Fine-tuned on**: environmental causal attribution corpus (German) |
| | - **Tasks**: |
| | 1. Token classification (BIO tags for INDICATOR / ENTITY) |
| | 2. Relation classification (CAUSE, EFFECT, INTERDEPENDENCY) |
| |
|
| | ## Usage |
| | Find the custom [library](https://github.com/norygami/causalbert). Once installed, run inference like so: |
| | ```python |
| | from transformers import AutoTokenizer |
| | from causalbert.infer import load_model, analyze_sentence_with_confidence |
| | |
| | model, tokenizer, config, device = load_model("norygano/C-BERT") |
| | result = analyze_sentence_with_confidence( |
| | model, tokenizer, config, "Autoverkehr verursacht Bienensterben.", [] |
| | ) |
| | ``` |
| |
|
| | ## Training |
| |
|
| | - **Base model**: `google-bert/bert-base-german-cased` |
| | - **Epochs**: 3, **LR**: 2e-5, **Batch size**: 8 |
| | - See [train.py](https://github.com/norygami/causalbert/blob/main/causalbert/train.py) for details. |
| |
|
| | ## Limitations |
| |
|
| | - Only German. |
| | - Sentence-level; doesn’t handle cross-sentence causality. |
| | - Relation classification depends on detected spans — errors in token tagging propagate. |