causal-narrative's picture
Upload README.md with huggingface_hub
5a2df2b verified
---
datasets:
- CausalNewsCorpus
language: en
library_name: transformers
license: mit
metrics:
- accuracy
- f1
- precision
- recall
tags:
- text-classification
- roberta
- causal-narrative
- sequence-classification
---
# RoBERTa Causal Narrative Classifier
This model is a fine-tuned version of `roberta-base` for causal narrative sentence classification.
## Model Description
- **Base Model**: roberta-base
- **Task**: Binary classification (causal vs non-causal sentences)
- **Training Data**: CausalNewsCorpus V2
## Training Results
- **Accuracy**: 83.82%
- **Precision**: 84.31%
- **Recall**: 83.20%
- **F1 Score**: 83.48%
## Usage
```python
from transformers import RobertaTokenizer, RobertaForSequenceClassification
import torch
# Load model and tokenizer
model_name = "causal-narrative/roberta-causal-narrative-classifier"
tokenizer = RobertaTokenizer.from_pretrained(model_name)
model = RobertaForSequenceClassification.from_pretrained(model_name)
# Predict
text = "The heavy rain caused flooding in the city."
inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True)
with torch.no_grad():
outputs = model(**inputs)
prediction = torch.argmax(outputs.logits, dim=-1).item()
print(f"Is causal: {prediction == 1}")
```
## Labels
- **0**: Non-causal sentence
- **1**: Causal narrative sentence