--- license: mit tags: - truthfulness - bert - text-classification - dual-classifier pipeline_tag: text-classification --- # Truthfulness Detection Model Fine-tuned BERT model for detecting truthfulness in text at both token and sentence levels. ## Model Description This model uses a dual-classifier architecture on top of BERT to: - Classify truthfulness at the sentence level (returns probability 0-1) - Classify truthfulness for each token (returns probability 0-1 per token) Low scores indicate likely false statements, high scores indicate likely true statements. ## Example Output For "The earth is flat.": - Sentence score: 0.0736 (7.36% - correctly identified as false) - Token scores: ~0.10 for each token ## Training - Base model: bert-base-uncased - Training samples: 6,330 - Epochs: 3 - Batch size: 16 - Training time: 49 seconds on H100 ## Custom Architecture Required ⚠️ This model uses a custom `BERTForDualTruthfulness` class. You cannot load it with standard AutoModel. See the [implementation code](https://huggingface.co/prompterminal/classifier/blob/main/model_architecture.py) for the model class definition.--- license: mit ---