YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
Fine-tuned DistilBERT for binary depression classification from social media text (Twitter + Reddit).
Trained as part of the ExplainDepression Pipeline β a research system for explainable and fair depression detection using SHAP attributions, Depression Explanation Score (DES), and Explanation Consistency Index across Demographics (ECID).
βββββββββββββββββββββββββββ
PERFORMANCE Accuracy β 96.17% F1 Score β 0.9617 AUC-ROC β 0.9937 ECID β 0.2063 (bias detected in 3 clinical categories)
βββββββββββββββββββββββββββ
TRAINING DETAILS Base Model β distilbert-base-uncased Epochs β 3 Batch Size β 64 Learning Rate β 3e-5 Max Length β 96 tokens Train Size β 11,200 posts Platform β Kaggle GPU T4
βββββββββββββββββββββββββββ
LABELS LABEL_0 β Not Depressed LABEL_1 β Depressed
βββββββββββββββββββββββββββ
HOW TO USE
from transformers import pipeline
clf = pipeline( "text-classification", model="mdsajjadullah/explainDepression-distilbert", return_all_scores=True )
result = clf("I feel hopeless and completely alone.")
LABEL_1 score = depression probability
βββββββββββββββββββββββββββ
INTENDED FOR RESEARCH ONLY. Not for clinical diagnosis or deployment without professional mental health oversight.
- Downloads last month
- 15