YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

Fine-tuned DistilBERT for binary depression classification from social media text (Twitter + Reddit).

Trained as part of the ExplainDepression Pipeline β€” a research system for explainable and fair depression detection using SHAP attributions, Depression Explanation Score (DES), and Explanation Consistency Index across Demographics (ECID).

━━━━━━━━━━━━━━━━━━━━━━━━━━━

PERFORMANCE Accuracy β†’ 96.17% F1 Score β†’ 0.9617 AUC-ROC β†’ 0.9937 ECID β†’ 0.2063 (bias detected in 3 clinical categories)

━━━━━━━━━━━━━━━━━━━━━━━━━━━

TRAINING DETAILS Base Model β†’ distilbert-base-uncased Epochs β†’ 3 Batch Size β†’ 64 Learning Rate β†’ 3e-5 Max Length β†’ 96 tokens Train Size β†’ 11,200 posts Platform β†’ Kaggle GPU T4

━━━━━━━━━━━━━━━━━━━━━━━━━━━

LABELS LABEL_0 β†’ Not Depressed LABEL_1 β†’ Depressed

━━━━━━━━━━━━━━━━━━━━━━━━━━━

HOW TO USE

from transformers import pipeline

clf = pipeline( "text-classification", model="mdsajjadullah/explainDepression-distilbert", return_all_scores=True )

result = clf("I feel hopeless and completely alone.")

LABEL_1 score = depression probability

━━━━━━━━━━━━━━━━━━━━━━━━━━━

INTENDED FOR RESEARCH ONLY. Not for clinical diagnosis or deployment without professional mental health oversight.

Downloads last month
15
Safetensors
Model size
67M params
Tensor type
F32
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support