DepressLLM: Interpretable domain-adapted language model for depression detection from real-world narratives
By: Sehwan Moon , Aram Lee , Jeong Eun Kim and more
Potential Business Impact:
Helps find depression from people's stories.
Advances in large language models (LLMs) have enabled a wide range of applications. However, depression prediction is hindered by the lack of large-scale, high-quality, and rigorously annotated datasets. This study introduces DepressLLM, trained and evaluated on a novel corpus of 3,699 autobiographical narratives reflecting both happiness and distress. DepressLLM provides interpretable depression predictions and, via its Score-guided Token Probability Summation (SToPS) module, delivers both improved classification performance and reliable confidence estimates, achieving an AUC of 0.789, which rises to 0.904 on samples with confidence $\geq$ 0.95. To validate its robustness to heterogeneous data, we evaluated DepressLLM on in-house datasets, including an Ecological Momentary Assessment (EMA) corpus of daily stress and mood recordings, and on public clinical interview data. Finally, a psychiatric review of high-confidence misclassifications highlighted key model and data limitations that suggest directions for future refinements. These findings demonstrate that interpretable AI can enable earlier diagnosis of depression and underscore the promise of medical AI in psychiatry.
Similar Papers
MDD-LLM: Towards Accuracy Large Language Models for Major Depressive Disorder Diagnosis
Computation and Language
Helps doctors find depression faster and better.
Interpretable Depression Detection from Social Media Text Using LLM-Derived Embeddings
Computation and Language
Finds sad posts to help people feel better.
Generating Medically-Informed Explanations for Depression Detection using LLMs
Computation and Language
Finds depression early from online posts.