TY - JOUR AU - Adhikary, Prottay Kumar AU - Srivastava, Aseem AU - Kumar, Shivani AU - Singh, Salam Michael AU - Manuja, Puneet AU - Gopinath, Jini K AU - Krishnan, Vijay AU - Gupta, Swati Kedia AU - Deb, Koushik Sinha AU - Chakraborty, Tanmoy PY - 2024 DA - 2024/7/23 TI - Exploring the Efficacy of Large Language Models in Summarizing Mental Health Counseling Sessions: Benchmark Study JO - JMIR Ment Health SP - e57306 VL - 11 KW - mental health KW - counseling summarization KW - large language models KW - digital health KW - artificial intelligence KW - AI AB - Background: Comprehensive session summaries enable effective continuity in mental health counseling, facilitating informed therapy planning. However, manual summarization presents a significant challenge, diverting experts’ attention from the core counseling process. Leveraging advances in automatic summarization to streamline the summarization process addresses this issue because this enables mental health professionals to access concise summaries of lengthy therapy sessions, thereby increasing their efficiency. However, existing approaches often overlook the nuanced intricacies inherent in counseling interactions. Objective: This study evaluates the effectiveness of state-of-the-art large language models (LLMs) in selectively summarizing various components of therapy sessions through aspect-based summarization, aiming to benchmark their performance. Methods: We first created Mental Health Counseling-Component–Guided Dialogue Summaries, a benchmarking data set that consists of 191 counseling sessions with summaries focused on 3 distinct counseling components (also known as counseling aspects). Next, we assessed the capabilities of 11 state-of-the-art LLMs in addressing the task of counseling-component–guided summarization. The generated summaries were evaluated quantitatively using standard summarization metrics and verified qualitatively by mental health professionals. Results: Our findings demonstrated the superior performance of task-specific LLMs such as MentalLlama, Mistral, and MentalBART evaluated using standard quantitative metrics such as Recall-Oriented Understudy for Gisting Evaluation (ROUGE)-1, ROUGE-2, ROUGE-L, and Bidirectional Encoder Representations from Transformers Score across all aspects of the counseling components. Furthermore, expert evaluation revealed that Mistral superseded both MentalLlama and MentalBART across 6 parameters: affective attitude, burden, ethicality, coherence, opportunity costs, and perceived effectiveness. However, these models exhibit a common weakness in terms of room for improvement in the opportunity costs and perceived effectiveness metrics. Conclusions: While LLMs fine-tuned specifically on mental health domain data display better performance based on automatic evaluation scores, expert assessments indicate that these models are not yet reliable for clinical application. Further refinement and validation are necessary before their implementation in practice. SN - 2368-7959 UR - https://mental.jmir.org/2024/1/e57306 UR - https://doi.org/10.2196/57306 DO - 10.2196/57306 ID - info:doi/10.2196/57306 ER -