<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Ment Health</journal-id><journal-id journal-id-type="publisher-id">mental</journal-id><journal-id journal-id-type="index">16</journal-id><journal-title>JMIR Mental Health</journal-title><abbrev-journal-title>JMIR Ment Health</abbrev-journal-title><issn pub-type="epub">2368-7959</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v12i1e72038</article-id><article-id pub-id-type="doi">10.2196/72038</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Explainable AI for Depression Detection and Severity Classification From Activity Data: Development and Evaluation Study of an Interpretable Framework</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Ahmed</surname><given-names>Iftikhar</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Brahmacharimayum</surname><given-names>Anushree</given-names></name><degrees>MA</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Ali</surname><given-names>Raja Hashim</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Khan</surname><given-names>Talha Ali</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Ahmad</surname><given-names>Muhammad Ovais</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Software Engineering, University of Europe for Applied Sciences, Germany</institution><addr-line>Potsdam</addr-line><country>Germany</country></aff><aff id="aff2"><institution>Department of Mathematics and Computer Science, Karlstad University</institution><addr-line>Universitetsgatan 2</addr-line><addr-line>Karlstad</addr-line><country>Sweden</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Torous</surname><given-names>John</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Shaheen</surname><given-names>Ifra</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Sherimon</surname><given-names>Vinu</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Zhou</surname><given-names>Zumu</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Muhammad Ovais Ahmad, PhD, Department of Mathematics and Computer Science, Karlstad University, Universitetsgatan 2, Karlstad, 65188, Sweden, 46 76-113 22 49; <email>ovais.ahmad@kau.se</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>11</day><month>9</month><year>2025</year></pub-date><volume>12</volume><elocation-id>e72038</elocation-id><history><date date-type="received"><day>01</day><month>02</month><year>2025</year></date><date date-type="rev-recd"><day>13</day><month>07</month><year>2025</year></date><date date-type="accepted"><day>17</day><month>07</month><year>2025</year></date></history><copyright-statement>&#x00A9; Iftikhar Ahmed, Anushree Brahmacharimayum, Raja Hashim Ali, Talha Ali Khan, Muhammad Ovais Ahmad. Originally published in JMIR Mental Health (<ext-link ext-link-type="uri" xlink:href="https://mental.jmir.org">https://mental.jmir.org</ext-link>), 11.9.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mental.jmir.org/">https://mental.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mental.jmir.org/2025/1/e72038"/><abstract><sec><title>Background</title><p>Depression is one of the most prevalent mental health disorders globally, affecting approximately 280 million people and frequently going undiagnosed or misdiagnosed. The growing ubiquity of wearable devices enables continuous monitoring of activity levels, providing a new avenue for data-driven detection and severity assessment of depression. However, existing machine learning models often exhibit lower performance when distinguishing overlapping subtypes of depression and frequently lack explainability, an essential component for clinical acceptance.</p></sec><sec><title>Objective</title><p>This study aimed to develop and evaluate an interpretable machine learning framework for detecting depression and classifying its severity using wearable-actigraphy data, while addressing common challenges such as imbalanced datasets and limited model transparency.</p></sec><sec sec-type="methods"><title>Methods</title><p>We used the Depresjon dataset and applied Adaptive Synthetic Sampling (ADASYN) to mitigate class imbalance. We extracted multiple statistical features (eg, power spectral density mean and autocorrelation) and demographic attributes (eg, age) from the raw activity data. Five machine learning algorithms (logistic regression, support vector machines, random forest, XGBoost, and neural networks) were assessed via accuracy, precision, recall, <italic>F</italic><sub>1</sub>-score, specificity, and Matthew correlation constant. We further used Shapley Additive Explanations (SHAP) and Local Interpretable Model-agnostic Explanations (LIME) to elucidate prediction drivers.</p></sec><sec sec-type="results"><title>Results</title><p>XGBoost achieved the highest overall accuracy of 84.94% for binary classification and 85.91% for multiclass severity. SHAP and LIME revealed power spectral density mean, age, and autocorrelation as top predictors, highlighting circadian disruptions&#x2019; role in depression.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>Our interpretable framework reliably identifies depressed versus nondepressed individuals and differentiates mild from moderate depression. The inclusion of SHAP and LIME provides transparent, clinically meaningful insights, emphasizing the potential of explainable artificial intelligence to enhance early detection and intervention strategies in mental health care.</p></sec></abstract><kwd-group><kwd>artificial intelligence</kwd><kwd>explainable AI</kwd><kwd>depression</kwd><kwd>mental health</kwd><kwd>machine learning</kwd><kwd>activity data</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Mental disorder is a global phenomenon with over a billion individuals suffering from various types of mental disorders. Mental health disorder usually affects a person&#x2019;s thinking, interpersonal relationships, and emotions. It results in serious consequences such as risk of other diseases and suicide if not identified and treated properly in time. Depression is one of the most prevalent mental health disorders that can negatively impact an individual&#x2019;s daily functioning and well-being. Studies have shown an increase in the number of depression and anxiety disorders during and post-COVID&#x2013;19 pandemic, affecting both males and female demographics [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]. Younger demographics are found to suffer more from mental disorders than older citizens [<xref ref-type="bibr" rid="ref3">3</xref>]. It is estimated that approximately 280 million of the population is suffering from depression [<xref ref-type="bibr" rid="ref4">4</xref>].</p><p>Depression is characterized by changes in one&#x2019;s temper or a sudden hysterical attitude to any circumstances or situation. Depression is regarded as the second most common reason for death, and 0.8 million deaths are attributed to depression [<xref ref-type="bibr" rid="ref5">5</xref>]. Depression is an emotional disorder that triggers a sense of desolation, pessimism, and anhedonia that affect a person&#x2019;s life. This leads to the disruption of daily activities in all aspects of life, influencing others around them as well. The etiology of these types of mental illness results from environmental and genetic factors [<xref ref-type="bibr" rid="ref6">6</xref>]. A depressive illness can be classified as mild, moderate, and severe by degree of severity [<xref ref-type="bibr" rid="ref7">7</xref>]. It is challenging for the physicians to diagnose it in the early stages and is often misdiagnosed as conditions such as vitamin deficiency [<xref ref-type="bibr" rid="ref8">8</xref>]. It has also been linked with the rise of risk in cancer [<xref ref-type="bibr" rid="ref9">9</xref>] and cardiovascular ailment [<xref ref-type="bibr" rid="ref10">10</xref>] which can potentially decrease the life span of an individual. Even though it has high prevalence, it is often left untreated in many of the cases [<xref ref-type="bibr" rid="ref11">11</xref>].</p><p>Depression rating scale is a screening tool that comprises a questionnaire that is used for detecting the severity phases of depression [<xref ref-type="bibr" rid="ref12">12</xref>]. Montgomery-Asberg Depression Rating Scale (MADRS) was designed by Marie Asberg [<xref ref-type="bibr" rid="ref13">13</xref>]. It focuses on ten signs of depression, which include apparent sadness, reported sadness, inner tension, reduced sleep, reduced appetite, concentration difficulties, lassitude, inability to feel, pessimistic thoughts, and suicidal thoughts [<xref ref-type="bibr" rid="ref14">14</xref>]. The participant answers the questions with a score of 0 to 6. It is categorized as: 0 to 6 for normal or absence of symptom, 7 to 19 for mild, 20 to 34 for moderate and 35 to 60 for severe depression [<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>Variability is also seen in treatment response. While a person may respond well to treatment, others with the same symptoms may not respond to the same treatment in the same manner. Thus, posing a threat to an individual&#x2019;s life as prolonging its diagnosis and treatment can cause serious conditions [<xref ref-type="bibr" rid="ref16">16</xref>]. The diagnosis of depression is very challenging, and it is often misdiagnosed as it often mimics other types of disorders [<xref ref-type="bibr" rid="ref17">17</xref>]. One such example is it is often misdiagnosed as anxiety [<xref ref-type="bibr" rid="ref18">18</xref>]. Therefore, early detection and diagnosis of depression is paramount.</p><p>Over the preceding years, the domain of artificial intelligence (AI) has been evolving, and it has shown success in a plethora of fields including health care [<xref ref-type="bibr" rid="ref19">19</xref>]. Machine learning (ML, a sub-field of AI) can potentially predict whether a person has a mental disorder or not, and if there is a degree of inclination to suicide [<xref ref-type="bibr" rid="ref20">20</xref>]. Therefore, implementation of ML may benefit mental illness domain by characterizing those who are vulnerable and thus help the practitioners in the early identification of mental health disorders including depression.</p><p>The existing ML-based models for depression detection face several challenges, such as their inability to accurately differentiate between various types of depression. More advanced and sophisticated techniques such as deep neural networks are often restricted in their usage due to their &#x201C;black box&#x201D; approach, thereby lacking transparency and explainability [<xref ref-type="bibr" rid="ref19">19</xref>]. This lack of interpretability is a major hindrance in health care applications where understanding the decision-making process is crucial [<xref ref-type="bibr" rid="ref19">19</xref>]. In the context of diagnosis and treatment of depression, transparency and explainability are even more important.</p><p>AI models for depression require huge amounts of data which capture behavioral, psychological, and physiological indicators for effective modeling, and thus accurate detection and differentiation of depressive states. However, the collection of such datasets is often not possible due to challenges and limitations of data-collection methods [<xref ref-type="bibr" rid="ref21">21</xref>]. An alternative is to use wearable devices to collect continuous, real-time data on various aspects such as physical activity, sleep patterns, heart-rate variability, and many other physiological indicators [<xref ref-type="bibr" rid="ref21">21</xref>].</p><p>A number of studies have been conducted using a variety of datasets and ML models for diagnosis of several types of mental disorders. Paul and Juliet [<xref ref-type="bibr" rid="ref22">22</xref>] considered the prediction of mental health disorders problem based on a dataset of self-reported information by the patients. The authors used four ML algorithms, namely logistic regression, k-nearest neighbor, decision tree, and random forest (RF). RF is identified as the best performing algorithm achieving an accuracy of 81%. However, the authors did not discuss other performance measures such as recall, precision, and <italic>F</italic><sub>1</sub>-score. Zhang et al [<xref ref-type="bibr" rid="ref23">23</xref>] suggested an architecture that uses magnetoencephalography to classify posttraumatic stress disorder. The authors used support vector machine as the main classifier with a recursive RF for feature selection. The model with selective features achieved an area under the curve value up to 0.9. Mikolas et al [<xref ref-type="bibr" rid="ref24">24</xref>] focused on the automatic diagnosis of attention deficit hyperactivity disorder and argued that although recent advancements in the domain of ML helped in binary classification problem of identification of attention deficit hyperactivity disorder and healthy individuals, the problem of differentiating among multiple psychiatric conditions remains. The authors considered a dataset of 229 anonymized participants and observed that the best performance is obtained using automated feature selections. The study has certain limitations, such as a higher number of missing values in certain columns and lack of out-of-sample testing.</p><p>Spulber et al [<xref ref-type="bibr" rid="ref25">25</xref>] explored whether activity data collected via actigraphy and depression severity levels are correlated. The study involved individuals experiencing major depressive episodes who were not receiving antidepressant treatment. The authors trained models to predict the MADRS score from activity data using ML techniques, selecting models based on accuracy and precision. These models were ranked as best, intermediate, or worst based on root mean squared error values. External validation was also conducted using data from patients undergoing treatment. However, this external validation introduced potential biases due to differences in patient conditions and treatment status. Garcia-Ceja et al [<xref ref-type="bibr" rid="ref26">26</xref>] compared 2 models, namely RF and deep neural network (DNN), to identify healthy and nonhealthy participants from psychomotor data using the Depresjon dataset, which contains actigraphy data of control and condition groups. Due to class imbalance, different sampling methods such as random oversampling and Synthetic Minority Oversampling Technique (SMOTE) were adopted and were then compared with baseline and no oversampling after extracting three statistical features. It was found that RF with SMOTE achieved a higher <italic>F</italic><sub>1</sub>-score of 0.73.</p><p>Frogner et al [<xref ref-type="bibr" rid="ref27">27</xref>] implemented 1D convolutional neural networks (CNNs) to detect depression using activity data. Models were trained with different time segments, with 48-hours segments found to be optimal, to perform 3 classification tasks. The first task classifies depressive groups, the second classifies levels of depression (no depression, mild depression, and moderate depression based on MADRS score ranges: 0-10, 11-19, and &#x003E;20, respectively), and the last task predicts the MADRS score. The evaluation was conducted using leave-one-participant-out cross-validation combined with majority voting. An <italic>F</italic><sub>1</sub>-score of 70% was achieved for the first task, 30% for the second, and a mean square error (MSE) of 4.0 for the last one. The model performed well in classifying nondepressed participants but struggled with mild and moderate depression levels. Jakobsen et al [<xref ref-type="bibr" rid="ref28">28</xref>] examined whether objective metrics of activity data could enhance current diagnostic techniques for depression by analyzing activity patterns using ML models. Three statistical attributes (mean activity level, standard deviation, and proportion of zero activity) were derived for RF and DNN models. CNNs were trained using raw activity data represented as a 24 &#x00D7; 60 image matrix. DNN with SMOTE achieved an accuracy of 84%, a true positive rate of 82%, and a true negative rate of 84%. However, a significant amount of misclassifications was observed, particularly in groups with overlapping activity patterns, such as patients with agitation.</p><p>Espino-Salinas et al [<xref ref-type="bibr" rid="ref29">29</xref>] used the Depresjon dataset to create a framework for depression classification. Preprocessing involved converting motor activity data into vectors and creating matrices for input into neural networks. A 2D-CNN and DNN were then used to classify depressive and nondepressive subjects. The 2D-CNN achieved an accuracy of 76% and an <italic>F</italic><sub>1</sub>-score of 72%, while the DNN achieved an accuracy of 72% and an <italic>F</italic><sub>1</sub>-score of 69%. The <italic>F</italic><sub>1</sub>-score was selected as the primary metric due to the imbalanced dataset, as it better reflects the model&#x2019;s ability to handle unequal class distributions. The 2D-CNN outperformed DNN by achieving a better balance between sensitivity and specificity, demonstrating its potential for depression classification. To maintain focus and address space limitations, we direct readers to [<xref ref-type="bibr" rid="ref30">30</xref>-<xref ref-type="bibr" rid="ref35">35</xref>] for further details.</p><p>Despite significant advancements in ML techniques for mental health diagnosis, several gaps remain. A recurring limitation across studies is the relatively lower performance in distinguishing overlapping or nuanced conditions, particularly for depression severity levels. Models perform well in binary classifications (such as distinguishing between depressed and nondepressed individuals), but struggle with multiclass classifications or subtle variations in conditions, such as mild versus moderate depression. Another critical gap is the lack of explainable artificial intelligence (XAI) techniques in these studies. Current ML models, including CNNs and DNNs, operate as black box approaches. The black box approach means that these techniques offer limited insight into their decision-making processes. This lack of interpretability reduces trust and limits the practical application in clinical settings, where understanding why a model makes a specific prediction is as crucial as the prediction itself.</p><p>In this work, we use the data collected from wearable devices to investigate the application of ML algorithms in depression detection and severity level classification tasks. We use Adaptive Synthetic Sampling Technique (ADASYN) for addressing imbalanced datasets. Further, feature engineering is performed to extract meaningful features from the activity data. For the binary (depression: yes or no) and multiclass (predicting severity levels) classifications, we implement logistic regression, support vector machines, RF, XGBoost, and neural networks. The performance of the models is evaluated using accuracy, precision, recall, <italic>F</italic><sub>1</sub>-score, and Matthew correlation constant (MCC). To further contribute toward the explainability of the model, we used Shapley Additive Explanations (SHAP) and Local Interpretable Model-agnostic Explanations (LIME) methods to explain the decision-making process of the best-performing models.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Overview</title><p><xref ref-type="fig" rid="figure1">Figure 1</xref> is a graphical representation of our methodology. We acquired data from the Depresjon dataset and conducted exploratory data analysis to understand the nature of the data. We observed a class imbalance, which is addressed using sampling techniques, followed by normalization of data. We then applied feature engineering to extract features from the data. This was followed by designing models for binary (depressed vs nondepressed) and multiclass (severity levels) classification tasks. After model implementation, we used accuracy, precision, recall, specificity, MCC, and <italic>F</italic><sub>1</sub>-score as performance evaluation criteria. Finally, we applied the SHAP and LIME methods to explain the decision-making process of the selected model.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Graphical representation of methodology.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig01.png"/></fig></sec><sec id="s2-2"><title>Data Acquisition</title><p>The dataset adopted in this work is the &#x201C;Depresjon&#x201D; (Norwegian word for depression) dataset, available at Simula repository [<xref ref-type="bibr" rid="ref36">36</xref>]. The dataset consists of motor activity of individuals recorded per minute for several days by using an actigraphy watch. The dataset consists of 2 parts: The first part comprises measurements of activity levels related to each individual in the condition and control group. The files consist of timestamp, date, and activity columns. The second part comprises the demographic information related to each participant. The dataset is enriched with several variables: an identification number (number), days with measurements (days), gender (1 for female and 2 for male), and age group of the participant (age). It also records the type of affective disorder (afftype). The presence of melancholia is noted with 1 for melancholia and 2 for lack of melancholia (melanch). The dataset differentiates between inpatients &#x2018;1&#x2019; and outpatients &#x2018;2&#x2019; (inpatient). Education level is categorized by years completed (education). Marital status is recorded as 1 for married or cohabiting and 2 for single (marriage). Work status is coded as 1 for working or studying and 2 for unemployed or sick leave or pension (work). Finally, the dataset includes MADRS scores before (madrs1) and after (madrs2) the inspection.</p><p>The original dataset comprised 2 folders (condition and control); it was important to merge the data into a single coherent set. The resultant dataset after the merge operation has 1440 data points. For binary classification (depressed vs nondepressed), the &#x201C;condition&#x201D; group is labeled as 1, and the &#x201C;control&#x201D; group as 0. For multiclass classification (assessing the severity levels of depression), the labels were assigned as follows: 0 for normal, 1 for mild depression, and 2 for moderate depression. Participants were categorized based on their MADRS score: 0&#x2010;6 for &#x201C;Normal&#x201D; (no depression), 7&#x2010;19 for &#x201C;Mild&#x201D; depression, and 20&#x2010;34 for &#x201C;Moderate&#x201D; depression. Note that although the standard MADRS scale includes a &#x201C;Severe&#x201D; category for scores &#x003E;=35, our dataset contained no cases with scores in the severe range, so no &#x201C;Severe&#x201D; class was included.</p></sec><sec id="s2-3"><title>Addressing Data Imbalance</title><p>We observed that the data is imbalanced, which means that some output classes are more prevalent than others. Imbalance data can skew the performance of ML algorithms and require redressal. We used ADASYN for class imbalance problems. The intuition behind ADASYN is the usage of a weighted distribution for the less frequent instances in accordance with their difficulty of learning [<xref ref-type="bibr" rid="ref37">37</xref>]. For the working of ADASYN, we refer the reader to He et al [<xref ref-type="bibr" rid="ref37">37</xref>].</p></sec><sec id="s2-4"><title>Feature Engineering</title><p>The dataset consists of activity measurements recorded by an actigraph watch, resulting in a continuous stream of data collected at regular intervals for each patient. However, using these raw time series values is not an effective approach, as it can lead to an unnecessarily large number of interconnected features and fails to capture the nuanced behaviors and physiological rhythms indicative of mental health status. In the context of depression and related mood disorders, subtle variations in daily activity patterns, energy distribution across frequencies, and nonlinear fluctuations are particularly revealing. To address this, we summarized the continuous time-series data into key statistical features, thereby translating raw sensor data into indicators more closely tied to psychological and clinical insights. This feature engineering approach is a standard mechanism for managing continuous data streams while preserving essential patterns and variability [<xref ref-type="bibr" rid="ref38">38</xref>], effectively reducing data dimensionality and computational requirements.</p><p>To capture the multifaceted nature of human activity and its potential connections to depressive symptoms, we systematically extracted features from 5 broad categories: (1) time domain, (2) frequency domain, (3) circadian rhythm analysis, and (4) transition features. Each category addresses different aspects of activity that can reflect mood changes or disruptions in daily routines often associated with mental health conditions.</p></sec><sec id="s2-5"><title>Time-Domain Features</title><p>Time-domain features describe statistical properties and variability in the activity signal. Disruptions or heightened variability in these measures can be linked to changes in energy levels and daily routines often observed in individuals experiencing mood disturbances. Time-domain features (such as mean activity, variability, higher-order moments, autocorrelation, and entropy) have long anchored actigraphy research. Historic studies show these measures consistently distinguish depressive from nondepressive motor patterns, providing a compact, behaviorally interpretable window on psychomotor change [<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref27">27</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref38">38</xref>].</p><p>The following time domain features are extracted from the continuous stream of data.</p><list list-type="bullet"><list-item><p>Mean SD: summarize central tendency and overall variability of the recorded activity levels. Marked fluctuations might indicate irregular sleep-wake cycles or energy instabilities common in depressive episodes.</p></list-item><list-item><p>Skewness and kurtosis: capture the asymmetry and the &#x201C;peakedness&#x201D; of the activity distribution. Drastic changes in these metrics could signify periods of lethargy or bursts of hyperactivity, reflecting possible mood state shifts.</p></list-item><list-item><p>IQR: Measures the spread between the 25th and 75th percentiles, providing a robust indication of how condensed or spread out the daily activity is, which might correlate with mood variability.</p></list-item><list-item><p>Autocorrelation (AutoCorr): examines how similar activity levels are to themselves at varying time lags, helping detect repetitive patterns or circadian disruptions often seen in mental health disorders.</p></list-item><list-item><p>Entropy: assesses the randomness or irregularity in activity patterns. High entropy might suggest erratic daily schedules, potentially symptomatic of poor sleep hygiene or mood instability.</p></list-item></list></sec><sec id="s2-6"><title>Frequency-Domain Features</title><p>Analyzing the distribution of activity power across different frequencies can illuminate underlying periodicities. Alterations in these periodic components often correspond to disrupted sleep-wake cycles or abnormal energy distribution, which are common in depression. Multiple actigraphy studies link these spectral abnormalities to symptom severity, confirming their diagnostic value for mood inference [<xref ref-type="bibr" rid="ref25">25</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref38">38</xref>].</p><p>The features extracted are:</p><list list-type="bullet"><list-item><p>Power spectral density (mean): a measure of how the signal&#x2019;s power is spread across frequencies, reflecting the overall strength of periodic components in daily activity [<xref ref-type="bibr" rid="ref39">39</xref>].</p></list-item><list-item><p>Dominant frequency: identifies the frequency at which activity is most pronounced, commonly associated with daily (circadian) cycles. Deviations may mirror disrupted or shifted circadian rhythms in individuals with depressive symptoms.</p></list-item><list-item><p>Spectral entropy: quantifies how &#x201C;spread out&#x201D; or concentrated the activity power is across frequencies. Greater spread could imply less stable routines and more variability in daily patterns, potentially linked to mood fluctuations.</p></list-item></list></sec><sec id="s2-7"><title>Circadian Rhythm Analysis</title><p>Mood disorders are often accompanied by disruptions in circadian rhythms. Examining the timing of key transitions can yield valuable clues about sleep-wake irregularities [<xref ref-type="bibr" rid="ref40">40</xref>] (Activity Onset and Offset: identify the start and end of the primary daily activity period. Shifts in these values can reveal delayed or advanced phases of sleep, which are frequently observed in individuals with depression or other affective disorders).</p></sec><sec id="s2-8"><title>Transition Features</title><p>Activity does not always change smoothly throughout the day; abrupt transitions or steady patterns can be particularly telling of mental health status. Transition entropy captures the uncertainty or unpredictability inherent in the sequence of activity state transitions [<xref ref-type="bibr" rid="ref41">41</xref>]. Transition entropy quantifies the randomness in the transitions between different activity states by evaluating the distribution of transition probabilities. Elevated transition entropy may reflect instability in daily routines or mood fluctuations, which are often observed in individuals with certain mental health conditions.</p><p>Alongside the time- and frequency-domain metrics, we also tested a set of nonlinear descriptors (notably Fractal Dimension and the Hurst Exponent) to better capture the self-similar structure and long-range dependencies often observed in human activity signals. We implemented several published algorithms and tuned their parameters across multiple window lengths. Despite these efforts, the extraction pipeline consistently produced a large fraction of undefined or zero values, signaling numerical instability with our data resolution. After troubleshooting and sensitivity checks, the issue persisted and threatened to compromise both model robustness and interpretability. To safeguard the overall validity of the study, we therefore documented the attempted nonlinear analyses and excluded these features from the final feature set. <xref ref-type="fig" rid="figure2">Figure 2</xref> is a pictorial representation of the dataset after feature engineering.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Snap of data after feature engineering.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig02.png"/></fig><p>After feature engineering, we performed data normalization using <italic>z</italic>-score normalization technique as shown in <xref ref-type="disp-formula" rid="E1">Eq. (1)</xref>.</p><disp-formula id="E1"><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>z</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>&#x03BC;</mml:mi></mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mfrac></mml:mrow></mml:mstyle></mml:math></disp-formula><p>Note that <inline-formula><mml:math id="ieqn1"><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> and <inline-formula><mml:math id="ieqn2"><mml:msub><mml:mrow><mml:mi>z</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> represent the individual feature score before and after standardization, whereas <inline-formula><mml:math id="ieqn3"><mml:mi>&#x03BC;</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="ieqn4"><mml:mi>&#x03C3;</mml:mi></mml:math></inline-formula> represents the mean and standard deviation of the activity level scores.</p><sec id="s2-8-1"><title>Logistic Regression</title><p>Logistic Regression (LR) is a widely used supervised learning algorithm primarily employed for classification tasks [<xref ref-type="bibr" rid="ref42">42</xref>]. LR estimates the probability that a given input belongs to a particular class by applying the logistic (sigmoid) function to a linear combination of input features. LR is favored for its simplicity, interpretability, and efficiency, especially when the relationship between the features and the target variable is approximately linear. It also provides probabilistic insights, allowing for the assessment of confidence in predictions. In addition, LR can be extended to multiclass classification problems using strategies such as One-vs-Rest (OvR) or multinomial LR [<xref ref-type="bibr" rid="ref43">43</xref>]. Despite its simplicity, LR serves as a strong baseline model and is effective in scenarios where the decision boundary is linear or nearly linear.</p></sec><sec id="s2-8-2"><title>Support Vector Machines</title><p>Support vector machines (SVM) are a popular model notably for classification problems. The core concept of SVM is to identify an optimal hyperplane that separates or distinguishes data points into different classes in an n-dimensional space [<xref ref-type="bibr" rid="ref44">44</xref>]. For binary classification problems, SVM finds the optimal hyperplane that separates the 2 classes in the feature space. For the multiclass problem, two approaches are typically used: OvR and One-vs-One (OvO). In OvR, a binary classifier is used for each class against all other classes, whereas in OvO, a binary classifier is trained for every possible pair of classes [<xref ref-type="bibr" rid="ref44">44</xref>]. This means that for a <inline-formula><mml:math id="ieqn5"><mml:mi>k</mml:mi></mml:math></inline-formula>-class classification problem, OvO trains <inline-formula><mml:math id="ieqn6"><mml:mfrac><mml:mrow><mml:mi>N</mml:mi><mml:mo>&#x22C5;</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:mi>N</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac></mml:math></inline-formula> binary classifiers.</p></sec><sec id="s2-8-3"><title>Random Forest</title><p>RF is a supervised ensemble learning algorithm that integrates the results of multiple decision trees to produce a more precise outcome [<xref ref-type="bibr" rid="ref45">45</xref>]. Unlike decision trees, random forest has the ability to minimize overfitting [<xref ref-type="bibr" rid="ref45">45</xref>]. RF is based on the bagging method where multiple models are trained using the subsets of features that are randomly selected, and the resulting predictions are aggregated either by voting or average mechanism. For regression, the prediction is the average prediction across all the decision trees, and for classification, it is the majority vote class predicted across the decision trees [<xref ref-type="bibr" rid="ref46">46</xref>].</p></sec><sec id="s2-8-4"><title>XGBoost</title><p>XGBoost (Extreme gradient boosting) is a ML algorithm based on the Gradient Boosting framework [<xref ref-type="bibr" rid="ref47">47</xref>]. The idea of gradient boosting is based on developing multiple weak learners (like decision trees) with the expectation to aggregate the result to obtain a final strong learning model [<xref ref-type="bibr" rid="ref47">47</xref>]. XGBoost builds an ensemble of decision trees sequentially, where each new tree attempts to correct the errors of the previous ones. XGBoost is particularly useful in dealing with imbalanced datasets [<xref ref-type="bibr" rid="ref48">48</xref>].</p></sec><sec id="s2-8-5"><title>Neural Networks</title><p>Neural networks (NN) are a class of powerful, flexible ML models inspired by the human brain&#x2019;s network of neurons. Comprising layers of interconnected nodes (neurons), neural networks are capable of modeling complex, nonlinear relationships between input features and target variables. Each neuron applies a weighted sum of its inputs followed by a nonlinear activation function, enabling the network to learn intricate patterns and representations from the data [<xref ref-type="bibr" rid="ref49">49</xref>].</p><p>In the context of mental health analysis, neural networks can effectively capture the complex temporal and spatial dependencies present in actigraphy data, making them well-suited for tasks such as depression classification and severity assessment [<xref ref-type="bibr" rid="ref50">50</xref>]. Their ability to automatically learn feature representations from raw data reduces the need for extensive feature engineering, although they typically require larger datasets and more computational resources compared to traditional algorithms. The hyper-parameter settings are reported in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p><p>The hyper-parameter settings are reported in <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>.</p></sec></sec><sec id="s2-9"><title>Evaluation Metrics</title><p>In this study, we use accuracy, recall, specificity, precision, <italic>F</italic><sub>1</sub>-score, and MCC as performance evaluation criteria for ML models. These performance evaluation measures are based on the confusion matrix, which represents a summary of accurate and misclassified instances for a classification problem. A graphical representation of the confusion matrix is given in <xref ref-type="fig" rid="figure3">Figure 3</xref>. A confusion matrix has 4 key elements called true positive (TP), false positive (FP), false negative (FN), and true negative (TN), defined as follows [<xref ref-type="bibr" rid="ref51">51</xref>].</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Confusion matrix.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig03.png"/></fig><p>TP is an instance when the model classifies a positive instance correctly. FP is an instance when the model classifies a negative instance as a positive instance. FN is an instance when the model classifies a positive instance as a negative instance. TN is an instance when the model classifies a negative instance as a negative instance. Based on the definitions of various elements of the confusion matrix, in <xref ref-type="table" rid="table1">Table 1</xref>, we define our performance evaluation metrics.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Performance metrics.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Metric</td><td align="left" valign="bottom">Formula</td></tr></thead><tbody><tr><td align="left" valign="top">Accuracy (Acc)</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn7"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr><tr><td align="left" valign="top">Recall (Rec)</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn8"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr><tr><td align="left" valign="top">Specificity (Spec)</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn9"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr><tr><td align="left" valign="top">Precision (Prec)</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn10"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr><tr><td align="left" valign="top"><italic>F</italic><sub>1</sub>-score (F1)</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn11"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#x22C5;</mml:mo><mml:mfrac><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr><tr><td align="left" valign="top">MCC</td><td align="left" valign="top"><inline-formula><mml:math id="ieqn12"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mfrac><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>&#x00D7;</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:msqrt><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:msqrt></mml:mfrac></mml:mrow></mml:mstyle></mml:math></inline-formula></td></tr></tbody></table></table-wrap></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Performance Evaluation for Binary and Multiclass Classification Task</title><p><xref ref-type="table" rid="table2">Table 2</xref> highlights the performance of various ML models on the binary classification task. Among the evaluated models, XGBoost achieved the highest accuracy of 84.94%, significantly outperforming all other models. In contrast, LR was the least effective, with an accuracy of 70.68%. This ranking based on accuracy is consistently reflected across other metrics: XGBoost also leads in precision (83.40%), recall (87.45%), <italic>F</italic><sub>1</sub>-score (85.29%), and MCC (0.7012), while LR scores the lowest in each of these categories. The remaining models, including SVM, RF, and neural networks, show a progressive improvement in performance, with SVM outperforming LR and RF and neural networks closely trailing XGBoost. This uniform trend across all evaluation metrics underscores the robustness of XGBoost as the superior model for this classification task.</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Performance comparison of various models.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Model</td><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">Precision</td><td align="left" valign="bottom">Recall</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub>-score</td><td align="left" valign="bottom">Matthew correlation coefficient</td></tr></thead><tbody><tr><td align="left" valign="top">Logistic regression</td><td align="left" valign="top">0.7068</td><td align="left" valign="top">0.6937</td><td align="left" valign="top">0.7471</td><td align="left" valign="top">0.7187</td><td align="left" valign="top">0.4156</td></tr><tr><td align="left" valign="top">SVM</td><td align="left" valign="top">0.7789</td><td align="left" valign="top">0.7648</td><td align="left" valign="top">0.8056</td><td align="left" valign="top">0.7842</td><td align="left" valign="top">0.5595</td></tr><tr><td align="left" valign="top">XGBoost</td><td align="left" valign="top">0.8494</td><td align="left" valign="top">0.8340</td><td align="left" valign="top">0.8745</td><td align="left" valign="top">0.8529</td><td align="left" valign="top">0.7012</td></tr><tr><td align="left" valign="top">Random forest</td><td align="left" valign="top">0.8301</td><td align="left" valign="top">0.8096</td><td align="left" valign="top">0.8663</td><td align="left" valign="top">0.8358</td><td align="left" valign="top">0.6639</td></tr><tr><td align="left" valign="top">Neural networks</td><td align="left" valign="top">0.8368</td><td align="left" valign="top">0.8275</td><td align="left" valign="top">0.8543</td><td align="left" valign="top">0.8397</td><td align="left" valign="top">0.6758</td></tr></tbody></table></table-wrap><p>Furthermore, XGBoost exhibited an enhancement over the average performance of the other models across all evaluation metrics. Specifically, XGBoost surpassed the average accuracy, precision, recall, <italic>F</italic><sub>1</sub>-score, and MCC by 7.2%, 6.42%, 6.83%, and 17.47%, respectively. This improvement underscores XGBoost&#x2019;s superior capability in capturing complex patterns and interactions within the data, making it the most effective model in our study. The findings demonstrate the effectiveness of ensemble methods like XGBoost in outperforming traditional and other ML models, thereby offering enhanced predictive accuracy and reliability.</p><p><xref ref-type="table" rid="table3">Table 3</xref> presents the performance of various ML models on the multiclass classification task. XGBoost achieved the highest accuracy of 85.91%, along with precision, recall, and <italic>F</italic><sub>1</sub>-scores exceeding 85%, and a MCC of 0.7895. In contrast, SVM and LR recorded the lowest accuracies of 53.22% and 54.12%, respectively, and underperformed across other metrics as well. RF and neural networks demonstrated strong performance with accuracies of 82.85% and 83.52%, respectively, and consistently higher precision, recall, <italic>F</italic><sub>1</sub>-score, and MCC. These results indicate that XGBoost surpassed the average accuracy, precision, recall, <italic>F</italic><sub>1</sub>-score, and MCC by 20.35%, 20.13%, 20.34%, 20.41%, and 33.07%, respectively.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>Performance comparison of various models for multi-class classification.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Model</td><td align="left" valign="bottom">Accuracy</td><td align="left" valign="bottom">Precision</td><td align="left" valign="bottom">Recall</td><td align="left" valign="bottom">F1</td><td align="left" valign="bottom">Matthew correlation constant</td></tr></thead><tbody><tr><td align="left" valign="top">Logistic regression</td><td align="left" valign="top">0.5412</td><td align="left" valign="top">0.5414</td><td align="left" valign="top">0.5412</td><td align="left" valign="top">0.5386</td><td align="left" valign="top">0.3136</td></tr><tr><td align="left" valign="top">SVM</td><td align="left" valign="top">0.5322</td><td align="left" valign="top">0.5362</td><td align="left" valign="top">0.5322</td><td align="left" valign="top">0.5315</td><td align="left" valign="top">0.2999</td></tr><tr><td align="left" valign="top">XGBoost</td><td align="left" valign="top">0.8591</td><td align="left" valign="top">0.8602</td><td align="left" valign="top">0.8591</td><td align="left" valign="top">0.8581</td><td align="left" valign="top">0.7895</td></tr><tr><td align="left" valign="top">Random forest</td><td align="left" valign="top">0.8285</td><td align="left" valign="top">0.8308</td><td align="left" valign="top">0.8285</td><td align="left" valign="top">0.8276</td><td align="left" valign="top">0.7446</td></tr><tr><td align="left" valign="top">Neural networks</td><td align="left" valign="top">0.8352</td><td align="left" valign="top">0.8396</td><td align="left" valign="top">0.8352</td><td align="left" valign="top">0.8342</td><td align="left" valign="top">0.7554</td></tr></tbody></table></table-wrap></sec><sec id="s3-2"><title>Statistical Significance Testing</title><p>To verify that the superior performance of XGBoost was not due to random chance, we conducted rigorous statistical significance testing, as detailed below. We ran a 10&#x00D7;3 repeated stratified cross-validation protocol (30 folds in total), preserving the original class distribution in every split. Within each fold, the model was fit to the training indices and its binary predictions on the held-out set were scored with the <italic>F</italic><sub>1</sub>-metric, which is appropriate for moderately imbalanced problems. We then applied paired 2-tailed <italic>t</italic> tests between XGB and each rival model, followed by Holm correction to control the family-wise error rate at <italic>&#x03B1;</italic>=.05.</p><p>For binary classification, across the 30 cross-validation folds, XGB achieved the highest mean <italic>F</italic><sub>1</sub>-score (0.851 &#x00B1; 0.034) while its nearest competitor RF averaged 0.832; SVM, LR, and NN trailed at 0.659, 0.723, and 0.687, respectively. Paired significance testing confirmed that these differences are not due to sampling noise: the Holm-adjusted <italic>P</italic> values for XGB versus RF, LR, SVM, and NN were 3.6&#x00D7;10&#x207B;&#x2074;, 2.3&#x00D7;10&#x207B;&#x00B9;&#x2076;, 2.7&#x00D7;10&#x207B;&#x00B9;&#x2077;, and 1.7&#x00D7;10&#x207B;&#x00B2;&#x2070;, all well below 0.05, confirming that the superior performance of XGBoost is not by chance. Hence, XGB&#x2019;s performance advantage is statistically significant against every other baseline in the study, supporting its selection as the primary model for downstream explainability analysis. The same trend was observed for multiclass classification.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>In this section, we critically interpret XAI techniques (SHAP and LIME) to reveal the feature patterns that drive the depression-classification models, compare these explanations across XGBoost and neural-network architectures, and relate the findings to existing clinical evidence. One of the key critiques of ML models is their &#x201C;black-box&#x201D; nature [<xref ref-type="bibr" rid="ref52">52</xref>]. ML models evolved over the last decade to become more complex, resulting in improved performance by virtue of modeling intricate relationships among the input features. However, the increased complexity also meant lower interpretability and explainability powers. In many use cases such as health care, the lack of transparency in decision making is critical, and the reasoning behind a model&#x2019;s outcome is crucial for trust, accountability, and decision-making [<xref ref-type="bibr" rid="ref52">52</xref>]. A variety of techniques is proposed in XAI to enhance the explainability of complex ML models. SHAP is a type of XAI framework derived from the concept of coalition game where a prediction is regarded as the &#x201C;payoff&#x201D; and the worthiness of the feature is assumed as the &#x201C;player&#x201D; [<xref ref-type="bibr" rid="ref53">53</xref>]. It calculates the Shapley values and explains the prediction by showing how the features influenced the output. The importance of each feature that influenced the effectiveness of the ML algorithm is calculated as given in <xref ref-type="disp-formula" rid="E2">Eq. (2)</xref>.</p><disp-formula id="E2"><mml:math id="eqn2"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msub><mml:mi>&#x03D5;</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>f</mml:mi><mml:mo>,</mml:mo><mml:mi>x</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>S</mml:mi><mml:mo>&#x2286;</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mfrac><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>S</mml:mi><mml:mo>|</mml:mo></mml:mrow><mml:mo>!</mml:mo><mml:mo>&#x22C5;</mml:mo><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>N</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mrow><mml:mo>|</mml:mo><mml:mi>S</mml:mi><mml:mo>|</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>!</mml:mo></mml:mrow><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mi>N</mml:mi><mml:mo>|</mml:mo></mml:mrow><mml:mo>!</mml:mo></mml:mrow></mml:mfrac><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mo>&#x222A;</mml:mo><mml:mrow><mml:mo>{</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>}</mml:mo></mml:mrow></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>f</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mi>x</mml:mi><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><p>Where: <inline-formula><mml:math id="ieqn13"><mml:mi>f</mml:mi></mml:math></inline-formula> represents the prediction model, <inline-formula><mml:math id="ieqn14"><mml:mi>x</mml:mi></mml:math></inline-formula> is the input feature vector, <inline-formula><mml:math id="ieqn15"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mstyle></mml:math></inline-formula> is the set of features, <inline-formula><mml:math id="ieqn16"><mml:mi>S</mml:mi></mml:math></inline-formula> is a subset of features excluding the feature <inline-formula><mml:math id="ieqn17"><mml:mi>i</mml:mi></mml:math></inline-formula> that is, feature under XAI analysis, <inline-formula><mml:math id="ieqn18"><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>s</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the feature vector with features in subset <inline-formula><mml:math id="ieqn19"><mml:mi>S</mml:mi></mml:math></inline-formula> replaced by baseline values and <inline-formula><mml:math id="ieqn20"><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the value of feature <inline-formula><mml:math id="ieqn21"><mml:mi>i</mml:mi></mml:math></inline-formula> in the input vector <inline-formula><mml:math id="ieqn22"><mml:mi>x</mml:mi></mml:math></inline-formula>. For XAI analysis, we restrict to the <sc>Xgboost</sc> model only as it was identified as the best performing model.</p><p>LIME is a popular framework in XAI that helps make ML predictions easier to understand. LIME explains a model&#x2019;s predictions by simplifying it in the local context of a specific instance. It does this by slightly altering the input data around the instance and analyzing how these changes affect the model&#x2019;s predictions. This process highlights which features are most influential in the local neighborhood. To determine the importance of each feature, LIME builds a simple, interpretable model (such as a linear regression model) based on weighted data points. These weights are assigned according to how close the altered data points are to the original instance.</p></sec><sec id="s4-2"><title>XAI for Binary Classification</title><p>Two primary SHAP plots (<xref ref-type="fig" rid="figure4">Figures 4</xref> and <xref ref-type="fig" rid="figure5">5</xref>) indicate how each feature affects the model&#x2019;s decision boundary across the entire dataset. <xref ref-type="fig" rid="figure4">Figure 4</xref> shows the average magnitude of SHAP values (mean absolute SHAP) per feature, highlighting the top predictors. <xref ref-type="fig" rid="figure5">Figure 5</xref> depicts a dot summary where each point corresponds to a specific sample&#x2019;s SHAP value. Points to the right (positive SHAP) push the classification toward &#x201C;Depressed,&#x201D; whereas points to the left (negative SHAP) favor &#x201C;Non-Depressed.&#x201D; Colors (red to blue) represent the actual feature value (eg, higher PSD Mean in red and lower PSD Mean in blue).</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Binary classification feature importance.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig04.png"/></fig><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>SHAPE summary plots for binary classification.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig05.png"/></fig><p>From these plots, we observe the following. PSD means consistently have the largest average impact. Larger values of the PSD mean (possibly reflecting a stronger low-frequency power in the activity data) correlate with a higher likelihood of depression. Mean activity level and age follow closely, indicating that demographic factors (eg, older individuals) and overall daily activity levels significantly shape the classification. Activity offset and AutoCorr also stand out, aligning with clinical findings that disrupted or rigid circadian rhythms (late offsets and high autocorrelation) are often associated with depressive behaviors.</p><p>PSD Mean consistently has the largest average impact. Larger values of the PSD mean (possibly reflecting a stronger low-frequency power in the activity data) correlate with a higher likelihood of depression.</p><p>Mean activity level and age follow closely, indicating that demographic factors (eg, older individuals) and overall daily activity levels significantly shape the classification.</p><p>We applied LIME to explain a single test instance in <xref ref-type="fig" rid="figure6">Figure 6</xref>. LIME creates locally linear &#x201C;rules&#x201D; indicating how small perturbations around this instance shift its predicted probability for depression. Features with red bars push the instance toward &#x201C;Depressed,&#x201D; while green bars push it toward &#x201C;Non-Depressed.&#x201D; For example, a condition like &#x201C;Gender&#x2264;1.0&#x201D; (female) or older &#x201C;Age&#x201D; might increase the model&#x2019;s belief in depression if other activity-related features also align. LIME reported numeric weights for each condition, revealing how thresholds in AutoCorr, Entropy, or Skewness shape local decision boundaries. These rules coincide well with the SHAP findings (eg, higher autocorrelation frequently signals depressive risk).</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Local explanation for the non-depressed class.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig06.png"/></fig><p>While SHAP and LIME are theoretically different (SHAP uses game-theoretic Shapley values, and LIME uses local linear approximations), both converge on similar high-impact features, boosting confidence in the model&#x2019;s learned representations. Specifically, PSD mean, mean activity level, and daily rhythm indicators (like activity offset, AutoCorr) repeatedly appear among the strongest predictors.</p></sec><sec id="s4-3"><title>XAI for Multiclass Classification</title><p>We extended our explainable AI approach to a 3-class setting, distinguishing among &#x201C;Normal Case,&#x201D; &#x201C;Mild Depression,&#x201D; and &#x201C;Moderate Depression.&#x201D; SHAP was used for global insights (per-class and overall feature importance), while LIME was used for local (instance-level) explanations. <xref ref-type="fig" rid="figure7">Figure 7</xref> displays the mean absolute SHAP values aggregated in all 3 classes, highlighting key characteristics that consistently influence the predictions of the model. Notably, AutoCorr and age emerge as dominant factors, aligning with clinical observations that older age groups, coupled with repetitive or rigid activity patterns, may exhibit a higher risk of depression.</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Multi-class feature importance.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig07.png"/></fig><p>For a more in-depth analysis of each class, we generated separate SHAP summary plots. Normal (<xref ref-type="fig" rid="figure8">Figure 8A</xref>): Lower AutoCorr typically favors the normal class, suggesting that a more flexible or varied daily rhythm (less repetitive activity) aligns with nondepressed status. Moderate values of age and PSD mean also push predictions away from mild or moderate depression.</p><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>SHAP summary plots for multi-class classification.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig08.png"/></fig><p>Mild depression (<xref ref-type="fig" rid="figure8">Figure 8B</xref>): Mild cases show strong contributions from AutoCorr and transition entropy. In many instances, moderate daily rhythms&#x2014;coupled with subtle changes in the frequency composition of activity (as captured by PSD Mean)&#x2014;tilt the model toward mild depression. Moderate depression (<xref ref-type="fig" rid="figure8">Figure 8C</xref>): This class often manifests higher or more entrenched irregularities. Age, AutoCorr, and an elevated PSD mean appear to push an instance from mild toward moderate severity. Here, demographic and circadian disruptions intertwine more intensively, reflecting a deeper depressive state.</p><p>Across all classes, AutoCorr consistently ranks among the most critical features. However, its specific influence varies. Moderate autocorrelation can lead a case from normal to mild, while higher autocorrelation plus advanced age often indicates moderate depression. These distinctions underscore how the same feature can have class-dependent impacts based on subtle variations in its actual range or interaction with other features.</p><p>To validate SHAP&#x2019;s per-class findings at the instance level, we used LIME for a single test example. <xref ref-type="fig" rid="figure9">Figure 9</xref> focuses on the model&#x2019;s explanation for the &#x201C;Moderate Depression&#x201D; class. LIME produces threshold-based &#x201C;rules&#x201D; (eg, &#x201C;0.74&#x003C;AutoCorr&#x2264;0.78&#x201D;) that locally shift the probability of class membership. Green bars push the model toward selecting &#x201C;Moderate Depression.&#x201D; For instance, an <italic>AutoCorr</italic> range above 0.74 or age exceeding a certain threshold may substantially raise this class probability. Red bars pull the prediction away from &#x201C;Moderate Depression,&#x201D; indicating features or thresholds that conflict with the typical profile for moderate severity (eg, a lower PSD mean or higher Kurtosis).</p><fig position="float" id="figure9"><label>Figure 9.</label><caption><p>LIME explanation for the moderate depression class.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v12i1e72038_fig09.png"/></fig><p>This local view converges with the SHAP observations: circadian-related variables (eg, AutoCorr and Transition entropy) and demographic factors (eg, age) become pivotal. In mild cases, these metrics may appear only partially disrupted, whereas for moderate severity, they consistently register in ranges linked to stronger depressive indications.</p></sec><sec id="s4-4"><title>Comparing Explanations Across XGBoost and Neural Network Models</title><p>Although XGBoost delivered the highest accuracy, we repeated the XAI analysis for artificial-neural-network (ANN) to test whether the explanatory story was model-dependent. For binary classification, both models ranked PSD mean as the dominant driver of the depressed versus nondepressed decision, confirming the feature&#x2019;s robustness. Beyond that, the explanations diverged: XGBoost distributed moderate importance across roughly ten additional variables (eg, mean, age, activity offset, and AutoCorr), whereas the ANN concentrated nearly all remaining attribution on just three features (activity onset, kurtosis, and activity offset). This comparison strengthens interpretability in 2 ways: the cross-model agreement on PSD mean validates it as a consistent physiological marker and the differing secondary patterns reveal architecture-specific sensitivities, offering complementary insights that would be missed if explanations were reported for XGBoost alone.</p><p>For binary classification, both models ranked PSD mean as the dominant driver of the depressed versus nondepressed decision, confirming the feature&#x2019;s robustness. Beyond that, the explanations diverged: XGBoost distributed moderate importance across roughly ten additional variables (eg, mean, age, activity offset, and AutoCorr), whereas the ANN concentrated nearly all remaining attribution on just 3 features (activity onset, kurtosis, activity offset). This comparison strengthens interpretability in 2 ways: (1) the cross-model agreement on PSD mean validates it as a consistent physiological marker; and (2) the differing secondary patterns reveal architecture-specific sensitivities, offering complementary insights that would be missed if explanations were reported for XGBoost alone.</p><p>For multiclass classification, inspection of the mean-|SHAP| values reveals that XGBoost and the neural network rely on partly overlapping but ultimately different information in multiclass classification, and they do so on strikingly dissimilar attribution scales. XGBoost&#x2019;s strongest signals come from temporal-structure and demographic variables&#x2014;AutoCorr carries the largest impact, closely followed by age, with gender and transition entropy also substantial&#x2014;whereas features such as activity offset contribute almost nothing. The ANN tells a nearly opposite story: its 2 most influential cues are the higher-order statistic kurtosis and activity offset, while the variables that dominate XGBoost shrink to the noise floor. Both models nevertheless converge on PSD mean as a relevant marker, which bolsters confidence in that spectral feature&#x2019;s physiological validity even though its weight is modest in XGBoost and comparatively prominent in the ANN. It is also worth noting that the ANN&#x2019;s absolute SHAP magnitudes are roughly 2 orders of magnitude smaller than the tree ensemble&#x2019;s, reflecting the smoother decision surface learned by the network and the conservative estimates produced by KernelExplainer; consequently, comparisons should be made within each model&#x2019;s scale rather than across scales. Taken together, the agreement on PSD mean and the divergence elsewhere show that the 2 architectures extract complementary patterns&#x2014;trees capitalize on heterogeneous temporal-demographic interactions, whereas the network distills the data into distribution-shape metrics&#x2014;providing a richer, cross-validated picture of how depression severity is encoded.</p></sec><sec id="s4-5"><title>Discussion of XAI Findings</title><p>The combined SHAP&#x2013;LIME analysis reveals several consistent themes.</p><sec id="s4-5-1"><title>Circadian Rhythms and Activity Measures</title><p>Our model&#x2019;s emphasis on autocorrelation and spectral activity features matches a large body of research linking disrupted daily rhythms to mood disorders. For example, clinical and experimental evidence shows that blunted or shifted circadian cycles are closely associated with depression [<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref55">55</xref>]. A recent longitudinal study using wearable devices found that circadian phase shifts often precede worsening mood symptoms in depressed patients [<xref ref-type="bibr" rid="ref56">56</xref>]. Furthermore, therapies that realign circadian timing (bright light therapy, social rhythm therapy, etc) reliably alleviate depressive symptoms [<xref ref-type="bibr" rid="ref55">55</xref>], suggesting a causal role for rhythm disruption. That said, mood and circadian processes are bidirectionally linked [<xref ref-type="bibr" rid="ref55">55</xref>,<xref ref-type="bibr" rid="ref57">57</xref>], so altered activity patterns may both influence and reflect depressive states. In short, our findings are consistent with known associations; however, they do not by themselves prove that these rhythm-related features cause depression.</p></sec><sec id="s4-5-2"><title>Age and Demographic Influence</title><p>Age and gender differences in depression risk are well documented [<xref ref-type="bibr" rid="ref57">57</xref>,<xref ref-type="bibr" rid="ref58">58</xref>], and our model captures these patterns. For instance, older adults tend to show earlier and less variable sleep&#x2013;wake timing [<xref ref-type="bibr" rid="ref59">59</xref>], which may interact with late-life depression risk. Epidemiological studies report that women have roughly twice the odds of major depression compared with men [<xref ref-type="bibr" rid="ref58">58</xref>]; our model&#x2019;s gender feature trends are in line with that. However, like other risk factors, age and gender effects are multifaceted and often indirect. Biological aging brings many changes (health, lifestyle, and brain function), and gender differences reflect hormonal, psychosocial, and diagnostic factors. Thus, while age and demographic features correlate with depression, the mechanistic pathways remain speculative. In practice, these variables likely modulate risk in complex, context-dependent ways [<xref ref-type="bibr" rid="ref57">57</xref>], so caution is warranted in attributing causality to them without further study.</p></sec><sec id="s4-5-3"><title>Distinction Between Mild and Moderate Depression</title><p>The multiclass SHAP plots show that certain transitions from mild to moderate often involve heightened daily-activity irregularities. The interplay of transition entropy and AutoCorr is particularly telling, resembling findings in psychological studies that link more chaotic sleep&#x2013;wake transitions to deeper depressive states [<xref ref-type="bibr" rid="ref60">60</xref>]. Nevertheless, these are still observational correlations. For example, it is unclear whether irregular sleep drives depression or mainly reflects its impact on physiology. Disentangling this requires targeted studies (eg, manipulating sleep schedules) beyond what SHAP and LIME alone can show.</p></sec><sec id="s4-5-4"><title>Local Explanation Value</title><p>Per-instance force plots or LIME rules allow for clinically relevant breakdowns of risk factors, highlighting how subtle changes in demographic or circadian metrics can shift a patient&#x2019;s classification from &#x201C;Normal&#x201D; to &#x201C;Mild&#x201D; or &#x201C;Moderate Depression.&#x201D; This local interpretability aligns with the literature&#x2019;s emphasis on personalized medicine and the necessity of context-aware diagnostics [<xref ref-type="bibr" rid="ref61">61</xref>].</p><p>In general, the XAI approach confirms that data-driven detection of depression is highly dependent on circadian and demographic characteristics&#x2014;echoing established work on the connection between daily behavioral patterns and mental health status. By offering both a holistic (global) and case-specific (local) view, SHAP and LIME collectively bolster the transparency and reliability of ML in mental health applications. Note that although these associations are promising, confirming true causality will require targeted longitudinal and interventional studies beyond the correlations revealed by SHAP and LIME.</p><p>In this study, we presented a ML framework augmented with explainable AI for the detection and severity-level classification of depression based on wearable-actigraphy data. We demonstrated how XGBoost, coupled with ADASYN and feature selection, consistently achieved superior performance compared to other methods. The SHAP and LIME explanations revealed that circadian-related and demographic factors (eg, PSD mean, age, and AutoCorr) play central roles in identifying depressive states. By providing both global and local interpretability, the proposed system can help clinicians and researchers understand and trust the model&#x2019;s decisions, supporting more targeted treatment plans and facilitating broader acceptance of AI-based tools in mental health care.</p></sec></sec></sec></body><back><notes><sec><title>Data Availability</title><p>The dataset is available for download from The Depresjon Dataset [<xref ref-type="bibr" rid="ref62">62</xref>].</p></sec></notes><fn-group><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">ADASYN</term><def><p>Adaptive Synthetic Sampling Technique</p></def></def-item><def-item><term id="abb2">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb3">ANN</term><def><p>artificial-neural-network</p></def></def-item><def-item><term id="abb4">CNN</term><def><p>convolutional neural networks</p></def></def-item><def-item><term id="abb5">DNN</term><def><p>deep neural network</p></def></def-item><def-item><term id="abb6">FN</term><def><p>false negative</p></def></def-item><def-item><term id="abb7">FP</term><def><p>false positive</p></def></def-item><def-item><term id="abb8">LIME</term><def><p>Interpretable Model-agnostic Explanations</p></def></def-item><def-item><term id="abb9">LR</term><def><p>logistic regression</p></def></def-item><def-item><term id="abb10">MADRS</term><def><p>Montgomery-Asberg Depression Rating Scale</p></def></def-item><def-item><term id="abb11">MCC</term><def><p>Mathew correlation constant</p></def></def-item><def-item><term id="abb12">ML</term><def><p>machine learning</p></def></def-item><def-item><term id="abb13">MSE</term><def><p>mean square error</p></def></def-item><def-item><term id="abb14">NN</term><def><p>neural networks</p></def></def-item><def-item><term id="abb15">OvO</term><def><p>One-vs-One</p></def></def-item><def-item><term id="abb16">OvR</term><def><p>One-vs-Rest</p></def></def-item><def-item><term id="abb17">PSD</term><def><p>power spectral density</p></def></def-item><def-item><term id="abb18">RF</term><def><p>random forest</p></def></def-item><def-item><term id="abb19">SHAP</term><def><p>Shapley Additive Explanations</p></def></def-item><def-item><term id="abb20">SMOTE</term><def><p>Synthetic Minority Oversampling Technique</p></def></def-item><def-item><term id="abb21">SVM</term><def><p>support vector machines</p></def></def-item><def-item><term id="abb22">TN</term><def><p>true negative</p></def></def-item><def-item><term id="abb23">TP</term><def><p>true positive</p></def></def-item><def-item><term id="abb24">XAI</term><def><p>explainable artificial intelligence</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Santomauro</surname><given-names>DF</given-names> </name><name name-style="western"><surname>Mantilla Herrera</surname><given-names>AM</given-names> </name><name name-style="western"><surname>Shadid</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Global prevalence and burden of depressive and anxiety disorders in 204 countries and territories in 2020 due to the COVID-19 pandemic</article-title><source>The Lancet</source><year>2021</year><month>11</month><volume>398</volume><issue>10312</issue><fpage>1700</fpage><lpage>1712</lpage><pub-id pub-id-type="doi">10.1016/S0140-6736(21)02143-7</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gulland</surname><given-names>A</given-names> </name></person-group><article-title>Women have higher rates of mental disorders than men, NHS survey finds</article-title><source>BMJ</source><year>2016</year><fpage>i5320</fpage><pub-id pub-id-type="doi">10.1136/bmj.i5320</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Varma</surname><given-names>P</given-names> </name><name name-style="western"><surname>Junge</surname><given-names>M</given-names> </name><name name-style="western"><surname>Meaklim</surname><given-names>H</given-names> </name><name name-style="western"><surname>Jackson</surname><given-names>ML</given-names> </name></person-group><article-title>Younger people are more vulnerable to stress, anxiety and depression during COVID-19 pandemic: a global cross-sectional survey</article-title><source>Prog Neuropsychopharmacol Biol Psychiatry</source><year>2021</year><month>07</month><day>13</day><volume>109</volume><fpage>110236</fpage><pub-id pub-id-type="doi">10.1016/j.pnpbp.2020.110236</pub-id><pub-id pub-id-type="medline">33373680</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="web"><article-title>World health organization</article-title><source>Depressive disorder (depression) - who.int</source><year>2023</year><access-date>2025-09-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.who.int/news-room/fact-sheets/detail/depression">https://www.who.int/news-room/fact-sheets/detail/depression</ext-link></comment></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="web"><article-title>World health organization</article-title><source>Suicide - India - who.int</source><access-date>2025-09-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.who.int/india/health-topics/suicide">https://www.who.int/india/health-topics/suicide</ext-link></comment></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Chand</surname><given-names>SP</given-names> </name><name name-style="western"><surname>Arif</surname><given-names>H</given-names> </name></person-group><source>Depression</source><year>2024</year><access-date>2025-09-04</access-date><publisher-name>StatPearls Publishing</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://www.ncbi.nlm.nih.gov/books/NBK430847/">https://www.ncbi.nlm.nih.gov/books/NBK430847/</ext-link></comment></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tolentino</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Schmidt</surname><given-names>SL</given-names> </name></person-group><article-title>DSM-5 criteria and depression severity: implications for clinical practice</article-title><source>Front Psychiatry</source><year>2018</year><volume>9</volume><fpage>450</fpage><pub-id pub-id-type="doi">10.3389/fpsyt.2018.00450</pub-id><pub-id pub-id-type="medline">30333763</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>McKee</surname><given-names>J</given-names> </name><name name-style="western"><surname>Brahm</surname><given-names>N</given-names> </name></person-group><article-title>Medical mimics: differential diagnostic considerations for psychiatric symptoms</article-title><source>Ment Health Clin</source><year>2016</year><month>11</month><volume>6</volume><issue>6</issue><fpage>289</fpage><lpage>296</lpage><pub-id pub-id-type="doi">10.9740/mhc.2016.11.289</pub-id><pub-id pub-id-type="medline">29955484</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Currier</surname><given-names>MB</given-names> </name><name name-style="western"><surname>Nemeroff</surname><given-names>CB</given-names> </name></person-group><article-title>Depression as a risk factor for cancer: from pathophysiological advances to treatment implications</article-title><source>Annu Rev Med</source><year>2014</year><volume>65</volume><issue>1</issue><fpage>203</fpage><lpage>221</lpage><pub-id pub-id-type="doi">10.1146/annurev-med-061212-171507</pub-id><pub-id pub-id-type="medline">24215332</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>H</given-names> </name><name name-style="western"><surname>Luiten</surname><given-names>PGM</given-names> </name><name name-style="western"><surname>Eisel</surname><given-names>ULM</given-names> </name><name name-style="western"><surname>Dejongste</surname><given-names>MJL</given-names> </name><name name-style="western"><surname>Schoemaker</surname><given-names>RG</given-names> </name></person-group><article-title>Depression after myocardial infarction: TNF-&#x03B1;-induced alterations of the blood&#x2013;brain barrier and its putative therapeutic implications</article-title><source>Neuroscience &#x0026; Biobehavioral Reviews</source><year>2013</year><month>05</month><volume>37</volume><issue>4</issue><fpage>561</fpage><lpage>572</lpage><pub-id pub-id-type="doi">10.1016/j.neubiorev.2013.02.004</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kohn</surname><given-names>R</given-names> </name><name name-style="western"><surname>Saxena</surname><given-names>S</given-names> </name><name name-style="western"><surname>Levav</surname><given-names>I</given-names> </name><name name-style="western"><surname>Saraceno</surname><given-names>B</given-names> </name></person-group><article-title>The treatment gap in mental health care</article-title><source>Bull World Health Organ</source><year>2004</year><month>11</month><volume>82</volume><issue>11</issue><fpage>858</fpage><lpage>866</lpage><pub-id pub-id-type="medline">15640922</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sharp</surname><given-names>LK</given-names> </name><name name-style="western"><surname>Lipsky</surname><given-names>MS</given-names> </name></person-group><article-title>Screening for depression across the lifespan: a review of measures for use in primary care settings</article-title><source>Am Fam Physician</source><year>2002</year><month>09</month><day>15</day><volume>66</volume><issue>6</issue><fpage>1001</fpage><lpage>1008</lpage><pub-id pub-id-type="medline">12358212</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Montgomery</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Asberg</surname><given-names>M</given-names> </name></person-group><article-title>A new depression scale designed to be sensitive to change</article-title><source>Br J Psychiatry</source><year>1979</year><month>04</month><volume>134</volume><issue>4</issue><fpage>382</fpage><lpage>389</lpage><pub-id pub-id-type="doi">10.1192/bjp.134.4.382</pub-id><pub-id pub-id-type="medline">444788</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Quilty</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Robinson</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Rolland</surname><given-names>JP</given-names> </name><name name-style="western"><surname>Fruyt</surname><given-names>FD</given-names> </name><name name-style="western"><surname>Rouillon</surname><given-names>F</given-names> </name><name name-style="western"><surname>Bagby</surname><given-names>RM</given-names> </name></person-group><article-title>The structure of the Montgomery&#x2013;&#x00C5;sberg depression rating scale over the course of treatment for depression</article-title><source>Int J Methods Psych Res</source><year>2013</year><month>09</month><volume>22</volume><issue>3</issue><fpage>175</fpage><lpage>184</lpage><pub-id pub-id-type="doi">10.1002/mpr.1388</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Herrmann</surname><given-names>N</given-names> </name><name name-style="western"><surname>Black</surname><given-names>SE</given-names> </name><name name-style="western"><surname>Lawrence</surname><given-names>J</given-names> </name><name name-style="western"><surname>Szekely</surname><given-names>C</given-names> </name><name name-style="western"><surname>Szalai</surname><given-names>JP</given-names> </name></person-group><article-title>The Sunnybrook Stroke Study: a prospective study of depressive symptoms and functional outcome</article-title><source>Stroke</source><year>1998</year><month>03</month><volume>29</volume><issue>3</issue><fpage>618</fpage><lpage>624</lpage><pub-id pub-id-type="doi">10.1161/01.str.29.3.618</pub-id><pub-id pub-id-type="medline">9506602</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Remes</surname><given-names>O</given-names> </name><name name-style="western"><surname>Mendes</surname><given-names>JF</given-names> </name><name name-style="western"><surname>Templeton</surname><given-names>P</given-names> </name></person-group><article-title>Biological, psychological, and social determinants of depression: a review of recent literature</article-title><source>Brain Sci</source><year>2021</year><month>12</month><day>10</day><volume>11</volume><issue>12</issue><fpage>1633</fpage><pub-id pub-id-type="doi">10.3390/brainsci11121633</pub-id><pub-id pub-id-type="medline">34942936</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Smith</surname><given-names>KM</given-names> </name><name name-style="western"><surname>Renshaw</surname><given-names>PF</given-names> </name><name name-style="western"><surname>Bilello</surname><given-names>J</given-names> </name></person-group><article-title>The diagnosis of depression: current and emerging methods</article-title><source>Compr Psychiatry</source><year>2013</year><month>01</month><volume>54</volume><issue>1</issue><fpage>1</fpage><lpage>6</lpage><pub-id pub-id-type="doi">10.1016/j.comppsych.2012.06.006</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Richter</surname><given-names>T</given-names> </name><name name-style="western"><surname>Fishbain</surname><given-names>B</given-names> </name><name name-style="western"><surname>Markus</surname><given-names>A</given-names> </name><name name-style="western"><surname>Richter-Levin</surname><given-names>G</given-names> </name><name name-style="western"><surname>Okon-Singer</surname><given-names>H</given-names> </name></person-group><article-title>Using machine learning-based analysis for behavioral differentiation between anxiety and depression</article-title><source>Sci Rep</source><year>2020</year><volume>10</volume><issue>1</issue><fpage>16381</fpage><pub-id pub-id-type="doi">10.1038/s41598-020-72289-9</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Muhammad</surname><given-names>D</given-names> </name><name name-style="western"><surname>Ahmed</surname><given-names>I</given-names> </name><name name-style="western"><surname>Ahmad</surname><given-names>MO</given-names> </name><name name-style="western"><surname>Bendechache</surname><given-names>M</given-names> </name></person-group><article-title>Randomized explainable machine learning models for efficient medical diagnosis</article-title><source>IEEE J Biomed Health Inform</source><year>2024</year><month>11</month><day>13</day><volume>PP</volume><fpage>1</fpage><lpage>10</lpage><pub-id pub-id-type="doi">10.1109/JBHI.2024.3491593</pub-id><pub-id pub-id-type="medline">40030196</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title><source>Nat Med</source><year>2019</year><month>01</month><volume>25</volume><issue>1</issue><fpage>44</fpage><lpage>56</lpage><pub-id pub-id-type="doi">10.1038/s41591-018-0300-7</pub-id><pub-id pub-id-type="medline">30617339</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abd-Alrazaq</surname><given-names>A</given-names> </name><name name-style="western"><surname>AlSaad</surname><given-names>R</given-names> </name><name name-style="western"><surname>Aziz</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Wearable artificial intelligence for anxiety and depression: scoping review</article-title><source>J Med Internet Res</source><year>2023</year><month>01</month><day>19</day><volume>25</volume><fpage>e42672</fpage><pub-id pub-id-type="doi">10.2196/42672</pub-id><pub-id pub-id-type="medline">36656625</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Paul</surname><given-names>N</given-names> </name><name name-style="western"><surname>Juliet</surname><given-names>S</given-names> </name></person-group><article-title>Comparative analysis of machine learning techniques for mental health prediction</article-title><year>2023</year><conf-name>2023 8th International Conference on Communication and Electronics Systems (ICCES)</conf-name><conf-loc>Coimbatore, India</conf-loc><pub-id pub-id-type="doi">10.1109/ICCES57224.2023.10192763</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>J</given-names> </name><name name-style="western"><surname>Richardson</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Dunkley</surname><given-names>BT</given-names> </name></person-group><article-title>Classifying post-traumatic stress disorder using the magnetoencephalographic connectome and machine learning</article-title><source>Sci Rep</source><year>2020</year><volume>10</volume><issue>1</issue><fpage>5937</fpage><pub-id pub-id-type="doi">10.1038/s41598-020-62713-5</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mikolas</surname><given-names>P</given-names> </name><name name-style="western"><surname>Vahid</surname><given-names>A</given-names> </name><name name-style="western"><surname>Bernardoni</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Training a machine learning classifier to identify ADHD based on real-world clinical data from medical records</article-title><source>Sci Rep</source><year>2022</year><month>07</month><day>28</day><volume>12</volume><issue>1</issue><fpage>12934</fpage><pub-id pub-id-type="doi">10.1038/s41598-022-17126-x</pub-id><pub-id pub-id-type="medline">35902654</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Spulber</surname><given-names>S</given-names> </name><name name-style="western"><surname>Elberling</surname><given-names>F</given-names> </name><name name-style="western"><surname>Svensson</surname><given-names>J</given-names> </name><name name-style="western"><surname>Tiger</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ceccatelli</surname><given-names>S</given-names> </name><name name-style="western"><surname>Lundberg</surname><given-names>J</given-names> </name></person-group><article-title>Patterns of activity correlate with symptom severity in major depressive disorder patients</article-title><source>Transl Psychiatry</source><year>2022</year><month>06</month><day>2</day><volume>12</volume><issue>1</issue><fpage>226</fpage><pub-id pub-id-type="doi">10.1038/s41398-022-01989-9</pub-id><pub-id pub-id-type="medline">35654778</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Garcia-Ceja</surname><given-names>E</given-names> </name><name name-style="western"><surname>Riegler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jakobsen</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Motor activity based classification of depression in unipolar and bipolar patients</article-title><year>2018</year><conf-name>2018 IEEE 31st International Symposium on Computer-Based Medical Systems (CBMS)</conf-name><conf-loc>Karlstad</conf-loc><pub-id pub-id-type="doi">10.1109/CBMS.2018.00062</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Frogner</surname><given-names>JI</given-names> </name><name name-style="western"><surname>Noori</surname><given-names>FM</given-names> </name><name name-style="western"><surname>Halvorsen</surname><given-names>P</given-names> </name><etal/></person-group><article-title>One-dimensional convolutional neural networks on motor activity measurements in detection of depression</article-title><year>2019</year><month>10</month><day>15</day><conf-name>MM &#x2019;19</conf-name><conf-loc>Nice France</conf-loc><fpage>9</fpage><lpage>15</lpage><comment><ext-link ext-link-type="uri" xlink:href="https://dl.acm.org/doi/proceedings/10.1145/3347444">https://dl.acm.org/doi/proceedings/10.1145/3347444</ext-link></comment><pub-id pub-id-type="doi">10.1145/3347444.3356238</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Jakobsen</surname><given-names>P</given-names> </name><name name-style="western"><surname>Garcia-Ceja</surname><given-names>E</given-names> </name><name name-style="western"><surname>Riegler</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Applying machine learning in motor activity time series of depressed bipolar and unipolar patients compared to healthy controls</article-title><source>PLoS ONE</source><year>2020</year><volume>15</volume><issue>8</issue><fpage>e0231995</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0231995</pub-id><pub-id pub-id-type="medline">32833958</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Espino-Salinas</surname><given-names>CH</given-names> </name><name name-style="western"><surname>Galv&#x00E1;n-Tejada</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Luna-Garc&#x00ED;a</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Two-dimensional convolutional neural network for depression episodes detection in real time using motor activity time series of depresjon dataset</article-title><source>Bioengineering (Basel)</source><year>2022</year><month>09</month><day>9</day><volume>9</volume><issue>9</issue><fpage>458</fpage><pub-id pub-id-type="doi">10.3390/bioengineering9090458</pub-id><pub-id pub-id-type="medline">36135004</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Y</given-names> </name></person-group><article-title>Application of machine learning to predict mental health disorders and interpret feature importance</article-title><year>2023</year><conf-name>2023 3rd International Symposium on Computer Technology and Information Science (ISCTIS)</conf-name><fpage>257</fpage><lpage>261</lpage><pub-id pub-id-type="doi">10.1109/ISCTIS58954.2023.10213032</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Malik</surname><given-names>A</given-names> </name><name name-style="western"><surname>Shabaz</surname><given-names>M</given-names> </name><name name-style="western"><surname>Asenso</surname><given-names>E</given-names> </name></person-group><article-title>Machine learning based model for detecting depression during Covid-19 crisis</article-title><source>Sci Afr</source><year>2023</year><month>07</month><volume>20</volume><fpage>e01716</fpage><pub-id pub-id-type="doi">10.1016/j.sciaf.2023.e01716</pub-id><pub-id pub-id-type="medline">37214195</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Handing</surname><given-names>EP</given-names> </name><name name-style="western"><surname>Strobl</surname><given-names>C</given-names> </name><name name-style="western"><surname>Jiao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Feliciano</surname><given-names>L</given-names> </name><name name-style="western"><surname>Aichele</surname><given-names>S</given-names> </name></person-group><article-title>Predictors of depression among middle-aged and older men and women in Europe: a machine learning approach</article-title><source>Lancet Reg Health Eur</source><year>2022</year><month>07</month><volume>18</volume><fpage>100391</fpage><pub-id pub-id-type="doi">10.1016/j.lanepe.2022.100391</pub-id><pub-id pub-id-type="medline">35519235</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Haque</surname><given-names>UM</given-names> </name><name name-style="western"><surname>Kabir</surname><given-names>E</given-names> </name><name name-style="western"><surname>Khanam</surname><given-names>R</given-names> </name></person-group><article-title>Detection of child depression using machine learning methods</article-title><source>PLoS ONE</source><year>2021</year><volume>16</volume><issue>12</issue><fpage>e0261131</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0261131</pub-id><pub-id pub-id-type="medline">34914728</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Pacheco-Gonz&#x00E1;lez</surname><given-names>SL</given-names> </name><name name-style="western"><surname>Zanella-Calzada</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Galv&#x00E1;n-Tejada</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Ch&#x00E1;vez-Lamas</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Rivera-G&#x00F3;mez</surname><given-names>JF</given-names> </name><name name-style="western"><surname>Galv&#x00E1;n-Tejada</surname><given-names>JI</given-names> </name></person-group><article-title>Evaluation of five classifiers for depression episodes detection</article-title><source>RCS</source><year>2019</year><volume>148</volume><issue>10</issue><fpage>129</fpage><lpage>138</lpage><pub-id pub-id-type="doi">10.13053/rcs-148-10-11</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Galv&#x00E1;n-Tejada</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Zanella-Calzada</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Gamboa-Rosales</surname><given-names>H</given-names> </name><etal/></person-group><article-title>Depression episodes detection in unipolar and bipolar patients: a methodology with feature extraction and feature selection with genetic algorithms using activity motion signal as information source</article-title><source>Mobile Information Systems</source><year>2019</year><month>04</month><day>23</day><volume>2019</volume><issue>1</issue><fpage>1</fpage><lpage>12</lpage><pub-id pub-id-type="doi">10.1155/2019/8269695</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Garcia-Ceja</surname><given-names>E</given-names> </name><name name-style="western"><surname>Riegler</surname><given-names>M</given-names> </name><name name-style="western"><surname>Jakobsen</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Depresjon: a motor activity database of depression episodes in unipolar and bipolar patients</article-title><year>2018</year><conf-name>Multimedia System Conference 2018</conf-name><conf-loc>Amsterdam</conf-loc><pub-id pub-id-type="doi">10.1145/3204949.3208125</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Garcia</surname><given-names>EA</given-names> </name><name name-style="western"><surname>He</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li S.</surname><given-names>A</given-names> </name></person-group><article-title>ADASYN: Adaptive synthetic sampling approach for imbalanced learning</article-title><year>2008</year><conf-name>2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence)</conf-name><conf-loc>Hong Kong</conf-loc><fpage>1322</fpage><lpage>1328</lpage><pub-id pub-id-type="doi">10.1109/IJCNN.2008.4633969</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zanella-Calzada</surname><given-names>LA</given-names> </name><name name-style="western"><surname>Galv&#x00E1;n-Tejada</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Ch&#x00E1;vez-Lamas</surname><given-names>NM</given-names> </name><etal/></person-group><article-title>Feature extraction in motor activity signal: towards a depression episodes detection in unipolar and bipolar patients</article-title><source>Diagnostics (Basel)</source><year>2019</year><month>01</month><day>10</day><volume>9</volume><issue>1</issue><fpage>8</fpage><pub-id pub-id-type="doi">10.3390/diagnostics9010008</pub-id><pub-id pub-id-type="medline">30634621</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Macz&#x00E1;k</surname><given-names>B</given-names> </name><name name-style="western"><surname>Gingl</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Vadai</surname><given-names>G</given-names> </name></person-group><article-title>General spectral characteristics of human activity and its inherent scale-free fluctuations</article-title><source>Sci Rep</source><year>2024</year><month>01</month><day>31</day><volume>14</volume><issue>1</issue><fpage>2604</fpage><pub-id pub-id-type="doi">10.1038/s41598-024-52905-8</pub-id><pub-id pub-id-type="medline">38297022</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Wilson</surname><given-names>CE</given-names> </name><name name-style="western"><surname>Crouse</surname><given-names>JJ</given-names> </name><name name-style="western"><surname>Carpenter</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Hickie</surname><given-names>IB</given-names> </name></person-group><person-group person-group-type="editor"><name name-style="western"><surname>Kushida</surname><given-names>CA</given-names> </name></person-group><article-title>The circadian component of mood disorders: the sleep-wake cycle, biological rhythms, and chronotherapeutics</article-title><source>Encyclopedia of Sleep and Circadian Rhythms</source><year>2023</year><edition>2</edition><publisher-name>Academic Press</publisher-name><fpage>766</fpage><lpage>779</lpage><pub-id pub-id-type="doi">10.1016/B978-0-12-822963-7.00147-X</pub-id><pub-id pub-id-type="other">978-0-323-91094-1</pub-id></nlm-citation></ref><ref id="ref41"><label>41</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Langensiepen</surname><given-names>C</given-names> </name><name name-style="western"><surname>Lotfi</surname><given-names>A</given-names> </name></person-group><article-title>Uncertainty measures in an ambient intelligence environment</article-title><year>2017</year><conf-name>2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE)</conf-name><fpage>1</fpage><lpage>5</lpage><pub-id pub-id-type="doi">10.1109/FUZZ-IEEE.2017.8015765</pub-id></nlm-citation></ref><ref id="ref42"><label>42</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Hosmer Jr</surname><given-names>DW</given-names> </name><name name-style="western"><surname>Lemeshow</surname><given-names>S</given-names> </name><name name-style="western"><surname>Sturdivant</surname><given-names>RX</given-names> </name></person-group><source>Applied Logistic Regression</source><year>2013</year><publisher-name>John Wiley &#x0026; Sons</publisher-name><pub-id pub-id-type="doi">10.1002/9781118548387</pub-id></nlm-citation></ref><ref id="ref43"><label>43</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Do</surname><given-names>TN</given-names> </name><name name-style="western"><surname>Poulet</surname><given-names>F</given-names> </name></person-group><article-title>Parallel multiclass logistic regression for classifying large scale image datasets</article-title><source>Advanced Computational Methods for Knowledge Engineering Advances in Intelligent Systems and Computing</source><year>2015</year><publisher-name>Springer</publisher-name><fpage>255</fpage><lpage>266</lpage><pub-id pub-id-type="doi">10.1007/978-3-319-17996-4_23</pub-id></nlm-citation></ref><ref id="ref44"><label>44</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chandra</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Bedi</surname><given-names>SS</given-names> </name></person-group><article-title>Survey on SVM and their application in image classification</article-title><source>Int j inf tecnol</source><year>2021</year><month>10</month><volume>13</volume><issue>5</issue><fpage>1</fpage><lpage>11</lpage><pub-id pub-id-type="doi">10.1007/s41870-017-0080-1</pub-id></nlm-citation></ref><ref id="ref45"><label>45</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Breiman</surname><given-names>L</given-names> </name></person-group><article-title>Random forests</article-title><source>Machine Learning</source><year>2001</year><volume>45</volume><publisher-name>Springer</publisher-name><fpage>5</fpage><lpage>32</lpage><pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id></nlm-citation></ref><ref id="ref46"><label>46</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Cutler</surname><given-names>A</given-names> </name><name name-style="western"><surname>Cutler</surname><given-names>DR</given-names> </name><name name-style="western"><surname>Stevens</surname><given-names>JR</given-names> </name></person-group><article-title>Random forests</article-title><source>Ensemble Machine Learning: Methods and Applications</source><year>2012</year><publisher-name>Springer</publisher-name><fpage>157</fpage><lpage>175</lpage><pub-id pub-id-type="doi">10.1007/978-1-4419-9326-7_5</pub-id></nlm-citation></ref><ref id="ref47"><label>47</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>T</given-names> </name><name name-style="western"><surname>Guestrin</surname><given-names>C</given-names> </name></person-group><article-title>Xgboost: a scalable tree boosting system</article-title><source>arXiv</source><comment>Preprint posted online on  Mar 19, 2016</comment><pub-id pub-id-type="doi">10.48550/arXiv.1603.02754</pub-id></nlm-citation></ref><ref id="ref48"><label>48</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Rahman</surname><given-names>RU</given-names> </name><name name-style="western"><surname>Ahmad</surname><given-names>I</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Sparks</surname><given-names>R</given-names> </name><name name-style="western"><surname>Saad</surname><given-names>AB</given-names> </name><name name-style="western"><surname>Mullen</surname><given-names>A</given-names> </name></person-group><article-title>Singletrome: a method to analyze and enhance the transcriptome with long noncoding rnas for single cell analysis</article-title><source>bioRxiv</source><year>2022</year><month>11</month><day>2</day><pub-id pub-id-type="doi">10.1101/2022.10.31.514182v1</pub-id></nlm-citation></ref><ref id="ref49"><label>49</label><nlm-citation citation-type="book"><person-group person-group-type="author"><name name-style="western"><surname>Goodfellow</surname><given-names>I</given-names> </name><name name-style="western"><surname>Bengio</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Courville</surname><given-names>A</given-names> </name></person-group><source>Deep Learning</source><year>2016</year><publisher-name>The MIT Press</publisher-name><pub-id pub-id-type="other">9780262035613</pub-id></nlm-citation></ref><ref id="ref50"><label>50</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Misgar</surname><given-names>MM</given-names> </name><name name-style="western"><surname>Bhatia</surname><given-names>MPS</given-names> </name></person-group><article-title>Detection of depression from iomt time series data using UMAP features</article-title><year>2022</year><conf-name>2022 International Conference on Computing, Communication, and Intelligent Systems (ICCCIS)</conf-name><fpage>623</fpage><lpage>628</lpage><pub-id pub-id-type="doi">10.1109/ICCCIS56430.2022.10037751</pub-id></nlm-citation></ref><ref id="ref51"><label>51</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Ahmad</surname><given-names>I</given-names> </name><name name-style="western"><surname>A. Alqarni</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ali Almazroi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Tariq</surname><given-names>A</given-names> </name></person-group><article-title>Experimental evaluation of clickbait detection using machine learning models</article-title><source>Intelligent Automation &#x0026; Soft Computing</source><year>2020</year><volume>26</volume><issue>4</issue><fpage>1335</fpage><lpage>1344</lpage><pub-id pub-id-type="doi">10.32604/iasc.2020.013861</pub-id></nlm-citation></ref><ref id="ref52"><label>52</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Muhammad</surname><given-names>D</given-names> </name><name name-style="western"><surname>Bendechache</surname><given-names>M</given-names> </name></person-group><article-title>Unveiling the black box: a systematic review of explainable artificial intelligence in medical image analysis</article-title><source>Comput Struct Biotechnol J</source><year>2024</year><month>12</month><volume>24</volume><fpage>542</fpage><lpage>560</lpage><pub-id pub-id-type="doi">10.1016/j.csbj.2024.08.005</pub-id><pub-id pub-id-type="medline">39252818</pub-id></nlm-citation></ref><ref id="ref53"><label>53</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Su-In</surname><given-names>L</given-names> </name><name name-style="western"><surname>Scott</surname><given-names>M</given-names> </name></person-group><article-title>A unified approach to interpreting model predictions</article-title><source>arXiv</source><comment>Preprint posted online on  May 22, 2017</comment><pub-id pub-id-type="doi">10.48550/arXiv.1705.07874</pub-id></nlm-citation></ref><ref id="ref54"><label>54</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yang</surname><given-names>HJ</given-names> </name><name name-style="western"><surname>Cheng</surname><given-names>WJ</given-names> </name><name name-style="western"><surname>Hsiao</surname><given-names>MC</given-names> </name><etal/></person-group><article-title>Rest-activity rhythm associated with depressive symptom severity and attention among patients with major depressive disorder: a 12-month follow-up study</article-title><source>Front Psychiatry</source><year>2023</year><volume>14</volume><fpage>1214143</fpage><pub-id pub-id-type="doi">10.3389/fpsyt.2023.1214143</pub-id><pub-id pub-id-type="medline">37663595</pub-id></nlm-citation></ref><ref id="ref55"><label>55</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Walker</surname><given-names>WH</given-names>  <suffix>2nd</suffix></name><name name-style="western"><surname>Walton</surname><given-names>JC</given-names> </name><name name-style="western"><surname>DeVries</surname><given-names>AC</given-names> </name><name name-style="western"><surname>Nelson</surname><given-names>RJ</given-names> </name></person-group><article-title>Circadian rhythm disruption and mental health</article-title><source>Transl Psychiatry</source><year>2020</year><month>01</month><day>23</day><volume>10</volume><issue>1</issue><fpage>28</fpage><pub-id pub-id-type="doi">10.1038/s41398-020-0694-0</pub-id><pub-id pub-id-type="medline">32066704</pub-id></nlm-citation></ref><ref id="ref56"><label>56</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Song</surname><given-names>YM</given-names> </name><name name-style="western"><surname>Jeong</surname><given-names>J</given-names> </name><name name-style="western"><surname>de Los Reyes</surname><given-names>AA</given-names>  <suffix>5th</suffix></name><etal/></person-group><article-title>Causal dynamics of sleep, circadian rhythm, and mood symptoms in patients with major depression and bipolar disorder: insights from longitudinal wearable device data</article-title><source>EBioMedicine</source><year>2024</year><month>05</month><volume>103</volume><fpage>105094</fpage><pub-id pub-id-type="doi">10.1016/j.ebiom.2024.105094</pub-id><pub-id pub-id-type="medline">38579366</pub-id></nlm-citation></ref><ref id="ref57"><label>57</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Mimikou</surname><given-names>C</given-names> </name><name name-style="western"><surname>Kokkotis</surname><given-names>C</given-names> </name><name name-style="western"><surname>Tsiptsios</surname><given-names>D</given-names> </name><etal/></person-group><article-title>Explainable machine learning in the prediction of depression</article-title><source>Diagnostics (Basel)</source><year>2025</year><month>06</month><day>2</day><volume>15</volume><issue>11</issue><fpage>1412</fpage><pub-id pub-id-type="doi">10.3390/diagnostics15111412</pub-id><pub-id pub-id-type="medline">40506984</pub-id></nlm-citation></ref><ref id="ref58"><label>58</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Salk</surname><given-names>RH</given-names> </name><name name-style="western"><surname>Hyde</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Abramson</surname><given-names>LY</given-names> </name></person-group><article-title>Gender differences in depression in representative national samples: meta-analyses of diagnoses and symptoms</article-title><source>Psychol Bull</source><year>2017</year><month>08</month><volume>143</volume><issue>8</issue><fpage>783</fpage><lpage>822</lpage><pub-id pub-id-type="doi">10.1037/bul0000102</pub-id><pub-id pub-id-type="medline">28447828</pub-id></nlm-citation></ref><ref id="ref59"><label>59</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Maglione</surname><given-names>JE</given-names> </name><name name-style="western"><surname>Ancoli-Israel</surname><given-names>S</given-names> </name><name name-style="western"><surname>Peters</surname><given-names>KW</given-names> </name><etal/></person-group><article-title>Depressive symptoms and circadian activity rhythm disturbances in community-dwelling older women</article-title><source>Am J Geriatr Psychiatry</source><year>2014</year><month>04</month><volume>22</volume><issue>4</issue><fpage>349</fpage><lpage>361</lpage><pub-id pub-id-type="doi">10.1016/j.jagp.2012.09.003</pub-id><pub-id pub-id-type="medline">23567424</pub-id></nlm-citation></ref><ref id="ref60"><label>60</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wichers</surname><given-names>M</given-names> </name><name name-style="western"><surname>Groot</surname><given-names>PC</given-names> </name><collab>Psychosystems, ESM Group, EWS Group</collab></person-group><article-title>Critical slowing down as a personalized early warning signal for depression</article-title><source>Psychother Psychosom</source><year>2016</year><volume>85</volume><issue>2</issue><fpage>114</fpage><lpage>116</lpage><pub-id pub-id-type="doi">10.1159/000441458</pub-id><pub-id pub-id-type="medline">26821231</pub-id></nlm-citation></ref><ref id="ref61"><label>61</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Stiglic</surname><given-names>G</given-names> </name><name name-style="western"><surname>Kocbek</surname><given-names>P</given-names> </name><name name-style="western"><surname>Fijacko</surname><given-names>N</given-names> </name><name name-style="western"><surname>Zitnik</surname><given-names>M</given-names> </name><name name-style="western"><surname>Verbert</surname><given-names>K</given-names> </name><name name-style="western"><surname>Cilar</surname><given-names>L</given-names> </name></person-group><article-title>Interpretability of machine learning&#x2010;based prediction models in healthcare</article-title><source>WIREs Data Min &#x0026; Knowl</source><year>2020</year><month>09</month><volume>10</volume><issue>5</issue><fpage>e1379</fpage><pub-id pub-id-type="doi">10.1002/widm.1379</pub-id></nlm-citation></ref><ref id="ref62"><label>62</label><nlm-citation citation-type="web"><article-title>Simula</article-title><source>The Depresjon Dataset</source><access-date>2025-09-04</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://datasets.simula.no/depresjon/">https://datasets.simula.no/depresjon/</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Hyperparameters of models.</p><media xlink:href="mental_v12i1e72038_app1.docx" xlink:title="DOCX File, 14 KB"/></supplementary-material></app-group></back></article>