Published on in Vol 13 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/85319, first published .
The Performance of Wearable Device–Based Artificial Intelligence in Detecting Depression: Systematic Review and Meta-Analysis

The Performance of Wearable Device–Based Artificial Intelligence in Detecting Depression: Systematic Review and Meta-Analysis

The Performance of Wearable Device–Based Artificial Intelligence in Detecting Depression: Systematic Review and Meta-Analysis

1Liuzhou Railway Vocational Technical College, 2 Wenyuan Road, Yufeng District, Liuzhou, China

2Faculty of Humanities, Management and Science, Universiti Putra Malaysia, Sarawak, Malaysia

3School of Mechanical and Automotive Engineering, Guangxi University of Science and Technology, Liuzhou, China

4School of Automation, Guangxi University of Science and Technology, Liuzhou, China

Corresponding Author:

Jiawen Liu, MA


Background: In recent years, advances in wearable sensor technology and artificial intelligence (AI) have provided new possibilities for detecting and monitoring depression.

Objective: This study systematically reviewed and meta-analyzed the diagnostic and predictive performance of wearable device–based AI models for detecting depression and predicting depressive episodes and explored factors influencing outcomes.

Methods: Following PRISMA-DTA (Preferred Reporting Items for a Systematic Review and Meta-Analysis of Diagnostic Test Accuracy) guidelines, the PubMed, Embase, Web of Science, and PsycINFO databases were searched from inception to May 27, 2025. Eligible studies used AI algorithms on wearable device data for depression detection or episode prediction. Sensitivity, specificity, diagnostic odds ratio, and area under the curve (AUC) were pooled using a bivariate random effects model. Risk of bias was assessed using Prediction Model Risk of Bias Assessment Tool plus artificial intelligence (PROBAST+ AI), and certainty of evidence was assessed using the Grading of Recommendations Assessment, Development, and Evaluation (GRADE) tool.

Results: We included 16 studies (32 datasets) with 1189 patients and 13,593 samples. For depression detection, pooled sensitivity and specificity were 0.89 (95% CI 0.83‐0.93) and 0.93 (95% CI 0.87‐0.96), with a diagnostic odds ratio of 110.47 (95% CI 33.33‐366.17) and AUC of 0.96 (95% CI 0.94‐0.98). Random forest models showed the best performance (sensitivity=0.89, specificity=0.91, AUC=0.97). Subgroup analyses indicated that study design, AI method, reference standard, and input type significantly affected diagnostic accuracy (P<.05). For depressive episode prediction (3 datasets), pooled sensitivity was 0.86 (95% CI 0.80‐0.91), and pooled specificity was 0.65 (95% CI 0.59‐0.71). The overall risk of bias was low to moderate, with no evidence of publication bias.

Conclusions: Wearable device–based AI models achieved high accuracy for detecting depression and moderate utility in predicting episodes. However, heterogeneity, reliance on retrospective and public datasets, and lack of standardized methods limited generalizability.

Trial Registration: PROSPERO CRD420251070778; https://www.crd.york.ac.uk/PROSPERO/view/CRD420251070778

JMIR Ment Health 2026;13:e85319

doi:10.2196/85319

Keywords



Depression is a highly prevalent psychiatric disorder. According to World Health Organization (WHO) statistics, there are more than 350 million people with depression worldwide, and it is predicted that, by 2030, depression will become the leading cause of the global disease burden [1]. Patients with depression often experience persistent low mood, anhedonia, sleep disturbances, and cognitive impairment accompanied by a significantly increased risk of self-harm and suicide [2]. Depression not only has a significant impact at the individual level but also imposes a heavy economic burden on health care systems and society as a whole [3].

Traditionally, the diagnosis of depression relies on standardized clinical criteria and rating scales. The Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-IV) and Fifth Edition (DSM-V) provide operationalized criteria. These still depend on subjective symptom assessment and patient self-report, making them susceptible to reporting bias and interrater variability [4]. The Hamilton Depression Rating Scale (HDRS) and Montgomery-Åsberg Depression Rating Scale (MADRS) are commonly used to quantify disease severity. However, their accuracy depends on the clinician’s expertise and may vary due to different interpretations of symptom severity [5]. The Patient Health Questionnaire-9 (PHQ-9) is widely used in clinical and research settings due to its ease of administration; however, it remains limited by recall bias and the patient’s willingness or ability to express psychological distress accurately [6]. Although these scale and interview-based diagnostic methods are well-established, they lack objective biomarkers. They are limited in providing real-time and ecologically valid assessments, especially as symptom presentation may fluctuate over short periods.

In recent years, advances in wearable sensor technology and artificial intelligence (AI) have provided new possibilities for detecting and monitoring depression [7]. Wearable devices, such as wristbands and smartwatches, can collect longitudinal, multimodal physiological and behavioral data (eg, heart rate variability, sleep patterns, skin temperature, geolocation) [8,9]. This provides a more objective and continuous approach to assessing depressive symptoms. AI methods based on these data sources have demonstrated promising accuracy for depression classification and severity prediction, with studies reporting identification accuracies ranging from approximately 76% to >90% depending on the sample size, data type, and analytical strategy [10,11]. However, significant differences exist in the diagnostic performance reported across studies, largely due to variations in algorithms, wearable devices, and study populations [12,13]. Prior reviews, such as that by Abd-Alrazaq et al [14], have examined digital or sensor-based approaches for mental health detection, but these have mainly focused on feasibility or cross-sectional screening rather than on the predictive capacity of wearable device–based AI for future depressive episodes [15-17]. As systematic evaluations centered specifically on wearable-derived physiological and behavioral data remain limited, our review adds value by assessing the ability of wearable AI models to forecast depressive episodes, incorporating additional summary metrics such as the area under the curve (AUC) and diagnostic odds ratio (DOR) for a more comprehensive evaluation, and conducting subgroup analyses by algorithm type to clarify how methodological factors influence diagnostic performance.

We aimed to conduct a systematic review and meta-analysis to comprehensively assess the diagnostic performance of wearable device–based AI in depression detection and depressive episode prediction. Furthermore, we sought to evaluate how this performance is influenced by key methodological variables through subgroup analyses, focusing on factors such as study design, reference standard, AI algorithm type, and data source.


The meta-analysis rigorously adhered to the PRISMA-DTA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses of Diagnostic Test Accuracy) reporting guidelines [18]. The completed checklist is available as Checklist 1. Prior to study initiation, the research protocol was registered in the PROSPERO registry (registration ID: CRD420251070778).

Ethical Considerations

Because this is a systematic review and meta-analysis, ethics approval and consent to participate were not applicable. The manuscript does not include any identifiable participant image nor other personal or clinical details.

Search Strategy

A comprehensive literature search was conducted in 4 electronic databases: PubMed, Embase, Web of Science, and PsycINFO. The search was finalized on May 27, 2025. The search strategy incorporated 3 key groups of terms: (1) AI-related terms (eg, “artificial intelligence,” “machine learning,” “deep learning”), (2) disease-related terms (eg, “mood disorders,” “depression,” “psychological stress”), and (3) wearable device–related terms (eg, “wearable electronic devices,” “smart watch,” “accelerometer”). Both free-text terms and MeSH terms were used in combination. No restrictions were applied regarding language or publication year during the search. Detailed search strategies for each database are provided in Table S1 in Multimedia Appendix 1. In addition to systematic electronic searches, we performed backward and forward reference list checking of all included studies to find relevant publications in similar meta-analyses [14].

Inclusion and Exclusion Criteria

The PITROS (participants, index text, target conditions, reference standard, outcomes, and settings) framework was developed for the inclusion criteria. The detail is shown in Table 1. Studies were excluded if (1) the title or abstract was not relevant; (2) the publication type was a review, preprint, meta-analysis, conference abstract, or letter to the editor; (3) the study was not published in English; (4) AI was applied solely to predict treatment or intervention effects for depression, rather than detection or diagnosis; (5) data collection was not performed using wearable devices, including cases where only traditional medical equipment, handheld devices (such as smartphones), or implantable devices were used; or (6) data were collected exclusively via questionnaires or interviews without wearable device input. The screening process was conducted independently by two reviewers (JL and JW), with initial selection based on titles and abstracts, followed by a full-text assessment according to the inclusion and exclusion criteria. Duplicate references were identified and removed using EndNote and manual verification. Discrepancies were resolved through discussion and, if unresolved, by consulting a third reviewer (ZW).

Table 1. Summary of the inclusion criteria using the PITROS framework.
CriteriaDetails
Participants (P)Individuals with a clinical diagnosis of depression as well as healthy controls
Index test (I)Use of noninvasive wearable devices (eg, smartwatches) to collect physiological data for the development and evaluation of artificial intelligence (AI) algorithms aimed at detecting depression or predicting a depressive episode
Target conditions (T)Positive group: individuals who met the standardized criteria for depression or those experiencing a depressive episode; negative group: healthy individuals or those not experiencing an episode
Reference standard (R)Validated diagnostic scales or criteria such as the Montgomery-Åsberg Depression Rating Scale (MADRS); Patient Health Questionnaire-9 (PHQ-9); Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition (DSM-IV); Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (DSM-V); and Hamilton Depression Rating Scale (HDRS) used to verify wearable device–based AI algorithm performance
Outcomes (O)Sensitivity, specificity, diagnostic odds ratio (DOR), and area under the curve (AUC)
Settings (S)Retrospective or prospective studies conducted in contexts such as public databases or local hospitals

Quality Assessment

The quality assessment of the included studies was performed using the latest Prediction Model Risk of Bias Assessment Tool plus artificial intelligence (PROBAST+ AI) tool, which has replaced the previous PROBAST 2019 version [19]. This comprehensive instrument is structured into 2 phases—model development and model evaluation—with each phase comprising 7 domains. These domains address key aspects, including participants and data sources, predictors, outcome assessment, and analysis. Each domain is rated as a low (L), high (H), or unclear (U) risk of bias based on clearly defined signaling questions. The signaling questions are categorized as “Yes,” “Probably Yes,” “Probably No,” “No,” “No Information,” and, where applicable, “Not Applicable.” A domain is considered at low risk of bias if all signaling questions are answered with “Yes” or “Probably Yes.” Conversely, the presence of any “No” or “Probably No” responses in a domain indicates a potential high risk of bias. If neither of these are present but “No Information” is shown, the risk of bias is classified as unclear. Detailed signaling questions and summary tables are provided in Multimedia Appendix 1 (Table S2 and Table S3).

 To improve the accuracy during the quality assessment process, two reviewers (JL and JW) independently evaluated the risk of bias for each included study using the PROBAST+ AI tool. In cases where discrepancies arose, consensus was achieved through discussion and critical analysis.

Certainty of Evidence

The Grading of Recommendations, Assessment, Development, and Evaluation (GRADE) framework was used to assess the certainty of evidence for sensitivity, specificity, and DOR in detecting depression [20]. The evaluation focused on 5 key domains: risk of bias, inconsistency, indirectness, imprecision, and publication bias. If sufficient rationale was identified in any of these domains, the certainty of evidence was downgraded accordingly. Detailed assessment criteria and procedures are provided in Table S4 in Multimedia Appendix 1.

Data Extraction

Data extraction was conducted independently by 2 reviewers (JL and JW) who assessed the full-text articles to determine their potential eligibility for inclusion. In instances of disagreement, a third reviewer (ZW) acted as an arbitrator to achieve consensus. Extracted data encompassed study characteristics, patient characteristics, and technical parameters, including first author, year of publication, country, study design, target condition, reference standard, age, number of female participants, total participants, the positive sample size, and total sample size. Additionally, information specific to AI technical information was collected, such as the name of wearable devices, placement of wearable devices, data input, dataset source, AI method, AI algorithm, validation approach, and diagnostic performance measures.

True positive (TP) was defined as cases in which the AI model, based on wearable device data, identified depression or a depressive episode, and this was confirmed using the reference standard (MADRS, HDRS, PHQ-9, DSM-IV, or DSM-V). True negative (TN) was defined as nondepressed cases or a nondepressive episode as determined using both the AI model and the reference standard. False positive (FP) referred to cases where the AI model identified depression or a depressive episode, but this was not confirmed using the reference standard. Conversely, false negative (FN) referred to cases where the AI model failed to identify depression or depressive episodes that were confirmed using the reference standard. For eligible studies included in the systematic review but lacking sufficient data for meta-analysis, the corresponding authors were contacted via email to obtain the necessary information. If diagnostic contingency tables were not directly available in the publications, TP, FP, FN, and TN values were primarily back-calculated using reported sensitivity, specificity, the positive sample sizes identified using the reference standard, and the total sample size.

Outcome Measures

The core outcome parameters analyzed in this systematic review consisted of sensitivity, specificity, DOR, and AUC values extracted from internal validation populations. Sensitivity metrics, also designated as recall or TP rates, quantified the AI model’s capability to properly identify confirmed depression cases or a depressive episode, calculated as TP/(TP+ FN). Specificity, representing TN rates, reflected the algorithm’s accuracy for recognizing nondepressed participants or a nondepressive episode determined via TN/(TN+ FP). AUC represents the area under the receiver operating characteristic curve and serves as a comprehensive measure of the model’s ability to distinguish between positive (depressed or depressive episode) and negative (nondepressed or nondepressive episode) cases. The DOR, an integrative measure of diagnostic performance, combines both sensitivity and specificity, expressing the odds of a positive test result among patients with depression relative to those without depression [21]. For studies evaluating multiple types of AI algorithms, all relevant results were extracted to facilitate direct comparison across different algorithmic methods.

Statistical Analysis

A bivariate random effects model was used to meta-analyze and assess the diagnostic performance [22]. Forest plots were used to visually present the pooled sensitivity, specificity, and DOR for internal validation datasets, while a summary receiver operating characteristic curve was constructed to display the 95% confidence and prediction regions for the overall estimates [23]. The prediction region indicates the likely range of sensitivity and specificity for future studies, and the pooled AUC was also computed. Heterogeneity across studies was assessed using the Higgins I² statistic, with values of 25%, 50%, and 75% denoting low, moderate, and high heterogeneity, respectively [24]. For internal validation datasets with substantial heterogeneity (I²>50%), we used bivariate boxplots to identify outliers beyond the 95% CI and to explore potential sources of heterogeneity. In addition, according to the predefined protocol, subgroup analyses were conducted using z tests to evaluate the impact of potential variables on the results. Variables included study design (retrospective vs prospective), reference standard (MADRS vs other reference standards), AI method (machine learning vs deep learning), type of input data (only activity data vs others), and data source (open vs closed).

Radar plots were generated to illustrate the distribution of algorithms among the included studies, and bubble plots were used to visualize the trends in the DOR of AI models over time. The Fagan nomogram was applied to evaluate the clinical impact of the AI models [25]. Assessment of publication bias was performed using the Deeks funnel plot asymmetry test, involving a regression of the adequate sample size against the log DOR. All statistical analyses were conducted using the “midas” and “metadta” packages in Stata version 15.1 and R version 4.3.1. A P value <.05 was considered statistically significant.


Study Selection

A total of 1656 potentially relevant articles were identified through the searches of the 4 primary databases. After removing 513 duplicates, 1143 unique records remained for preliminary screening. During this phase, 1114 records were excluded due to apparent irrelevance, as determined by their titles and abstracts, or because they did not meet the required publication types. As a result, 60 articles proceeded to the full-text review. Following a detailed assessment, exclusions included the following: 32 studies for insufficient or incomplete diagnostic data (lack of TP, FP, FN, or TN information), 7 for overlapping populations, 2 for not focusing on depression, 2 for not using wearable devices for data collection, and 1 for being published in a language other than English. In addition, we identified 3 records through nondatabase sources. Ultimately, 16 studies fulfilled all inclusion criteria and were included in the meta-analysis [10-13,16,17,26-35]. The study selection process adhered to the PRISMA guidelines, as detailed in Figure 1.

Figure 1. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram illustrating the study selection process. AI: artificial intelligence, FN: false negative, FP: false positive, TN: true negative, TP: true positive.

Study Description and Quality Assessment

A total of 16 eligible studies were included, encompassing 32 internal validation datasets with an aggregate of 1189 patients and 13,593 samples. Data derived from the Depresjon dataset were used in 11 studies [10,11,13,17,26-30,32,35]. The included studies were published between 2019 and 2023. Geographically, one-half (8/16, 50%) were conducted in the Americas (Mexico: n=6; United States, n=2) [10,11,17,27-29,32,35], while the remainder were carried out in Asia (Japan, n=2; China, n=1; Korea, n=1; India, n=1; Saudi Arabia, n=1) [12,13,16,31,33,34] and Europe (Poland, n=1; Norway, n=1) [26,30]. The study designs comprised 10 retrospective [10,11,13,26-30,32,35] and 6 prospective [12,16,17,31,33,34] studies. Regarding reference standards, 10 studies used the MADRS [10,11,13,26-30,32,35], 2 studies used the PHQ-9 [16,17], 1 study used the DSM-IV [34], 1 study used the DSM-V [33], 1 study used the HDRS [12], and 1 study used both the DSM-V and the PHQ-9 [31]. Depression diagnosis was evaluated in 14 studies [10-13,26-35], whereas 2 focused on the identification of depressive episodes [16,17].

Data collection was primarily performed using wrist-worn wearable devices (15 studies) [10-13,16,17,26-33,35], while head-worn devices were used in 1 study [34]. In 10 studies, the data input consisted solely of activity data [10,11,13,26-30,32,35], A combination of activity and sleep data was used in 1 study [33], 1 study relied on electrocardiogram data [34] , and 4 studies incorporated more than 3 types of parameters such as activity data, sleep data, and heart rate for model training [12,16,17,31]. The majority of studies (13/16, 81%) applied machine learning methods [10-13,16,17,26,28,29,31-33,35], while 2 studies used deep learning methods and 1 study used both machine learning and deep learning methods [30]. Data sources were classified as open in 11 studies [10,11,13,26-30,32,34,35] and closed in 5 studies [12,16,17,31,33]. Summaries of the study, patient, and technical characteristics are provided in Table 2 and Multimedia Appendix 1 (Table S5 and Table S6). Among AI algorithms, random forest (RF) was the most frequently implemented (10/32, 31%), with a detailed distribution of the algorithm shown in Figure S1 in Multimedia Appendix 1.

Table 2. Study and patient characteristics of the included studies.
AuthorYearCountryStudy designTarget conditionReference standardAge (years)Gender (female), n (%)Total participants, nPositive sample size, nTotal sample size, n
Adamczyk et al [26]2021PolandRetrospectiveDepression vs healthyMADRSa40.1b30 (55)552355
Bai et al [16]2021ChinaProspectiveDepressive episodePHQ-9c18‐60dNRe261201201
Espino-Salinas et al [27]2022MexicoRetrospectiveDepression vs healthyMADRS40.1b30 (55)5536116
Galván-Tejada et al [28]2019MexicoRetrospectiveDepression vs healthyMADRS40.1b30 (55)559622483
Jacobson et al [29]2019United StatesRetrospectiveDepression vs healthyMADRS40.1b30 (55)552355
Jakobsen et al [30]2020NorwayRetrospectiveDepression vs healthyMADRS40.1b30 (55)552355
Mullick et al [17]2022United StatesProspectiveDepressive episodePHQ-915.5 (12‐18)f41 (75)55355470
Narziev et al [31]2020KoreaProspectiveDepression vs healthyDSM‐5g, PHQ-9NRNR20430600
Pacheco-González et al [10]2019MexicoRetrospectiveDepression vs healthyMADRS40.1b30 (55)552355
Rodríguez-Ruiz et al [32]2020MexicoRetrospectiveDepression vs healthyMADRS40.1b30 (55)5512293574
Rodríguez-Ruiz et al [11]2022MexicoRetrospectiveDepression vs healthyMADRS40.1b53 (49)1094391293
Sato et al [33]2023JapanProspectiveDepression vs healthyDSM-535.6 (11.3)h36 (52)694069
Sharma et al [34]2023IndiaProspectiveDepression vs healthyDSM-4i39.3 (12‐77)f24 (38)643464
Tazawa et al [12]2020JapanProspectiveDepression vs healthyHDRSj60.2b40 (47)86112236
Zanella-Calzada et al [35]2019MexicoRetrospectiveDepression vs healthyMADRS40.1b30 (55)5516544125
Zakariah and Alotaibi [13]2023Saudi ArabiaRetrospectiveDepression vs healthyMADRS20‐69d30 (55)5546142

aMADRS: Montgomery-Åsberg Depression Rating Scale.

bMean.

cPHQ-9: Patient Health Questionnaire-9.

dRange.

eNR: not reported.

fMedian (range).

gDSM-5: Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition.

hMean (SD).

iDSM-4: Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition.

jHDRS: Hamilton Depression Rating Scale.

The risk of bias for the included studies was assessed using the PROBAST+ AI quality assessment tool, with results summarized in Figure 2 and Multimedia Appendix 1 (Table S3 and Table S4). For model development, the overall quality judgment rated 37.5% (6/16) of studies as high risk and the remaining 62% (10/16) as low risk. In terms of applicability concerns, 6% (1/16) of studies were rated as high risk, while 94% (15/16) were rated as low risk. For model validation, the overall risk of bias judgment classified 44% (7/16) of studies as high risk, 19% (3/16) as unclear risk, and 38% (6/16) as low risk. Regarding applicability concerns for model validation, 19% (3/16) of the studies were considered high risk, and 81% (13/16) were rated as low risk. Overall, high-risk items were relatively infrequent, with most studies judged as low risk, indicating that the general methodological quality of the included studies is acceptable. The certainty of evidence, as evaluated using the GRADE framework, ranged from low to high across outcomes, suggesting that the certainty of the evidence was generally moderate (Table S4 in Multimedia Appendix 1).

Figure 2. PROBAST+ AI (Prediction Model Risk of Bias Assessment Tool plus artificial intelligence) quality assessment, including risk of bias (high, low, or unclear) of the included studies: (A) model development and (B) model validation.

Diagnostic Performance of Individual Studies

In the study by Rodríguez-Ruiz et al [32], the detection of depression using an RF model trained on activity data achieved the highest sensitivity (0.99) among all included studies. Their study also reported the highest specificity (0.99) for detecting depression. In addition, for predicting depressive episodes, the study by Bai et al [16] achieved the highest sensitivity (0.90) using a k nearest neighbor (KNN) model trained on activity data, heart rate data, location information, sleep data, smartphone use data, and social interaction data. This study also reported the highest specificity (0.66) for depressive episode prediction using an RF model trained on the same multimodal data sources.

Diagnostic Performance of Different AI Algorithms for Depression in the Internal Validation Set

Within the internal validation cohorts, the DOR for depression detection using wearable device–based AI algorithms demonstrated a progressive increase from 2019 to 2023, with the highest DOR observed for the Uniform Manifold Approximation and Projection and neural network (UMAP&NN) algorithm and the lowest for the naïve Bayes algorithm (Figure S2 in Multimedia Appendix 1). Subgroup analysis according to algorithm type revealed that the RF algorithm was used in 9 datasets and exhibited robust and promising performance: Sensitivity was 0.89 (95% CI 0.81‐0.94), specificity was 0.91 (95% CI 0.80‐0.96), and the AUC was 0.97 (95% CI 0.95‐0.98; Table S7 in Multimedia Appendix 1).

Diagnostic Performance of Wearable Device–Based AI Models for Depression in the Internal Validation Set

A total of 29 datasets from 14 studies were included in the pooled analysis of diagnostic performance. The combined sensitivity of AI models for depression detection was 0.89 (95% CI 0.83‐0.93; moderate certainty), and the specificity was 0.93 (95% CI 0.87‐0.96; moderate certainty). The pooled DOR was 110.47 (95% CI 33.33‐366.17; low certainty), as illustrated in Figures 3 and 4. Additionally, the AUC for the models was 0.96 (95% CI 0.94‐0.98; Figure 5). Using a pretest probability of 20%, the Fagan nomogram demonstrated a positive likelihood ratio of 76% and a negative likelihood ratio of 3% (Figure S3 in Multimedia Appendix 1). No evidence of publication bias was detected according to the Deeks funnel plot asymmetry test (P=.23; Figure S4 in Multimedia Appendix 1).

Figure 3. Forest plot of sensitivity and specificity for depression detection using artificial intelligence (AI) in wearable devices, with each row representing an individual study evaluating the performance of different AI algorithms for detecting depression [10,11,12,13,26,27,28,29,30,32,33,34,35]. Boxes: point estimates; horizontal lines: 95% CIs; diamond at the bottom: pooled estimate. CNN: convolutional neural network; DNN: deep neural network; KNN: k nearest neighbor; LR: logistic regression; LSTM: long short-term memory; NB: naïve Bayes; NN: neural network; RF: random forest; SVM: support vector machine; UMAP: Uniform Manifold Approximation and Projection.
Figure 4. Forest plot of diagnostic performance of wearable device–based artificial intelligence (AI) models for depression in the internal validation set [10,11,12,13,26,27,28,29,30,31,32,33,34,35]. CNN: convolutional neural network; DNN: deep neural network; KNN: k nearest neighbor; LR: logistic regression; LSTM: long short-term memory; NB: naïve Bayes; NN: neural network; RF: random forest; SVM: support vector machine; UMAP: Uniform Manifold Approximation and Projection.
Figure 5. Summary receiver operating characteristic (SROC) curve for depression diagnosis using artificial intelligence (AI) in wearable devices, showing the individual study estimates of sensitivity (SENS) and specificity (SPEC; blue circles) and the summary operating point with pooled sensitivity and specificity and corresponding 95% CIs (red diamond). AUC: area under the curve.

Bivariate Boxplot and Subgroup Analysis of Wearable Device–Based AI Models in Internal Validation Set

Substantial heterogeneity was observed in both sensitivity (I²=96.52%) and specificity (I²=97.43%) for depression diagnosis. The bivariate boxplot illustrates that the results from Pacheco-González et al [10] (naïve Bayes and KNN), Zakariah and Alotaibi [13] (UMAP&NN), and Rodríguez-Ruiz et al [32] (RF) fall outside the 95% confidence region, suggesting that these studies may represent potential sources of heterogeneity (Figure 6). Subgroup analyses further demonstrated significant differences in the sensitivity and specificity of AI models according to study design (retrospective vs prospective), reference standard (MADRS vs other reference standards), AI method (machine learning vs deep learning), and data input (only activity data vs other types of data), with all comparisons yielding P<.05 (Table S8 in Multimedia Appendix 1).

Figure 6. Bivariate boxplot of logit-transformed sensitivity and specificity for wearable device–based artificial intelligence (AI) depression detection [10,13,32]. The inner shaded oval indicates the median distribution of the data points, while the outer shaded oval represents the 95% confidence boundary. KNN: k nearest neighbor; NB: naïve Bayes; NN: neural network; RF: random forest; UMAP: Uniform Manifold Approximation and Projection.

Predictive Performance of Wearable Device–Based AI Models for Depressive Episodes in the Internal Validation Set

A total of 3 datasets from 2 studies were included in the pooled analysis of predictive performance for depressive episodes. The wearable device–based AI models demonstrated a sensitivity of 0.86 (95% CI 0.80‐0.91; high certainty) and specificity of 0.65 (95% CI 0.59‐0.71; high certainty), as shown in Figure S5 in Multimedia Appendix 1.


Principle Findings

Our results demonstrate that wearable device–based AI achieves promising diagnostic performance in depression diagnosis, with sensitivity, specificity, DOR, and AUC of 0.89, 0.93, 110.47, and 0.96, respectively, showing extremely high diagnostic performance. The superior performance of wearable AI systems in depression detection can be attributed to several key mechanisms. First, these devices enable continuous and objective monitoring of multiple physiological and behavioral parameters intrinsically related to depressive symptoms, including motor activity data, sleep data, heart rate data, and disruptions to the circadian rhythm [35,36]. Studies consistently show that, compared with healthy controls, patients with depression exhibit significantly reduced motor activity, altered sleep patterns, and decreased step counts, with these differences being particularly pronounced during specific periods (11 AM to 6 PM) [7]. Second, machine learning algorithms, particularly when combined with dimensionality reduction techniques such as UMAP&NN and ensemble methods like RF and deep neural networks, can identify complex, nonlinear relationships among multiple sensor modalities that may not be apparent in traditional clinical assessments [13]. The integration of features such as skin temperature and sleep duration–related parameters, along with their correlations, has proven particularly valuable in predictive models [37]. Third, the longitudinal nature of wearable data collection enables the detection of subtle changes in behavioral patterns over time, allowing for the early identification of depressive episodes even before their clinical manifestations. This continuous monitoring approach minimizes the biases inherent in traditional self-report measures and provides a more comprehensive assessment of patient status [38].

Our results show that wearable device–based AI has a sensitivity of 0.86 and a specificity of 0.65 for detecting depressive episodes, indicating relatively low specificity. The low specificity of wearable AI systems for detecting depressive episodes may be attributed to several factors. First, in the studies we included, patient populations were patients with major depression who tend to have more complex clinical presentations and comorbidities, which may lead to the overlap of physiological signaling patterns with other psychiatric or physical disorders, thus increasing the rate of FPs [16,17,39]. Additionally, data collected by wearable devices are often affected by user compliance, device wearing habits, and external interferences, which can lead to increased data noise and thus affect the specificity of the model [40]. However, it is important to note that our study on depressive episodes had only 3 datasets and a relatively small sample size, which may affect the stability and reliability of the results. Small sample sizes limit the efficacy of statistical analyses, potentially leading to wider CIs and increased uncertainty in the results. Future prospective studies with larger sample sizes are needed to obtain more robust results.

Interestingly, the results of our subgroup analyses showed that both sensitivity and specificity from prospective studies were significantly higher than those from retrospective studies. The superior diagnostic performance demonstrated by prospective studies in wearable AI depression diagnosis can be attributed to several key mechanisms. First, the prospective study design enables researchers to standardize control over the data collection process, ensuring consistency in wearable device configuration, wearing time, and data quality, whereas retrospective studies often face inconsistent data collection standards. Second, prospective studies can minimize selection bias and recall bias because patients and controls are identified at the beginning of the study, thereby avoiding the risk of sample selection based on known outcomes [12].

Consistent with previous knowledge, our subgroup analysis results showed that both sensitivity and specificity from deep learning algorithms were significantly higher than those from traditional machine learning algorithms. The mechanism by which deep learning algorithms exhibit superior diagnostic performance in wearable AI depression detection can be attributed to their unique architectural advantages and data processing capabilities. First, deep learning models such as deep neural networks and convolutional neural networks possess powerful automatic feature extraction capabilities. This allows them to learn and recognize complex nonlinear patterns directly from raw wearable device data. In contrast, traditional machine learning methods, such as RF, tend to rely on manually engineered statistical features [41]. Second, deep learning models can capture higher-order interactions and temporal dependencies in the data through a multilayer neural network structure, which is crucial for understanding the complex physiological and behavioral patterns of individuals with depression [42]. The results of this study are consistent with those of the study by Jakobsen et al [30], where the sensitivity (0.65) and specificity (0.78) of the convolutional neural network algorithm were similarly shown to be superior to that of the RF algorithm (sensitivity=0.61, specificity=0.75) in their study.

Another interesting finding of our meta-analysis was that AI models trained solely on activity data demonstrated significantly lower sensitivity and specificity than algorithms trained on a combination of multiple data types, such as sleep, heart rate, and activity data. This difference in performance can be attributed to the fact that depression manifests itself through multiple pathophysiological pathways that affect autonomic nervous system function, circadian rhythms, and behavioral patterns [41]. Unimodal approaches may miss critical diagnostic information captured by other sensors, such as heart rate variability, which plays a key role in depression detection as an essential indicator of autonomic dysfunction [43]. The accuracy of the algorithm trained using a combination of multiple datasets (76.67%) was significantly better than that of the AI model trained on activity data (69.24%), as reported by Bai et al [16], further confirming the results of this study.

Compared With Previous Studies

In 2023, Abd-Alrazaq et al [44] conducted a meta-analysis of wearable device–based AI anxiety tests and reported a sensitivity of 0.79, a specificity of 0.92, and an accuracy of 0.82. Compared with these benchmark results, our study demonstrated that wearable device–based AI depression tests achieved superior diagnostic performance, with sensitivity and specificity of 0.89 and 0.91, respectively. These differences in diagnostic performance may stem from fundamental differences in disease specificity. As noted in the literature, anxiety and depression are related but have different manifestations [44]. Specifically, physiologic indicators of anxiety focus on acute physiologic responses such as elevated heart rate, sweating, and muscle tension, whereas physiologic indicators of depression are more often characterized by chronic changes in sleep patterns, physical activity levels, and mood states [45]. Because of this, chronic physiological changes in depression may be more easily and accurately captured and recognized by existing wearable device technologies than acute physiological responses to anxiety.

Additionally, in 2023, Abd-Alrazaq et al [14] conducted another meta-analysis on the performance of wearable device–based AI for depression detection, which showed a sensitivity of 0.87 and specificity of 0.93. Our results showed that the sensitivity and specificity of wearable device–based AI depression detection were 0.89 and 0.91, respectively, which were generally similar to the results of their study. However, compared with previous studies, our study offers several key innovations. First, it also evaluates the predictive ability for depressive episodes, expanding the scope beyond standard diagnostic assessment. Second, we incorporated additional metrics, such as the AUC and DOR, to provide a more comprehensive evaluation of the diagnostic accuracy of wearable devices. Third, we conducted separate analyses for specific AI algorithms and examined the trends in diagnostic performance over time. Finally, we included more detailed subgroup analyses based on study design, reference standard, AI method, type of input data, and data source, thereby assessing how these factors influence the outcome measures.

Heterogeneity

Our meta-analysis revealed substantial heterogeneity in the pooled estimates, which is a common and expected finding in meta-analyses of AI-based diagnostic tools. This high degree of heterogeneity reflects the inherent methodological diversity across studies, stemming from variability in datasets, wearable sensor types, data preprocessing pipelines, and AI model architectures. Although our bivariate box-and-whisker plots identified the study by Rodríguez-Ruiz et al [32] as a potential outlier, likely due to its exclusive use of nocturnal data for classification, the overall heterogeneity was multifaceted. Our subgroup analyses further confirmed that study design, AI methodology, input data type, and reference standard significantly influenced diagnostic performance metrics. Therefore, the observed heterogeneity should be interpreted not only as a limitation of the included studies but also as an indicator of the current state of a rapidly evolving and technically diverse research domain.

Interpreting the Results

When interpreting our results, this study demonstrated that wearable device–based AI exhibits promising performance for diagnosing depression but shows limited effectiveness for predicting depressive episodes. This finding suggests that wearable device–based AI has potential applications in the field of depression diagnosis, and the future deployment of such technologies in daily medical practice may facilitate early identification of and timely intervention in depression. However, it should be clear that these models, at the current level of technology, cannot yet be used as an independent diagnostic tool or basis for clinical decision-making, and the final diagnosis still needs to be comprehensively evaluated by clinical professionals in combination with standardized assessment tools [46]. It is noteworthy that the vast majority of the studies included in the analysis used wrist-worn wearable devices, while only a few studies explored the application of devices on other body parts [34]. This suggests that research on expanding the application sites of wearable devices may be a direction worth exploring in the future. Additionally, the current evidence is based predominantly on internal validation datasets. The general scarcity of rigorous external validation across the literature represents a critical barrier to the clinical translation and real-world reliability of these models, highlighting a key area for future research [45,47].

Limitations

Several limitations of this meta-analysis should be considered when interpreting the results. First, most of the included studies (10/16, 63%) used a retrospective design, which may have introduced potential selection bias and information bias, as retrospective studies rely on pre-existing data and may have incomplete data or variable quality [48]. Therefore, well-designed prospective studies are needed to validate the findings of this meta-analysis to provide higher-quality evidence. Second, regarding the definition of a gold standard, there were significant differences in the reference standards used across different studies, including variations in depression assessment scales and diagnostic criteria. We attempted to assess the impact of varying gold standards through subgroup analyses, which showed that different subgroup analyses did affect the estimates of sensitivity and specificity, suggesting that the results need to be interpreted with caution and future studies need to be more standardized and consistent in terms of the gold standard qualification [49]. Third, most studies (11/16, 69%) relied on a public dataset (Depresjon datasets) as their primary data sources, which may limit the model’s generalizability and applicability across different populations. In the future, there is a need to collect patient data from more regions and health care organizations to train AI models with more diverse data and validate the external validity of the results [47,50].

Conclusion

In conclusion, this study demonstrates that wearable device–based AI exhibits promising performance for diagnosing depression but shows limited ability for predicting depressive episodes. Deep learning algorithms and the integration of multimodal data inputs significantly outperformed traditional approaches. However, substantial heterogeneity among studies, along with the predominance of retrospective designs and reliance on public datasets, limits the generalizability of these findings. Future research should focus on large-scale, prospective studies with standardized protocols to enhance clinical applicability and ensure broader external validity.

Acknowledgments

The authors would like to thank all contributors and reviewers of the included studies.

This article was written by the author independently, without using any artificial intelligence (AI) tools or software to generate, edit, or modify the content.

Funding

No funding was received for this work.

Data Availability

All data generated or analyzed during this study are included in this published article. Further inquiries can be directed to the corresponding author.

Authors' Contributions

Conceptualization: JL

Data curation: JL

Formal analysis: JL

Investigation: JL, JW, ZW, MISBAAA

Methodology: JL, JW, ZW, MISBAAA

Writing – review & editing: JL

Conflicts of Interest

None declared.

Multimedia Appendix 1

Search strategy and additional assessments of the studies.

DOCX File, 3737 KB

Checklist 1

PRISMA-DTA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses of Diagnostic Test Accuracy) checklist.

PDF File, 86 KB

  1. Yadav U, Sharma AK, Patil D. Review of automated depression detection: social posts, audio and video, open challenges and future direction. Concurr Comput. Jan 10, 2023;35(1):e7407. URL: https://onlinelibrary.wiley.com/toc/15320634/35/1 [CrossRef]
  2. Ducasse D, Dubois J, Jaussent I, et al. Association between anhedonia and suicidal events in patients with mood disorders: a 3-year prospective study. Depress Anxiety. Jan 2021;38(1):17-27. [CrossRef] [Medline]
  3. de Sousa RD, Zagalo DM, Costa T, de Almeida JMC, Canhão H, Rodrigues A. Exploring depression in adults over a decade: a review of longitudinal studies. BMC Psychiatry. Apr 15, 2025;25(1):378. [CrossRef] [Medline]
  4. Nunes EV, Rounsaville BJ. Comorbidity of substance use with depression and other mental disorders: from Diagnostic and Statistical Manual of Mental Disorders, fourth edition (DSM-IV) to DSM-V. Addiction. Sep 2006;101 Suppl 1(s1):89-96. [CrossRef] [Medline]
  5. Carneiro AM, Fernandes F, Moreno RA. Hamilton Depression Rating Scale and Montgomery-Asberg Depression Rating Scale in depressed and bipolar I patients: psychometric properties in a Brazilian sample. Health Qual Life Outcomes. Apr 2, 2015;13:42. [CrossRef] [Medline]
  6. Levis B, Benedetti A, Thombs BD, DEPRESsion Screening Data (DEPRESSD) Collaboration. Accuracy of Patient Health Questionnaire-9 (PHQ-9) for screening to detect major depression: individual participant data meta-analysis. BMJ. Apr 9, 2019;365:l1476. [CrossRef] [Medline]
  7. Woll S, Birkenmaier D, Biri G, et al. Applying AI in the context of the association between device-based assessment of physical activity and mental health: systematic review. JMIR Mhealth Uhealth. Mar 6, 2025;13:e59660. [CrossRef] [Medline]
  8. Moshe I, Terhorst Y, Opoku Asare K, et al. Predicting symptoms of depression and anxiety using smartphone and wearable data. Front Psychiatry. 2021;12:625247. [CrossRef] [Medline]
  9. Shui X, Xu H, Tan S, Zhang D. Depression recognition using daily wearable-derived physiological data. Sensors (Basel). Jan 19, 2025;25(2):567. [CrossRef] [Medline]
  10. Pacheco-González SL, Zanella-Calzada LA, Galván-Tejada CE, Chávez-Lamas NM, Rivera-Gómez JF, Galván-Tejada JI. Evaluation of five classifiers for depression episodes detection. RCS. 2019;148(10):129-138. [CrossRef]
  11. Rodríguez-Ruiz JG, Galván-Tejada CE, Luna-García H, et al. Classification of depressive and schizophrenic episodes using night-time motor activity signal. Healthcare (Basel). Jul 5, 2022;10(7):1256. [CrossRef] [Medline]
  12. Tazawa Y, Liang KC, Yoshimura M, et al. Evaluating depression with multimodal wristband-type wearable device: screening and assessing patient severity utilizing machine-learning. Heliyon. Feb 2020;6(2):e03274. [CrossRef] [Medline]
  13. Zakariah M, Alotaibi YA. Unipolar and bipolar depression detection and classification based on actigraphic registration of motor activity using machine learning and uniform manifold approximation and projection methods. Diagnostics (Basel). Jul 10, 2023;13(14):2323. [CrossRef] [Medline]
  14. Abd-Alrazaq A, AlSaad R, Shuweihdi F, Ahmed A, Aziz S, Sheikh J. Systematic review and meta-analysis of performance of wearable artificial intelligence in detecting and predicting depression. NPJ Digit Med. May 5, 2023;6(1):84. [CrossRef] [Medline]
  15. Wang W, Chen J, Hu Y, et al. Integration of artificial intelligence and wearable Internet of Things for mental health detection. International Journal of Cognitive Computing in Engineering. 2024;5:307-315. [CrossRef]
  16. Bai R, Xiao L, Guo Y, et al. Tracking and monitoring mood stability of patients with major depressive disorder by machine learning models using passive digital data: prospective naturalistic multicenter study. JMIR Mhealth Uhealth. Mar 8, 2021;9(3):e24365. [CrossRef] [Medline]
  17. Mullick T, Radovic A, Shaaban S, Doryab A. Predicting depression in adolescents using mobile and wearable sensors: multimodal machine learning-based exploratory study. JMIR Form Res. Jun 24, 2022;6(6):e35807. [CrossRef] [Medline]
  18. Cohen JF, Deeks JJ, Hooft L, et al. Preferred reporting items for journal and conference abstracts of systematic reviews and meta-analyses of diagnostic test accuracy studies (PRISMA-DTA for Abstracts): checklist, explanation, and elaboration. BMJ. Mar 15, 2021;372:n265. [CrossRef] [Medline]
  19. Moons KGM, Damen JAA, Kaul T, et al. PROBAST+AI: an updated quality, risk of bias, and applicability assessment tool for prediction models using regression or artificial intelligence methods. BMJ. Mar 24, 2025;388:e082505. [CrossRef] [Medline]
  20. Gopalakrishna G, Mustafa RA, Davenport C, et al. Applying Grading of Recommendations Assessment, Development and Evaluation (GRADE) to diagnostic tests was challenging but doable. J Clin Epidemiol. Jul 2014;67(7):760-768. [CrossRef] [Medline]
  21. Gu Y, Xue J, Xia X, et al. Prediction of post stroke depression with machine learning: a national multicenter cohort study. J Psychiatr Res. Jul 2025;187:123-133. [CrossRef] [Medline]
  22. Arends LR, Hamza TH, van Houwelingen JC, Heijenbrok-Kal MH, Hunink MGM, Stijnen T. Bivariate random effects meta-analysis of ROC curves. Med Decis Making. 2008;28(5):621-638. [CrossRef] [Medline]
  23. Walter SD. Properties of the summary receiver operating characteristic (SROC) curve for diagnostic test data. Stat Med. May 15, 2002;21(9):1237-1256. [CrossRef] [Medline]
  24. Higgins JPT, Thompson SG. Quantifying heterogeneity in a meta-analysis. Stat Med. Jun 15, 2002;21(11):1539-1558. [CrossRef] [Medline]
  25. Safari S, Baratloo A, Elfil M, Negida A. Evidence based emergency medicine; part 4: pre-test and post-test probabilities and Fagan’s nomogram. Emerg (Tehran). 2016;4(1):48-51. [Medline]
  26. Adamczyk J, Malawski F. Comparison of manual and automated feature engineering for daily activity classification in mental disorder diagnosis. cai. 2021;40(4):850-879. [CrossRef]
  27. Espino-Salinas CH, Galván-Tejada CE, Luna-García H, et al. Two-dimensional convolutional neural network for depression episodes detection in real time using motor activity time series of depresjon dataset. Bioengineering (Basel). Sep 9, 2022;9(9):458. [CrossRef] [Medline]
  28. Galván-Tejada CE, Zanella-Calzada LA, Gamboa-Rosales H, et al. Depression episodes detection in unipolar and bipolar patients: a methodology with feature extraction and feature selection with genetic algorithms using activity motion signal as information source. Mobile Information Systems. Apr 23, 2019;2019(1):1-12. [CrossRef]
  29. Jacobson NC, Weingarden H, Wilhelm S. Digital biomarkers of mood disorders and symptom change. NPJ Digit Med. 2019;2(3):3. [CrossRef] [Medline]
  30. Jakobsen P, Garcia-Ceja E, Riegler M, et al. Applying machine learning in motor activity time series of depressed bipolar and unipolar patients compared to healthy controls. PLoS ONE. 2020;15(8):e0231995. [CrossRef] [Medline]
  31. Narziev N, Goh H, Toshnazarov K, Lee SA, Chung KM, Noh Y. STDD: short-term depression detection with passive sensing. Sensors (Basel). Mar 4, 2020;20(5):1396. [CrossRef] [Medline]
  32. Rodríguez-Ruiz JG, Galván-Tejada CE, Vázquez-Reyes S, Galván-Tejada JI, Gamboa-Rosales H. Classification of depressive episodes using nighttime data; a multivariate and univariate analysis. Program Comput Soft. Dec 2020;46(8):689-698. [CrossRef]
  33. Sato S, Hiratsuka T, Hasegawa K, et al. Screening for major depressive disorder using a wearable ultra-short-term HRV monitor and signal quality indices. Sensors (Basel). Apr 10, 2023;23(8):3867. [CrossRef] [Medline]
  34. Sharma G, Joshi AM, Gupta R, Cenkeramaddi LR. DepCap: a smart healthcare framework for EEG based depression detection using time-frequency response and deep neural network. IEEE Access. 2023;11:52327-52338. [CrossRef]
  35. Zanella-Calzada LA, Galván-Tejada CE, Chávez-Lamas NM, et al. Feature extraction in motor activity signal: towards a depression episodes detection in unipolar and bipolar patients. Diagnostics (Basel). Jan 10, 2019;9(1):8. [CrossRef] [Medline]
  36. Rykov Y, Thach TQ, Bojic I, Christopoulos G, Car J. Digital biomarkers for depression screening with wearable devices: cross-sectional study with machine learning modeling. JMIR Mhealth Uhealth. Oct 25, 2021;9(10):e24872. [CrossRef] [Medline]
  37. Nejadshamsi S, Karami V, Ghourchian N, et al. Development and feasibility study of HOPE model for prediction of depression among older adults using wi-fi-based motion sensor data: machine learning study. JMIR Aging. Mar 3, 2025;8:e67715. [CrossRef] [Medline]
  38. Ortiz A, Halabi R, Alda M, et al. Day-to-day variability in activity levels detects transitions to depressive symptoms in bipolar disorder earlier than changes in sleep and mood. Int J Bipolar Disord. Apr 2, 2025;13(1):13. [CrossRef] [Medline]
  39. Bladon S, Eisner E, Bucci S, et al. A systematic review of passive data for remote monitoring in psychosis and schizophrenia. NPJ Digit Med. Jan 27, 2025;8(1):62. [CrossRef] [Medline]
  40. Menassa M, Wilmont I, Beigrezaei S, et al. The future of healthy ageing: wearables in public health, disease prevention and healthcare. Maturitas. May 2025;196:108254. [CrossRef] [Medline]
  41. Zhang S, Li Y, Zhang S, et al. Deep learning in human activity recognition with wearable sensors: a review on advances. Sensors (Basel). Feb 14, 2022;22(4):1476. [CrossRef] [Medline]
  42. He L, Niu M, Tiwari P, et al. Deep learning for depression recognition with audiovisual cues: a review. Information Fusion. Apr 2022;80:56-86. [CrossRef]
  43. Ahmed A, Ramesh J, Ganguly S, Aburukba R, Sagahyroon A, Aloul F. Investigating the feasibility of assessing depression severity and valence-arousal with wearable sensors using discrete wavelet transforms and machine learning. Information. 2022;13(9):406. [CrossRef]
  44. Abd-Alrazaq A, AlSaad R, Harfouche M, et al. Wearable artificial intelligence for detecting anxiety: systematic review and meta-analysis. J Med Internet Res. Nov 8, 2023;25:e48754. [CrossRef] [Medline]
  45. Li F, Zhang D. Transformer-driven affective state recognition from wearable physiological data in everyday contexts. Sensors (Basel). Jan 27, 2025;25(3):761. [CrossRef]
  46. Liu H, Wu H, Yang Z, et al. An historical overview of artificial intelligence for diagnosis of major depressive disorder. Front Psychiatry. 2024;15:1417253. [CrossRef] [Medline]
  47. Dashti M, Azimi T, Khosraviani F, et al. Systematic review and meta-analysis on the accuracy of artificial intelligence algorithms in individuals gender detection using orthopantomograms. Int Dent J. Jun 2025;75(3):2157-2168. [CrossRef] [Medline]
  48. Shen Y, Yu J, Zhou J, Hu G. Twenty-five years of evolution and hurdles in electronic health records and interoperability in medical research: comprehensive review. J Med Internet Res. Jan 9, 2025;27:e59024. [CrossRef] [Medline]
  49. Gilbody S, Sheldon T, House A. Screening and case-finding instruments for depression: a meta-analysis. CMAJ. Apr 8, 2008;178(8):997-1003. [CrossRef] [Medline]
  50. Hassan L, Milton A, Sawyer C, et al. Utility of consumer-grade wearable devices for inferring physical and mental health outcomes in severe mental illness: systematic review. JMIR Ment Health. Jan 7, 2025;12:e65143. [CrossRef] [Medline]


AI: artificial intelligence
AUC: area under the curve
DOR: diagnostic odds ratio
DSM-IV: Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition
DSM-V: Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition
FN: false negative
FP: false positive
GRADE: Grading of Recommendations, Assessment, Development, and Evaluation
HDRS: Hamilton Depression Rating Scale
KNN: k nearest neighbor
MADRS: Montgomery-Åsberg Depression Rating Scale
PHQ-9: Patient Health Questionnaire-9
PITROS: participants, index text, target conditions, reference standard, outcomes, and settings
PRISMA-DTA: Preferred Reporting Items for a Systematic Review and Meta-Analysis of Diagnostic Test Accuracy
PROBAST + AI: Prediction Model Risk of Bias Assessment Tool plus artificial intelligence
RF: random forest
TN: true negative
TP: true positive
UMAP&NN: Uniform Manifold Approximation and Projection and neural network
WHO: World Health Organization


Edited by John Torous; submitted 05.Oct.2025; peer-reviewed by Farshad Khosraviani, Siân Bladon; final revised version received 25.Jan.2026; accepted 28.Jan.2026; published 10.Mar.2026.

Copyright

© Jiawen Liu, Junhui Wang, Zhaobin Wu, Mohamad Ibrani Shahrimin Bin Adam Assim. Originally published in JMIR Mental Health (https://mental.jmir.org), 10.Mar.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.