Review
Abstract
Background: Digital conversational agents (or “chatbots”) that can generate human-like conversations have recently been adapted as a means of administering mental health interventions. However, their development for youth seeking mental health services requires further investigation.
Objective: This youth-engaged scoping review synthesizes the recent research on digital conversational agents for youth seeking mental health or substance use services.
Methods: Studies were included if they were published between 2016 and 2025 and examined digital conversational agents for youth aged 11 to 24 years with mental health or substance use challenges in clinical settings. Systematic literature searches were conducted in February 2024 in multiple databases and updated in March 2025. Data were extracted using codeveloped forms and synthesized narratively.
Results: Ten studies were included, all focusing on mental health. Seven examined the acceptability and feasibility of digital conversational agents; others explored youth perceptions of use, design, and content, with some exploration of impact on mental health symptoms. Eight of ten studies reported high acceptability or positive user experiences. Three were randomized controlled trials that found potential reductions in depressive symptoms. Reporting on the ethical standards was limited. No studies focused on substance use alone.
Conclusions: Literature on digital conversational agents for treatment-seeking youth is emerging but limited. Future rigorous research is needed that prioritizes data security, safety measures, and youth co-design in the development of safe, engaging, digital conversational agents for youth with mental health conditions.
doi:10.2196/77098
Keywords
Introduction
Background
Digital conversational agents, or “chatbots,” are tools that can simulate a human interaction in text format, using artificial intelligence to respond intelligently to user inputs. They may use human-generated algorithms to produce conversations, or natural language processing to recognize and respond to inputs with a programmed personality. Digital conversational agents have been used in mental health for tasks such as providing psychoeducation, building coping skills, delivering cognitive-behavioral therapy, and providing diagnostic support [-].
There is emerging research available on digital conversational agents for mental health or substance use in treatment-seeking populations such as youth who are seeking or accessing services to support their mental health or substance use. A recent set of 2 reviews found minimal research on digital conversational agents for mental illness, with a specific gap around youth, and the studies were not necessarily in treatment-seeking populations [,]. They did find some acceptability and positive effects for the tool’s ability to establish diagnosis, but less so among youth. Some positive mental health impacts were found, but study limitations made these difficult to determine. Digital conversational agents in the psychiatric sphere have targeted various outcomes. For example, one was designed to promote activity engagement among individuals with severe mental illness []. Others have been designed to promote service utilization [], develop mindfulness and emotion regulation [], determine diagnoses [], or promote medication adherence []. Evidence of the development of something akin to a therapeutic alliance has been demonstrated [], while the responses of digital conversational agents have been rated as higher quality and more empathetic than physician responses []. However, the research evidence for treatment-seeking youth in clinical contexts (ie, receiving treatment) remains limited and has yet to be synthesized.
There is more research on the use of digital conversational agents as a prevention, promotion, or low-threshold mental health support tool, particularly for youth in school settings [-]. An acceptability study found that youth preferred to receive information through a digital conversational agent versus searching for it online, while the “personality” of the digital conversational agent was appreciated for its friendly, relational aspects []; the practical advice and referrals offered were also appreciated. Advantages include flexibility, accessibility, the capacity to reach more youth faster than in-person services, cost-effectiveness, and the ability to get help without stigma [,,]. However, evidence on efficacy remains limited [], and the need to empirically test these interventions has been highlighted [,,].
As digital conversational agents expand around the world, a number of ethical issues have been raised. One team set out recommendations for minimal ethical standards in the use of digital conversational agents for youth mental health []. Some factors include ensuring confidentiality, being transparent about uses and limits, generating appropriate crisis responses, and protecting against addictive use. The article further highlighted the importance of testing efficacy, being transparent with users about the evidence base, encouraging in-person services, and explaining that these tools are not designed to treat severe mental illness. For vulnerable youth in treatment-seeking contexts, privacy and confidentiality is important, and addictive use might be a particular concern. These concerns have been reflected in discussions of the promise and necessary cautions regarding the implementation of digital conversational agents [,]. These issues must be attended to in the development, testing, and scale of digital conversational agents for mental health.
Youth are high users of digital technology [] and often use digital means of seeking information about health []. They may be interested in receiving support through digital conversational agent technology. However, treatment-seeking youth with potentially complex mental health challenges have different needs and may have different preferences than school-based samples. Given ethical concerns about digital conversational agents for severe mental illness [], further exploratory work is an essential prerequisite to digital conversational agent development. It is important to gain the perspectives of treatment-seeking youth around the possible utility, functionality, acceptability, and safety of digital conversational agents to address mental health challenges among them. It is also important to continue to do up-to-date work in this area, given the fast pace of technology development that leaves older research quickly out of date.
Objective
This scoping review aims to synthesize the recent literature on digital conversational agents for youth in clinical contexts seeking treatment for mental health or substance use challenges.
Methods
Overview
We used the scoping review methodology given the breadth and nature of the research and the exploratory nature of the research question in an emerging field, using established scoping review processes [,]. The scoping review incorporated the following major steps: (1) defining the research question, (2) identifying relevant studies, (3) screening and selecting studies, (4) extracting the data, and (5) summarizing the data and reporting. PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines have been followed in the reporting of this review () [].
Youth Engagement
In accordance with the strategy for patient-oriented research [], we engaged 2 youth engagement specialists with lived experience in the conduct of this review. We endorsed pragmatism as a fundamental worldview and valued youth contributions, as per lived experience engagement paradigms []. Youth engagement specialists helped identify keywords for the systematic search, coselected information to extract from the resulting studies, and assisted with interpreting and reporting on the findings. Collaboration with the youth engagement specialists helped guard against bias and ensured that the study was relevant to the real-world experiences of young people, enhancing rigor.
Defining the Research Question
This scoping review aims to understand the published academic literature on digital conversational agents for treatment-seeking youth with mental health or substance use challenges, in clinical contexts. Based on the PICO (population, intervention, comparison and outcomes) framework [], this review focuses on youth with mental health or substance use challenges (population) and applications of digital conversational agents in treatment-seeking in clinical settings (intervention). Note that the treatment-seeking and clinical settings context was a key inclusion criterion that makes this review very specific to clinical contexts, which is a gap in the literature. It includes studies with or without comparison groups and reports on the full range of findings, regarding applications, acceptability, and outcomes related to digital conversational agents for the target population, as well as other findings highlighted in the selected literature. Literature on preventive interventions and nonclinical settings was excluded.
Identifying Relevant Studies
A comprehensive search strategy was developed with a health sciences librarian (TR). Considering the subject areas of the research question, the search was developed, tested, and finalized in Medline, then translated and run on February 22, 2024 in the following bibliographic databases: APA PsycInfo, Medline, Embase, Cumulative Index to Nursing and Allied Health Literature (CINAHL), Web of Science, Applied Social Sciences Index and Abstracts (ASSIA), and Cochrane Central Register of Controlled Trials (CENTRAL). The search was repeated and updated on March 18, 2025. The search strategies used database-specific subject headings and keywords in natural language, as well as advanced search operators such as truncation and adjacency operators, to capture 3 main concepts: youth, digital conversational agent, and mental health or substance use. The review was limited to articles published in the past 8 years (2016-2025) due to the rapid advancement of AI technology. There were no limitations placed on language or study type. Conference abstracts were removed through the search strategy when allowed by a database, but were otherwise included. A supplementary hand search of reference lists and Google Scholar was conducted. Given the nature of the research question, the gray literature search was limited to dissertations in APA PsycInfo, registered clinical trials in the CENTRAL search, and nontraditional publications indexed in CINAHL.
Study Screening and Selection
To be included, articles had to present descriptive information or research results on digital conversational agents for youth mental health or substance use, focusing on treatment-seeking populations in clinical contexts. Articles could report on any study design and sample size and could include descriptive process papers. The digital agent and study had to be focused on youth approximately aged 11-24 years, or separate analyses had to be available for a subgroup within this age range. Articles had to be published in English or French consistent with the language skills available on the team, although no non-English language texts were found. We included any outcomes identified by the article in question, including acceptability and feasibility outcomes as well as mental health or substance use outcomes of any kind or intensity. Excluded were any studies not reporting on digital conversational agents for youth mental health or substance use, research conducted on children (majority aged younger than 11 years) or adults (majority aged older than 24 years), and commentaries that did not describe either research results or a specific digital conversational agent. presents inclusion and exclusion criteria, which were developed to optimize the ability to address the research question. Selected articles were screened first at the title and abstract level by 2 of 3 research staff (include or exclude responses), with any conflicts resolved through discussion with the project lead. The resulting set of articles were screened in full using the same process. Article screening was conducted using Covidence software [].
Inclusion criteria
- Addresses digital conversational agents
- Agent designed for youth aged 11-24 years
- Participants with mental health or substance use challenges (disorders or symptoms)
- Participants seeking mental health or substance use treatment (ie, connected with clinical contexts)
- Published between 2016 and 2025
- Reports on efficacy, effectiveness, feasibility, or acceptability
- Reports on youth perspectives using youth-reported data (any method)
Exclusion criteria
- Not a youth-focused agent and study
- Prevention or promotion that does not support known mental health or substance use challenges
- Youth were not treatment-seeking or were not exposed to the digital conversational agent in a clinical context
- Published before 2016
- Papers without original data
Data Extraction
A data extraction form was developed collaboratively among team members, including youth. From the final selected articles, data were extracted by 1 staff member and verified by a second staff member using a spreadsheet. The data extraction form included (1) study general information (eg, reference, location, and data collection years), (2) digital conversational agent description (eg, goal or objective, application and content, context of use, development and testing, youth engagement, software and data strategy, feedback learning integration, delivery method, financial investment, privacy, content monitoring, and sustainability plan), (3) study methods (eg, study objective, design, sampling method, sample size, sample description, exclusion criteria, measures, and other treatments), (4) outcomes (eg, finding overview, youth perspectives, digital conversational agent benefits and limitations, adverse events, and study limitations), (eg, 5) ethical issues (privacy and confidentiality, efficacy, safety, and mitigating strategies), and (eg, 6) other information (author conclusions and recommendations).
Data Summarization and Reporting
The data were summarized narratively and in table format. Quality assessment was not conducted, as this is not a priority in scoping reviews in which a wide range of diverse research is mapped.
Ethical Considerations
Since this is a scoping review of the literature, ethics board approval was not required. There are no human participants in this article and informed consent is not required.
Results
Study Selection
Our search identified 8659 records, including 5875 records from the initial search in 2024 and 2784 from the updated search in 2025. After removing 3734 duplicates, we screened 4925 titles and abstracts, excluding 4900. We further assessed 25 full-text articles. Fifteen were excluded due to not meeting our inclusion criteria, which resulted in 10 studies included in the review. The study selection process is illustrated in the PRISMA flow diagram ().

summarizes the included articles and digital conversational agents studied. The 10 studies examined 9 digital conversational agents for youth mental health in clinical care contexts. Studies emerged from the United States [,,], Germany [,], Australia [,], Thailand [], Japan [], and Hong Kong [], with publication years ranging from 2017 to 2025. Only 3 articles stipulated that the digital conversational agents were co-designed with youth [,,].
| Author (year) | Country | Format | Technology described | Digital conversational agent objective | Intervention modality | Expressed digital conversational agent limitations | Codevelopment with youth |
| Wrightson-Hester et al [] | Australia | Web application | Artificial intelligence conversational agent | Help youth manage mental health using curious questioning | Method of levels therapy | Repetitive content and input comprehension issues | Yes |
| Greenwood et al [] | United States | Mobile app | Natural language processing and machine learning | Help youth develop emotion regulation skills | CBT, IPT-Ab, and DBTc | Effectiveness not rigorously tested for; not implemented in clinical practice | Not specified |
| Beilharz et al [] | Australia | Web-based chat accessible on mobile and desktop devices | Artificial intelligence conversational agent | Provide support for body image issues and eating disorders | Psychoeducation, coping skills via CBT, acceptance and commitment therapy, mindfulness | Limited to psychoeducation and simultaneous coping skills | Yes |
| Fitzpatrick et al [] | United States | Instant messenger app on mobile or desktop | Decision tree methodology with predefined responses, including natural language processing at certain points within the tree | Deliver self-help CBT to young adults with anxiety and depression symptoms. | CBT | Potential risk of violating psychotherapeutic processes related to limitations in conversational abilities | Not specified |
| Robinson et al [] | United States | Mobile app | Not specified | Deliver CBT to adolescents at adolescent outpatient mental health clinic. | CBT | Not specified | Not specified |
| Mahmud and Porntrakoon [] | Thailand | Not specified | AId-generated or DCAe mental health application | No specific digital conversational agent introduced | Not specified | Lack of customization, depersonalization, algorithm bias, privacy concerns | N/Af |
| Kuhlmeier et al [] | Germany | Mobile apps and web platforms | Text-based conversational agent prototype | Provide youth and young adults personalized therapy for depression | CBT, IPTg | Currently an early-stage prototype | Yes |
| Kuhlmeier et al [] | Germany | Prototype accessed via laptop | Text-based conversational agent prototype | Support youth with depression through a behavioral activation exercise | Behavioral activation | Currently an early-stage prototype | No, but participant feedback will influence the future design |
| Hasei et al [] | Japan | Accessible via mobile app LINE | GPT-4 large language model | Help pediatric, youth, and young adult cancer patient reduce anxiety and stress | Not specified | 4-turn conversations | Not specified |
| Li et al [] | Hong Kong | Accessible via Microsoft Teams | An interactive conversational system with natural language understanding | Support university students with self-reported depression through MBSRh | MBSR interventions | Technical issues with Microsoft Teams, required high self-discipline, limited interactivity | Not specified |
aCBT: cognitive behavioral therapy.
bIPT-A: interpersonal psychotherapy for adolescents.
cDBT: dialectical behavior therapy.
dAI: artificial intelligence.
eDCA: digital conversational agent.
fN/A: not applicable.
gIPT: interpersonal therapy.
hMBSR: mindfulness-based stress reduction.
Study Characteristics
Of the 10 included studies, 7 assessed feasibility or usability or the preliminary efficacy or effectiveness [,,,,,], while others investigated perceptions of using the agent [], perceptions of their design [], and youth’s interaction experience with them []. Three studies used qualitative methods [,,], 1 used quantitative methods [], and 6 used mixed methods [,,-]. Three studies used randomized controlled trials to evaluate 2 digital conversational agents [,,]. Participant demographics varied in the studies, covering age groups from adolescents to young adults with depression [,,], anxiety or depression [,,,,], anxiety and stress [], and eating disorders []. None focused on substance use disorders. Some studies also included mental health professionals [,] and caregivers [] as participants.
Intervention Characteristics
The technology driving the digital conversational agent was not always specified but included artificial intelligence, for example natural language processing, machine learning, and decision trees. While publications were dated 2016 to 2025, the year of the development of the digital conversational agent under examination was not reported. The interventions were delivered via a mobile app [,-,,] or a computer browser [,-]. Treatment modalities included cognitive behavioral therapy or components thereof [,,-]; an integration of cognitive behavioral therapy, interpersonal psychotherapy, and dialectical behavior therapy []; a method-of-levels approach []; mindfulness-based stress reduction interventions []; and unspecified modalities [,]. Some studies included screenshots or sample interactions to illustrate the digital conversational agent’s responses [,-,]. For example, one study shows a screenshot of coping skill recommendations made by the chatbot []. Another shows a screenshot of behavioral activation recommendations []. However, others did not provide details about the exact nature of the responses of the digital conversational agent. The digital conversational agents were not generally described as being integrated into a system of care to accompany clinical services, except for one that was designed for young cancer patients [].
Qualitative and Acceptability Findings
The methodology and outcomes of the included articles are summarized in . Most studies reported that the digital conversational agents were acceptable to youth, with high utilization [,], high satisfaction [], and overall positive perceptions [,,-]. Participants appreciated the ease of use [], appealing and accessible design [,], and friendly characters [,]. They also found that the content was helpful and appropriate [,,], equipped them with new mental health insight [], and offered them empathy [,]. Youth identified advantages, namely the neutrality, anonymity, and availability of these tools compared to traditional psychotherapy [,,]; some youth indicated their willingness to use digital conversational agents for depression and anxiety, with cautions regarding privacy and security [].
| Author (year) | Study objective | Sample size | Sample age (years) | Mental health or substance use conditions | Exclusion criteria | Measurement | Amount of use of digital conversational agent within study | Outcome summary |
| Wrightson-Hester et al [] | Assess feasibility and acceptability | 13 | 16-24 | Anxiety or depression | Severe depressive symptoms or frequent suicidal thoughts | Self-report measures of depression, anxiety, QoLa, related psychological constructs, and system usability and engagement Interview/focus group guide | Week 1: 1-4 days, 1-3 chats/day (average length 2-30 min); Week 2: 1-7 days, 1-5 chats/day (average length 5-15 min) | Feasible assessments; acceptable system engagement; acceptable interface; acceptable therapeutic conversations; some frustration, but resolved through use; recommendations for improvement; possible improvements in some psychological measures |
| Greenwood et al [] | Establish the feasibility, preliminary acceptability, effectiveness, and usability | 18 (17 analyzed) | 13-17 | Depression and anxiety | Parent-reported history of severe depression, substance use, psychotic illness, OCDb, PTSDc, panic disorder, or specific phobias; psychiatric hospitalization in the previous month; unaccompanied by a guardian at diagnostic visit; no access to a mobile device for regular use; unable to read and write English. | Self-report measures of depression (primary outcome), anxiety, and self-efficacy; app usage rates; acceptability, feasibility, and usability; safety monitoring. | Average use of 6 days, 55 mood check-ins, 313.17 sent messages, 13.63 lessons, and 9 tools over 4 weeks | Acceptability and feasibility confirmed; decreased depression and anxiety; increased self-efficacy; satisfactory usage rates; acceptable safety monitoring. |
| Beilharz et al [] | Evaluate the acceptability and feasibility | 25 (17 youth, 8 caregivers) | Youth: 13-18; caregivers: 46-57 | Body image concerns or eating disorders | Not specified | Focus group guide | Not specified | Appealing character and design; preference for brief content with a conversational tone; acceptable flow via button-based navigation; preference for free-text option; appreciation of accessibility. |
| Fitzpatrick et al [] | Assess feasibility, acceptability, and preliminary efficacy | 70 | 18-28 | Anxiety and depression | Not specified | Self-report measures of depression, anxiety, affect, and acceptability and usability; open-ended questions; usage tracking | Up to 20 sessions, average 12.14 check-ins over 2 weeks | Feasibility confirmed; reduction in depression symptoms, but not anxiety or positive or negative affect; high engagement and satisfaction; some positive comments on process and content; some challenges with process, technology, and content; limitations existed. |
| Robinson et al [] | Explore the feasibility and noninferiority | 141 | 13-17 | Depression or anxiety | Not specified | Self-report measure of depression and anxiety | Not specified | Noninferiority to clinician-delivered CBTd-skills group in reducing depression symptoms. |
| Mahmud and Porntrakoon [] | Explore perspectives on DCAse as alternative treatments | 15 (10 youth, 5 psychologists) | Youth: 20-25 | Depression or anxiety | Not specified | Interview guide; self-report measures of depression and anxiety. | Not specified | Interest in using DCA when experiencing depression or anxiety; concerns about privacy, confidentiality, and security; preferences varied on the extent and time of expert assistance |
| Kuhlmeier et al [] | Design a personalized conversational agent to treat youth depression | Problem awareness phase: 15 youth; evaluation phase: 5 experts, 5 potential users | Problem awareness phase: 14-17; evaluation phase: experts: mean 29; potential users: mean 24 | Depression | Not specified | Interview guide | Not specified | Personalization of the conversational agent and content is essential, combined with structure; areas of improvement identified |
| Kuhlmeier et al [] | Investigate how youth interact with the conversational agent for depression | 15 | 14-17 | Depression | Suicidal ideation, psychosis, or low cognitive functioning. | User interaction data and think-aloud sessions documenting thoughts about the DCA | 11:54-20:38 min | Engaging interaction style; preferences varied on emoji and language use; appreciation of predefined responses, optional reminders, and clear interface; varied perceptions of responding speed and the dialogue content |
| Hasei et al [] | Assess feasibility and potential impact in pediatric, youth, and young adult cancer patients | 5 | 13 to 20s | Not specified, but cancer patients usually experience anxiety and stress from cancer | Not specified | Self-report measure of anxiety and stress; open-ended questions. | Most used every 2-3 days for about 10-15 min/session over 1-2 weeks | Reduced anxiety and stress; possible increased cancer treatment engagement; possible increased self-expression and informal help-seeking. |
| Li et al [] | Evaluate feasibility, acceptability, safety, and preliminary efficacy | 30 (27 completed) | 18-25 | Depression | Diagnosed with a clinical psychotic condition preintervention or were currently involved in any mindfulness-based or other psychosocial interventions. | Self-report measure of depression, stress, anxiety, and mindfulness; open-ended questions. | 8 weeks; encouraged at least 1 hour of daily interaction. | Feasibility, acceptability, safety, and preliminary efficacy confirmed; limitations identified. |
aQoL: quality of life.
bOCD: obsessive compulsive disorder.
cPTSD: posttraumatic stress disorder.
dCBT: cognitive behavioral therapy.
eDCA: digital conversational agent.
A number of problems, challenges, or barriers were identified by participants across the studies. These included limited natural conversation [,,], limited customization [,,,], technical glitches [,], and concerns regarding the quality of the therapeutic interventions [,,]
Quantitative Findings
Six studies provided quantitative results on the digital conversational agents’ ability to improve participants’ mental health. A randomized controlled trial (RCT) of a digital conversational agent called Woebot found that, compared to a self-help e-book, the intervention significantly reduced depressive symptoms after 2 weeks, with a moderate effect size (d=0.44) []. In another RCT, 4 weeks of using Woebot was found to be noninferior to a standard clinician-led CBT group in reducing depression when comparing pre-post average scores and the number of participants with elevated depression; the effect size was not provided []. In a pilot RCT, the treatment group showed greater mean improvement in depression than the waitlist control group (d=0.98) []. No significance testing was reported for these trials [,]. A single arm study found a large effect size (d=1.26) for reducing problem-related distress and a medium effect size (d=–0.66) for improving goal conflict reorganization among participants with anxiety or depression []. Additionally, a single arm pretest-posttest study of a digital conversational agent delivering a mindfulness intervention reported significant and large improvements in depression, anxiety, stress, and mindfulness levels, with large effect sizes across all outcomes, at d=–1.95 for stress, d =–1.53 for depression, and d=–1.09 for anxiety [].
Ethical and Safety Considerations
The minimal ethical standards proposed for digital conversational agents revolve around (1) privacy and confidentiality, (2) efficacy, and (3) safety []. One study encompassed these requirements by providing user participants with an information package regarding privacy, efficacy, and safety issues as part of the application onboarding process []. Another study emphasized rigorous approval and safety measures that include anonymized and secure data handling and adherence to the Declaration of Helsinki []. However, most of the included articles did not discuss ethical issues in depth. While concerns over privacy and confidentiality were mentioned by many youths and stakeholders [,,,], only half of the articles included any information on their privacy policies. A case series assessing feasibility and acceptability for depression or anxiety reported that secure user identification and anonymized data collection was not conducted within the study, but would be implemented in the future []. One study informed participants that they would not be monitored by a counsellor [], while another disclosed that a psychologist would monitor participants’ activities asynchronously [].
Despite the ethical importance of ensuring efficacy, most of the articles described pilot trials examining acceptability and feasibility without adequate power to determine efficacy. Prominent limitations were small sample sizes and the lack of diversity. As a result, while many of the interventions were reportedly inspired by evidence-based therapeutic approaches [,,,,], the delivery of the interventions via these tools has not been demonstrated efficacious or effective in real-world settings and thus requires further evaluation.
Safety issues are also worthy of note. Only 6 of the selected studies described safety features [,,,,,]. Among these, 3 reported that participants were informed of about the tool and given explanation of its function and capacities [,,]. Only 1 study mentioned having a crisis response measure, in which the digital conversational agent could identify concerning language and confirm it with users []. One study emphasized that their safety measures include automated escalation systems for distress detection and professional involvement during development and testing []. Another study involved continuous monitoring by mental health professionals, weekly participants check-ins, and mental health support resources and referral pathways []. The digital conversational agents featured in these 4 trials offered additional mental health resources and provided easily accessible channels to connect users with suicide helplines or emergency services.
Discussion
Principal Findings
This review synthesized the recent literature on digital conversational agents for youth with mental health or substance use conditions, specifically in treatment-seeking clinical contexts. It was found that the literature is quite preliminary and is focused entirely on mental health, not substance use conditions. Yet, this is an emerging area of work that is a high priority to some research teams and health care organizations. Several of the digital conversational agents reviewed applied the principles of cognitive behavioral therapy, but this was not universal. Promising findings were identified in terms of feasibility and acceptability, together with reasonable usage rates and positive youth perspectives of acceptability. Preliminary findings suggested that digital conversational agents might be an acceptable means of delivering therapeutic content to treatment-seeking youth, with positive impacts on their mental health.
The literature was largely made up of feasibility and acceptability studies rather than efficacy trials. The digital conversational agents were not described as being embedded in intervention contexts, such as being integrated with in-person clinical care. Sample sizes were small and diversity was limited. People with different sociodemographic profiles and different clinical needs might be differentially interested in using e-health innovations [,], although differences across population subgroups have not been definitively established. This leaves the evidence base in a very premature state. For example, there is a lack of exploration of the impact the tools might have as an integrated component of youth mental health care among diverse youth.
All of the articles addressed mental health; none addressed substance use. It is important to consider mental health and substance use together, given the high levels of overlap and increasing trends toward treating concurrent disorders together []. Examining digital conversational agents for substance use challenges, or for concurrent mental health and substance use challenges, is a pressing area for future work. Investigations into suitability using different treatment modalities for specific diagnostic groups might also be worthy of consideration, given that specific symptoms may interact with whether and how one might use such a tool (eg, low motivation, paranoia, suicidal ideation, and mood instability).
This review focused on very recent literature on digital conversational agents, given the fast pace of technology. However, any article published in 2016 was likely examining a digital conversational agent developed at the latest in 2015, potentially earlier. These tools were not leveraging today’s latest technology, and the technology used was not thoroughly described. The rapid pace of technology advancement is not a new challenge, but rather one that the health care sector has been grappling with for years []. This leaves two problems: (1) the technology is already out of date by the time the research is complete, and (2) the evidence base that a new study leverages is out of date and may no longer be relevant by the time it is used. Ongoing rapid research and publication in this area is essential, with thorough descriptions of the technology underpinning the tool, including the year of tool development to help readers understand the state of technological advancement of the digital conversational agent in question.
In digital intervention development, co-design with the end users is critical []. This can be done using appropriate patient engagement frameworks []. Co-design helps ensure that the digital tool is appropriate for the end users and relevant to their needs. However, multiple articles reviewed did not discuss co-design, which may reduce acceptability. Future digital design teams should adhere to patient engagement principles for e-health innovation [] as well as youth engagement principles []. Co-design has the potential to be a meaningful opportunity for young people to gain experience in technology development and clinical research, supporting their personal and professional growth and development []. In conducting co-design, teams should incorporate principles of equity, diversity, inclusion and accessibility to ensure that a diverse range of voices are included, to generate a digital conversational agent that is appropriate for a wide range of diverse youth, and to mitigate the harms that can come from biases, discrimination, and inequities. Specifically, a lack of diversity at the co-design stage could perpetuate algorithmic bias, which should be guarded against. Engagement can also be leveraged to increase digital literacy and the use of resulting tools, potentially with the support of digital navigators [].
From our review, it was clear that the minimum ethical standards for digital conversational agents were not adequately attended to []. Few studies reported directly on privacy, confidentiality, and safety, and efficacy has not been definitively established. We call on developers and trialists to directly attend to the proposed ethical standards in their initiatives, proactively integrate these minimum standards into their technology and trials, integrate an ethical review stage into their initiative, and report clearly on the use of these standards. The ethics of delivering evidence-based intervention modalities via digital conversational agents should be considered—and when which modality is appropriate and for whom—since a clinician may naturally adapt their services to the individual youth, while digital tools may function differently using artificial intelligence in attempt to do the same. Matching the tool’s aim to the youth’s need, ensuring youth can withdraw at any time, and offering access to a person to talk to would all further enhance tools’ ethics and should be considered by interventionists. It should be transparent that a digital conversational agent can supplement clinical care but is not a replacement for it; integrating tools into clinical care rather than leaving them as stand-alone tools might address this challenge [].
The use of digital conversational agents is not without risk. A recent news story highlights a case of youth suicide after the youth’s ongoing interaction with a digital conversational agent []. It should be noted that youth are using digital conversational agents for mental health support even if these agents are not designed for that purpose. By developing a digital conversational agent with and for youth, with the explicit purpose of supporting mental health in appropriate ways using evidence-based content, it may be possible to provide young people with a safer tool and prevent such tragedies from occurring.
While some articles discussed soliciting youth feedback as part of a study, none discussed collecting youth feedback within the digital platform. Providing opportunities to submit feedback within the platform is a potential mechanism to enable continuous improvement and incremental alignment with the needs and preferences of the users. The feedback could be directed to programming teams for programming feedback, and potentially separately to clinical teams to support clinical care as part of an integrated treatment plan. It would be important to determine means to balance the importance of privacy and confidentiality [] with the potential for identified clinical support.
The next steps in this line of work would be to address some of the limitations, such as co-design and ethical concerns, with the ultimate goal of moving forward to full-scale trials to examine efficacy. This would have to be followed by implementation and ongoing refinement. The evaluation of complex interventions is expected to begin with iterative stages of intervention development, feasibility testing, evaluation, and implementation [,], and the research for this type of intervention is in the early stages of this process.
Limitations
Strengths and limitations of this review should be kept in mind. Important partners were engaged in the review, including industry partners and youth with lived experience. These partnerships ensured that the search terms, data extraction, and interpretations are relevant to the artificial intelligence industry and youth. However, readers should note the methodological and scope limitations: the search included only an 8-year period to ensure that the findings are as relevant as possible to recent technology, but may have missed early findings. We excluded any articles focusing on prevention or promotion, which may have relevant findings. Any publications indexed after March 2025 were not captured. The search was run only in English-based databases.
Conclusions
Emerging but preliminary literature is examining digital conversational agents for treatment-seeking youth with mental health conditions, but not substance use challenges. Digital technology already exists and will inevitably expand, making it essential that we support its expansion with rigorous research. This research must therefore continue, with attention to establishing efficacy, as well as ensuring appropriate protection of privacy and confidentiality and attention to safety and ethical concerns. While digital technology comes with challenges, it also comes with potential benefits. Co-design with youth, for youth, should be a priority. Nevertheless, promising early research justifies continuing this work with the possibility of codeveloping safe, accessible, youth-friendly, and potentially efficacious tools for youth mental health.
Acknowledgments
We would like to thank Shelby McKee for support with the article selection process. The authors disclosed receipt of the following financial support for the research, authorship, or publication of this article: this work was supported by the Ontario Brain Institute.
Data Availability
The data underlying this project are included herein. Additional data are available upon request.
Authors' Contributions
LDH conceptualized and led the project and drafted the manuscript. JH contributed to selecting the articles, extracting and verifying the data, and drafting the article. ATPN contributed to selecting the articles and extracting and verifying the data. TR formulated and conducted the search. All authors contributed to the design of the study and edited and approved the final manuscript.
Conflicts of Interest
Brian Ritchie is the founder and CEO of Kamazooie Development Corporation (“Kama.AI”), a private company specializing in artificial intelligence, including digital conversational agents. The other authors have no conflicts of interest to disclose.
PRISMA-SCR checklist.
PDF File (Adobe PDF File), 166 KBReferences
- Beilharz F, Sukunesan S, Rossell SL, Kulkarni J, Sharp G. Development of a positive body image chatbot (KIT) with young people and parents/carers: qualitative focus group study. J Med Internet Res. Jun 16, 2021;23(6):e27807. [FREE Full text] [CrossRef] [Medline]
- Fitzpatrick KK, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment Health. Jun 06, 2017;4(2):e19. [FREE Full text] [CrossRef] [Medline]
- Vaidyam AN, Linggonegoro D, Torous J. Changes to the psychiatric chatbot landscape: a systematic review of conversational agents in serious mental illness: changements du paysage psychiatrique des chatbots: une revue systématique des agents conversationnels dans la maladie mentale sérieuse. Can J Psychiatry. Apr 2021;66(4):339-348. [FREE Full text] [CrossRef] [Medline]
- Vaidyam AN, Wisniewski H, Halamka JD, Kashavan MS, Torous JB. Chatbots and conversational agents in mental health: a review of the psychiatric landscape. Can J Psychiatry. Jul 2019;64(7):456-464. [FREE Full text] [CrossRef] [Medline]
- Liu HT. Development of a chatbot to promote activity participation for people with serious mental illness. University of Alberta [dissertation]. University of Alberta; 2021. URL: https://ualberta.scholaris.ca/items/dacf408d-89af-453c-bb51-f146e62d3af8 [accessed 2025-09-24]
- Shah J, DePietro B, D'Adamo L, Firebaugh M, Laing O, Fowler LA, et al. Development and usability testing of a chatbot to promote mental health services use among individuals with eating disorders following screening. Int J Eat Disord. Sep 2022;55(9):1229-1244. [FREE Full text] [CrossRef] [Medline]
- Denecke K, Vaaheesan S, Arulnathan A. A mental health chatbot for regulating emotions (SERMO) - concept and usability test. IEEE Trans Emerg Topics Comput. Jul 1, 2021;9(3):1170-1182. [CrossRef]
- Philip P, Micoulaud-Franchi J, Sagaspe P, Sevin ED, Olive J, Bioulac S, et al. Virtual human as a new diagnostic tool, a proof of concept study in the field of major depressive disorders. Sci Rep. Feb 16, 2017;7:42656. [FREE Full text] [CrossRef] [Medline]
- Bickmore TW, Puskar K, Schlenk EA, Pfeifer LM, Sereika SM. Maintaining reality: relational agents for antipsychotic medication adherence. Interact Comput. Jul 2010;22(4):276-288. [CrossRef]
- Darcy A, Daniels J, Salinger D, Wicks P, Robinson A. Evidence of human-level bonds established with a digital conversational agent: cross-sectional, retrospective observational study. JMIR Form Res. May 11, 2021;5(5):e27868. [CrossRef] [Medline]
- Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. Jun 01, 2023;183(6):589-596. [FREE Full text] [CrossRef] [Medline]
- Høiland CG, Følstad A, Karahasanovic A. Hi, can I help? Exploring how to design a mental health chatbot for youths. Hum Technol. Aug 31, 2020;16(2):139-169. [CrossRef]
- Grové C. Co-developing a mental health and wellbeing chatbot with and for young people. Front Psychiatry. 2020;11:606041. [FREE Full text] [CrossRef] [Medline]
- Dekker I, De Jong EM, Schippers MC, De Bruijn-Smolders M, Alexiou A, Giesbers B. Optimizing students' mental health and academic performance: AI-enhanced life crafting. Front Psychol. 2020;11:1063. [FREE Full text] [CrossRef] [Medline]
- Lin APC, Trappey CV, Luan C, Trappey AJC, Tu KLK. A test platform for managing school stress using a virtual reality group chatbot counseling system. Appl Sci. Sep 29, 2021;11(19):9071. [CrossRef]
- Abd-Alrazaq AA, Rababeh A, Alajlani M, Bewick BM, Househ M. Effectiveness and safety of using chatbots to improve mental health: systematic review and meta-analysis. J Med Internet Res. Jul 13, 2020;22(7):e16021. [FREE Full text] [CrossRef] [Medline]
- Kretzschmar K, Tyroll H, Pavarini G, Manzini A, Singh I, NeurOx Young People’s Advisory Group. Can your phone be your therapist? Young people's ethical perspectives on the use of fully automated conversational agents (chatbots) in mental health support. Biomed Inform Insights. 2019;11:1178222619829083. [FREE Full text] [CrossRef] [Medline]
- Hamdoun S, Monteleone R, Bookman T, Michael K, Michael K. AI-based and digital mental health apps: balancing need and risk. IEEE Technol Soc Mag. Mar 2023;42(1):25-36. [CrossRef]
- Stade EC, Stirman SW, Ungar LH, Boland CL, Schwartz HA, Yaden DB, et al. Large language models could change the future of behavioral healthcare: a proposal for responsible development and evaluation. Npj Ment Health Res. Apr 02, 2024;3(1):12. [FREE Full text] [CrossRef] [Medline]
- Internet use by province and age group. Statistics Canada. 2023. URL: https://www150.statcan.gc.ca/t1/tbl1/en/tv.action?pid=2210013501 [accessed 2025-09-24]
- Selected online activities by gender, age group and highest certificate, diploma or degree completed. Statistics Canada. 2023. URL: https://www150.statcan.gc.ca/t1/tbl1/en/tv.action?pid=2210013701 [accessed 2025-09-24]
- Munn Z, Peters MDJ, Stern C, Tufanaru C, McArthur A, Aromataris E. Systematic review or scoping review? Guidance for authors when choosing between a systematic or scoping review approach. BMC Med Res Methodol. Nov 19, 2018;18(1):143. [CrossRef] [Medline]
- Bougioukas KI, Pamporis K, Vounzoulaki E, Karagiannis T, Haidich A. Types and associated methodologies of overviews of reviews in health care: a methodological study with published examples. J Clin Epidemiol. Jan 2023;153:13-25. [CrossRef] [Medline]
- Page M, McKenzie J, Bossuyt P, Boutron I, Hoffmann T, Mulrow C, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. PLoS Med. Mar 2021;18(3):e1003583. [FREE Full text] [CrossRef] [Medline]
- Strategy for patient-oriented research - patient engagement framework. Canadian Institutes of Health Research. URL: https://cihr-irsc.gc.ca/e/48413.html [accessed 2023-06-07]
- Allemang B, Sitter K, Dimitropoulos G. Pragmatism as a paradigm for patient-oriented research. Health Expect. Feb 2022;25(1):38-47. [FREE Full text] [CrossRef] [Medline]
- Huang X, Lin J, Demner-Fushman D. Evaluation of PICO as a knowledge representation for clinical questions. AMIA Annu Symp Proc. 2006;2006:359-363. [FREE Full text] [Medline]
- Covidence. URL: https://www.covidence.org/ [accessed 2025-09-24]
- Nicol G, Wang R, Graham S, Dodd S, Garbutt J. Chatbot-delivered cognitive behavioral therapy in adolescents with depression and anxiety during the COVID-19 pandemic: feasibility and acceptability study. JMIR Form Res. Nov 22, 2022;6(11):e40242. [FREE Full text] [CrossRef] [Medline]
- Robinson A, Williams A, Rapoport S, Birch A, Lang L, Wells NK, et al. 4.17 Noninferiority of a relational agent, Woebot, to reduce symptoms of depression relative to CBT skills group: a randomized controlled trial among adolescents seeking outpatient mental health care. J Am Acad Child Adolesc Psychiatry. Oct 2023;62(10):S238. [CrossRef]
- Kuhlmeier F, Gnewuch U, Lüttke S, Brakemeier E, Mädche A. A personalized conversational agent to treat depression in youth and young adults – a transdisciplinary design science research project. In: Drechsler A, Gerber A, Hevner A, editors. The Transdisciplinary Reach of Design Science Research. Cham, Switzerland. Springer; 2022.
- Kuhlmeier F, Gnewuch U, Metelmann L, Bauch L, Lüttke S. Exploring user experience with a conversational agent to treat depression in youth: a think-aloud study. 2022. Presented at: The 21st Annual Pre-ICIS Workshop on HCI Research in MIS; December 11, 2022; Copenhagen, Denmark. URL: https://publikationen.bibliothek.kit.edu/1000175413
- Wrightson-Hester A, Anderson G, Dunstan J, McEvoy PM, Sutton CJ, Myers B, et al. An artificial therapist (Manage Your Life Online) to support the mental health of youth: co-design and case series. JMIR Hum Factors. Jul 21, 2023;10:e46849. [CrossRef] [Medline]
- Mahmud T, Porntrakoon P. The use of AI chatbots in mental healthcare for university students in Thailand: a case study. 2023. Presented at: 7th International Conference on Business and Information Management; August 18-20, 2023; Bangkok, Thailand. [CrossRef]
- Hasei J, Hanzawa M, Nagano A, Maeda N, Yoshida S, Endo M, et al. Empowering pediatric, adolescent, and young adult patients with cancer utilizing generative AI chatbots to reduce psychological burden and enhance treatment engagement: a pilot study. Front Digit Health. 2025;7:1543543. [FREE Full text] [CrossRef] [Medline]
- Li Y, Chung TY, Lu W, Li M, Ho YWB, He M, et al. Chatbot-based mindfulness-based stress reduction program for university students with depressive symptoms: intervention development and pilot evaluation. J Am Psychiatr Nurses Assoc. 2025;31(4):398-411. [CrossRef] [Medline]
- Greenwood K, Robertson S, Vogel E, Vella C, Ward T, McGourty A, et al. SlowMo PatientPublic Involvement (PPI) team. The impact of patient and public involvement in the SlowMo study: reflections on peer innovation. Health Expect. Feb 2022;25(1):191-202. [FREE Full text] [CrossRef] [Medline]
- Wilkowsk W, Ziefle M, AlagÖZ F. How user diversity and country of origin impact the readiness to adopt E-health technologies: an intercultural comparison. Work. Jan 2012;41(S1):2072-2080. [CrossRef]
- Renner J, Täuber L, Nieder TO. Need for inclusive consideration of transgender and gender diverse people in e-health services: a systematic review. J Clin Med. Feb 18, 2022;11(4):1090. [FREE Full text] [CrossRef] [Medline]
- Henderson JL, Chaim G, Hawke LD, National Youth Screening Project Network. Screening for substance use and mental health problems in a cross-sectoral sample of Canadian youth. Int J Ment Health Syst. 2017;11:21. [CrossRef] [Medline]
- Stevens A, Milne R, Lilford R, Gabbay J. Keeping pace with new technologies: systems needed to identify and evaluate them. BMJ. Nov 13, 1999;319(7220):1291. [FREE Full text] [CrossRef] [Medline]
- Barony Sanchez RH, Bergeron-Drolet L, Sasseville M, Gagnon M. Engaging patients and citizens in digital health technology development through the virtual space. Front Med Technol. 2022;4:958571. [FREE Full text] [CrossRef] [Medline]
- Leung K, Lu-McLean D, Kuziemsky C, Booth RG, Collins Rossetti S, Borycki E, et al. Using patient and family engagement strategies to improve outcomes of health information technology initiatives: scoping review. J Med Internet Res. Oct 08, 2019;21(10):e14683. [FREE Full text] [CrossRef] [Medline]
- Hawke LD, Relihan J, Miller J, McCann E, Rong J, Darnay K, et al. Engaging youth in research planning, design and execution: practical recommendations for researchers. Health Expect. Dec 2018;21(6):944-949. [CrossRef] [Medline]
- Sheikhan NY, Kuluski K, McKee S, Hiebert M, Hawke LD. Exploring the impact of engagement in mental health and substance use research: a scoping review and thematic analysis. Health Expect. Oct 2023;26(5):1806-1819. [FREE Full text] [CrossRef] [Medline]
- Wisniewski H, Torous J. Digital navigators to implement smartphone and digital tools in care. Acta Psychiatr Scand. Apr 2020;141(4):350-355. [FREE Full text] [CrossRef] [Medline]
- Rodriguez-Villa E, Rauseo-Ricupero N, Camacho E, Wisniewski H, Keshavan M, Torous J. The digital clinic: implementing technology and augmenting care for mental health. Gen Hosp Psychiatry. 2020;66:59-66. [FREE Full text] [CrossRef] [Medline]
- Duffy C. 'There are no guardrails.' This mom believes an AI chatbot is responsible for her son’s suicide. CNN. 2024. URL: https://www.cnn.com/2024/10/30/tech/teen-suicide-character-ai-lawsuit [accessed 2025-09-25]
- Skivington K, Matthews L, Simpson SA, Craig P, Baird J, Blazeby JM, et al. A new framework for developing and evaluating complex interventions: update of Medical Research Council guidance. BMJ. Sep 30, 2021;374:n2061. [FREE Full text] [CrossRef] [Medline]
- Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M, et al. Medical Research Council Guidance. Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ. Sep 29, 2008;337:a1655. [FREE Full text] [CrossRef] [Medline]
Abbreviations
| ASSIA: Applied Social Sciences Index and Abstracts |
| CENTRAL: Cochrane Central Register of Controlled Trials |
| CINAHL: Cumulative Index to Nursing and Allied Health Literature |
| PICO: population, intervention, comparison and outcomes |
| PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews |
| RCT: randomized controlled trial |
Edited by J Torous; submitted 07.May.2025; peer-reviewed by X Jiang; comments to author 26.Aug.2025; revised version received 28.Aug.2025; accepted 29.Aug.2025; published 07.Nov.2025.
Copyright©Lisa D Hawke, Jingyi Hou, Anh T P Nguyen, Thalia Phi, Jamie Gibson, Brian Ritchie, Gillian Strudwick, Terri Rodak, Louise Gallagher. Originally published in JMIR Mental Health (https://mental.jmir.org), 07.Nov.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.

