Published on in Vol 6, No 10 (2019): October

Preprints (earlier versions) of this paper are available at, first published .
Conversational Agents in the Treatment of Mental Health Problems: Mixed-Method Systematic Review

Conversational Agents in the Treatment of Mental Health Problems: Mixed-Method Systematic Review

Conversational Agents in the Treatment of Mental Health Problems: Mixed-Method Systematic Review


Division of Psychology & Mental Health, School of Health Sciences, Faculty of Biology, Medicine and Health, University of Manchester, Manchester, United Kingdom

Corresponding Author:

Hannah Gaffney, BSc, MSc

School of Health Sciences, Faculty of Biology, Medicine and Health

University of Manchester

2nd Floor

Zochonis Building

Manchester, M13 9PL

United Kingdom

Phone: 44 161 306 0400


Background: The use of conversational agent interventions (including chatbots and robots) in mental health is growing at a fast pace. Recent existing reviews have focused exclusively on a subset of embodied conversational agent interventions despite other modalities aiming to achieve the common goal of improved mental health.

Objective: This study aimed to review the use of conversational agent interventions in the treatment of mental health problems.

Methods: We performed a systematic search using relevant databases (MEDLINE, EMBASE, PsycINFO, Web of Science, and Cochrane library). Studies that reported on an autonomous conversational agent that simulated conversation and reported on a mental health outcome were included.

Results: A total of 13 studies were included in the review. Among them, 4 full-scale randomized controlled trials (RCTs) were included. The rest were feasibility, pilot RCTs and quasi-experimental studies. Interventions were diverse in design and targeted a range of mental health problems using a wide variety of therapeutic orientations. All included studies reported reductions in psychological distress postintervention. Furthermore, 5 controlled studies demonstrated significant reductions in psychological distress compared with inactive control groups. In addition, 3 controlled studies comparing interventions with active control groups failed to demonstrate superior effects. Broader utility in promoting well-being in nonclinical populations was unclear.

Conclusions: The efficacy and acceptability of conversational agent interventions for mental health problems are promising. However, a more robust experimental design is required to demonstrate efficacy and efficiency. A focus on streamlining interventions, demonstrating equivalence to other treatment modalities, and elucidating mechanisms of action has the potential to increase acceptance by users and clinicians and maximize reach.

JMIR Ment Health 2019;6(10):e14166




Conversational agents are software programs that use artificial intelligence to simulate a conversation with a user through written text or voice. Recent everyday examples include digital assistants such as Siri (Apple), Cortana (Microsoft), Google Now, and Alexa (Amazon) [1]. The first conversational agent of this kind was ELIZA [2], which was programmed to mimic conversation with a Rogerian psychotherapist using typed text. In the 50 years since ELIZA, interest in conversational agents and artificial intelligence has waxed and waned, and this is reflected in publication rates over time [3]. However, significant advances in technology over the past 2 decades have facilitated the design of conversational agents that can undertake evermore complex tasks [4]. This has resulted in an explosion of publications in this area, particularly since 2009 [3].

Evidence has begun to accumulate around the potential benefits of conversational agents in diverse fields [5] within health and medical care [6] and specifically in mental health [7-11]. Increased access to information through the internet and mobile phones has highlighted the potential for conversational agents to provide autonomous, interactive, and crucially accessible mental health support. Existing digital therapies have suffered from low adherence and concerns about their efficiency without continued human support [12,13]. Existing digital therapy formats tend to focus on psychoeducation and a modular style of fixed content and duration that is inflexible for users. Conversational agents hold particular promise compared with other digital mental health interventions as they can provide greater interactivity that emulates therapeutic conversation and provides choice and control over session content and intensity. Research has demonstrated that users respond and connect to conversational agents in social ways, and they can encourage honest disclosure [14,15]. They also have potential for greater scalability compared with other therapy modalities such as human therapists, Wizard of Oz programs (where a therapist responds via a computer), or digital interventions that require ongoing support from a clinician to produce favorable outcomes.

The application of conversational agents in mental health is varied and includes diagnostic tools, symptom monitoring, and treatment or intervention [16]. Existing systematic and scoping reviews of conversational agent interventions in the mental health field have focused on a subset of conversational agents with a visual character (embodied) [8-10] or are now outdated [7]. As far as we are aware, this is the first comprehensive systematic review of conversational agents in the treatment of mental health problems.


We conducted a systematic review and synthesis of conversational agents in the treatment of mental health problems. Conversational agents are diverse in design [1] and include, for example, chatbots (eg, casual conversation delivered verbally or through text), embodied conversational agents (ECAs; a virtual visual character that simulates human style, face-to-face conversation with gestures, and nonverbal behavior), conversational agents with a physical presence (eg, robots), and conversational agents within virtual reality (VR). For this systematic review, studies that included an automated conversational agent that simulated a 2-way, real-time conversation, with text or verbal based input (either fixed response options or free text) and an independent (not supported by a human) stand-alone system were included. Studies that used Wizard of Oz methods, where a person or therapist responds through the computer or programs that required the ongoing support from a therapist or similar, were excluded. We followed the Preferred Reporting Items for Systematic Review and Meta-Analysis Protocols guidelines [17,18]. The protocol was registered prospectively at PROSPERO (registration number: CRD42018106652).

Literature Search

A systematic search of the literature was performed in September 2018 and updated in January 2019 using MEDLINE (1946 to August week 5, 2018), EMBASE (1974 to September 2018), PsycINFO (1806 to September 2018), Web of Science (1900 to September 2018), and the Cochrane library (All to September 2018). The search was not restricted by publication year or language. Overall, 3 categories of search terms were included: (1) relational agent, (2) mental health, and (3) intervention. The Boolean operator AND was used to bring together separate categories and OR was used to combine terms within categories. Keywords were collated from the existing literature, academics in the field of conversational agents, and the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition [19]. The search strategy included keyword truncations and mappings to subject heading (medical subject heading) that were adapted appropriately for each database. The reference lists of all included studies were handsearched to identify all relevant references. Gray literature, including conference abstracts or proceedings, and dissertations or theses identified through the database searches were also included for screening.

Eligibility Criteria

Studies were included if they reported on a conversational agent intervention for mental health; the agent was autonomous and could be used independently without support from a human; they simulated conversation; they relied on a turn-taking process with the user; and they reported on a mental health outcome. Review papers were included if all studies that were included met the inclusion criteria for this review. Studies were excluded if the output from the conversational agent was solely predetermined, for example, psychoeducation and not generated in response to user input; they used asynchronous communication, for example, email; they relied on a human user to generate responses (eg, Wizard of Oz methods); they required support from a person to operate, for example, a therapist or similar; they were limited to adherence to medication or physical health behaviors, for example, smoking cessation; they focused solely on the technical function, development or programming of the agent; and they lacked sufficient detail to determine eligibility (eg, short conference abstracts). Studies not written in English were translated as required. The review included a diverse range of study designs such as randomized controlled trials (RCTs), quasi-experimental designs, feasibility studies, and mixed method studies.


Studies identified through the database searches were exported to reference management software (Mendeley), and duplicates were deleted. Study selection was conducted by the first author (HG). Screening procedures were piloted before beginning the screening process. Abstracts and titles were initially screened, and articles not meeting the inclusion and exclusion criteria were removed. The first author (HG) then screened full texts and selected the articles for inclusion. Any lack of clarity over the eligibly of the studies was resolved through a discussion with a second author (WM). A random, 9.8% (26/264) sample of studies identified for full-text screening were also independently screened by a second reviewer. Cohen kappa was used to measure interrater agreement. Finally, reference lists of all included papers were screened for additional studies and the inclusion and exclusion criteria applied. See Figure 1 for a detailed breakdown of the flow of the included studies.

Figure 1. Flow diagram of included studies. Search updates were conducted until January 2019, with 2 new papers being identified.
View this figure

Data Extraction

Data from the included studies were extracted into a prespecified form, which included author, year of publication, study design, mental health domain, conversational agent name and description (including embodiment, access, theoretical approach, and input and output style), number and characteristics of participants (including age, gender, presence, and type of diagnosis or psychological problem), intervention description (including length and structure of intervention), control group description (if applicable), mental health outcome measures, user experience measures, attrition, and primary findings (primary mental health outcome and user experiences). Owing to the diversity in study designs, outcomes measured, intervention modalities, and durations and the varied use of active and inactive control groups, a meta-analysis would not have led to meaningful conclusions and was thus not undertaken. Instead, extracted data were narratively synthesized in line with guidance on the conduct of narrative synthesis in systematic reviews [20].

Risk-of-Bias Assessment

Risk-of-bias assessment of each study was conducted to ascertain the validity and reliability of the methods and findings to inform the narrative synthesis of the studies. The validated 16-item quality assessment tool for studies with diverse designs (QATSDD [21]) was deemed appropriate for this review to assess study quality as it includes quantitative (14 items), qualitative (14 items), and mixed-methods (16 items) items. Each of the 16 items is rated from 0 (not at all) to 3 (complete). Specifically, the tool assesses the clarity of the theoretical framework, study aims, study settings, the representativeness of the sample, rationale for data collection procedure, the appropriateness and reliability of data analysis, and the study’s strengths and limitations. For each included study, the scores for each item were summed and a percentage of the total possible score was calculated. If a study did not provide enough details to rate an item, the item was scored 0. The quality of each included study was assessed by the first author (HG).

Study Selection

The search identified 30,853 articles (see Figure 1) using the predefined search strategy outlined above. Duplicates were removed (8131), and articles not meeting the inclusion and exclusion criteria based on the title and abstract (22,388) were excluded. Handsearching through references resulted in an additional 5 studies being eligible for inclusion. The search was updated in January 2019, and 2 additional eligible studies were identified and included. Lack of clarity over the eligibility of articles (n=13) was resolved through a discussion with the second author (WM). Owing to the large number of articles identified from the initial search and limited researcher resource, interrater reliability was not assessed at the title and abstract stage. However, interrater reliability was assessed at full-text eligibility stage. A random sample of 26 studies (10% of the 264 studies identified for full-text screening) were independently screened by a second reviewer. The percentage agreement between first author (HG) and the independent rater was 96% (25/26 in agreement). Cohen kappa was 0.65, indicating substantial interrater agreement. Any differences in ratings were discussed, and an agreement was reached. A total of 13 articles were included in the review evaluating 11 different conversational agents.

Risk of Bias

The methodological quality of the included studies varied (see Table 1). Using the QATSDD [21] assessment tool, methodological quality ranged from the lowest score of 35% [22] to the highest score of 88% [23]. The average quality score was 59%. All of the included studies with percentage scores above 70% were RCTs [23-26].

All included studies received the maximum score of 3 for the criterion statement of aims or objectives in main body of report. All included studies scored a 2 or 3 for fit between research question and method of analysis and fit between stated research question and method of data collection. Most studies provided adequate descriptions of procedure for data collection and detailed recruitment data. Most studies (n=10) provided discussions of the key strengths and limitations of the study (scoring 2), and 3 studies gave thorough, complete discussions of strengths and limitations, obtaining a maximum score of 3. The lowest average scores were found for representative sample of target group of a reasonable size, good justification for analytic method selected, assessment of reliability of analytic process (qualitative only), and evidence of user involvement in design. See Table 1 for mean scores on each criterion across studies.

Table 1. List of criteria used to assess methodological quality and average score across studies.
1Explicit theoretical framework1.5
2Statement of aims/objectives in main body of report3.0
3Clear description of research setting2.1
4Evidence of sample size considered in terms of analysis1.5
5Representative sample of target group of a reasonable size1.3
6Description of procedure for data collection2.2
7Rationale for choice of data collection tool(s)1.8
8Detailed recruitment data2.2
9Statistical assessment of reliability and validity of measurement tool(s) (quantitative only)1.5
10Fit between stated research question and method of data collection (quantitative only)2.5
11Fit between stated research question and format and content of data collection tool, for example, interview schedule (qualitative only)1.9
12Fit between research question and method of analysis (quantitative only)2.5
13Good justification for analytic method selected1.2
14Assessment of reliability of analytic process (qualitative only)0
15Evidence of user involvement in design0.4
16Strengths and limitations critically discussed2.2

aScores can range from 0 (not at all) to 3 (complete).

Study Characteristics

The characteristics of the included studies are summarized in Multimedia Appendix 1. The 13 studies identified were conducted between 2013 and 2018 in 4 countries. Among them, 5 studies were conducted in the United Kingdom [23,24,27-29], 6 studies in the United States [22,25,26,30-32], 1 study in Sweden [33], and 1 study in Japan [34]. Across the studies, there was considerable heterogeneity in study design, intervention design, and outcome measures used. The majority of the included studies focused on interventions for common mental health problems, including depression [28-31] and/or anxiety [25,26], specific phobia (heights) [23], loneliness [22], and psychological distress [24,27]. Three studies focused on improving mental well-being [32-34]. A large proportion of studies (n=7) were preliminary and included feasibility [30], pilot RCTs [27,28,31-33], or nonrandomized trials [34]. In addition, 2 studies used quasi-experimental designs [22,29], and 4 studies were full-scale RCTs [23-26].

Most studies (n=8) used mixed methods [22,25,26,28-30,32,33], and the majority (n=9) of them reported on both mental health outcomes and user experiences [22,23,25,26,28-30,32,33].

Over half of the included studies (n=7) used specifically designed control groups, including screen or online psychoeducation [25,26,30,31], paper and CD-/MPEG-1 standard (MP3)-based psychoeducation [32] or an active control condition utilizing another conversational agent ELIZA [2,24,27]. Two studies used treatment as usual (TAU), which consisted of treatment for depression with a clinician [28] or corresponded to no treatment [23]. One study used a waitlist control group [33], and 1 study used a nonrandomized control group of participants who had expressed interest in taking part in the study but could not complete the intervention at that time [34]. Finally, 2 quasi-experimental studies did not use a control group [22,29]. However, Ring et al [22] compared groups that used 2 different versions (proactive and passive) of the conversational agent intervention.


Only 1 study, a pilot RCT, recruited participants from clinician caseloads or registers [28]. The remaining studies recruited self-selected participants from the community through outpatient clinics [32], universities [24-27,30,31,33], online advertisements [22,33,34], and radio advertisements [23] and by downloading the intervention app through the app store [29].

The included studies reported results from a total of 1200 participants. Study sample sizes ranged from 14 [22] to 454 [34]. Study participants ranged between 16 and 75 years old, and gender prevalence was 70.3% (692/985) female from studies that reported this data (12/13). One study with 129 participants [29] did not collect data on age or gender, and 1 study recruited only women [32]. Participants varied widely in severity of psychological distress from minimal psychological symptoms [22] to formal clinical diagnoses such as major depressive disorder [28] and acrophobia [23]. In addition, 5 of the 13 included studies recruited participants who self-reported symptoms of psychological distress to varying degrees.

Conversational Agent Interventions

Overall, 6 of the conversational agents were embodied (7 studies) [22,23,28,30-32,34]. Conversational agents used different technologies, with 3 conversational agents accessed on an app [25,29,33], 4 online (5 studies) [24,26,27,32,34], 3 using an offline computer program (4 studies) [22,28,30,31], and 1 VR program utilizing a VR headset [23].

The majority (8 out of 11 agents, evaluated in 9 studies) of the conversational agents included took natural language input either written [24-27,29,33,34] or spoken [23,28]. The remaining 3 agents took responses from participants using fixed onscreen response options (4 studies) [22,30-32]. The output mainly consisted of questions or written text (6 out of 11 agents, evaluated in 7 studies) [24-27,29,33,34]. Furthermore, 4 agents used spoken output [22,23,28,32]. In addition, 2 studies (1 conversational agent) [31] did not specify whether the conversational agent output was written or spoken.

The conversational agents provided interventions aimed at reducing symptoms [22-29], increasing well-being [32-34], or improving self-management [30,31]. Across the set of conversational agents, a range of therapeutic orientations were used, including cognitive behavioral therapy [23,25,28,34], method of levels (MOL) [24,27], mindfulness-based stress reduction [32], structured communication enhancement strategy [30,31], and eclectic interventions drawing on a wide variety of approaches [22,26,29,33]. Over half of the conversational agents (7 out of 11) focused on providing psychoeducation and self-management strategies [25,26,28,29,32-34], 1 agent (evaluated in 2 studies) utilized the principles of MOL therapy in a question-and-answer format [24,27], 1 agent offered social companionship [22], and 1 agent (evaluated in 2 studies) facilitated practice of effective communication with human health care professionals around psychological symptoms [30,31].

Conversational agent interventions varied widely in frequency and duration (see Table 2). From short interventions of 1 session (participant-determined length [24] up to 20 min [27]), 3 sessions (unspecified duration [31], 15-20 min each [30]), and 6 sessions (30 min each [23] through to daily usage over 2 weeks [25,26,33], 4 weeks [26,28], or a month [32]). One study only used data from participants who had engaged with the intervention at least every other day (>15 times) over a month [34]. Finally, 1 study installed 1 of 2 versions (passive, activated at will, and proactive, activated by a motion sensor) of the same conversational agent into participants’ homes for 1 week. One study enabled participants to continue TAU for depression with a clinician alongside the conversational agent intervention [28]. The majority of studies (n=9) set no upper limits on usage during the defined study period [22,24-26,28,29,32-34].

Table 2. Intervention engagement.
StudyTotal intervention lengthFrequency of use, mean (SD) or median (IQR)Intervention duration (min), mean (SD) or median (IQR)
Freeman et al, 2018 [23]6 × 30-min sessions over 2 weeks4.66 (SD 1.27)124.4 (SD 34.2)
Bird et al, 2018 [24]1 sessionNot applicable13 (SD NRa)
Fulmer et al, 2018 [26]Unlimited access for 2 weeks or 4 weeks192 interactions (SD NR)NR
Fitzpatrick et al, 2017 [25]Daily intervention for 2 weeks12.1 (SD 2.23)NR
Ly et al, 2017 [33]Daily intervention for 2 weeks17.71NR
Gaffney et al, 2014 [27]1 sessionNot applicable19.23 (SD 0.002)
Inkster et al, 2018 [29]Unlimited access for 2 weeks83% (90/108) of high-usage users (at least one use) used the app for more than 4 daysNR
Gardiner et al, 2017 [32]Unlimited access for 30 daysNR52 (IQR 101.4)
Pinto et al, 2016 [30]3 × 15-20-min sessions (baseline, 4 weeks, and 8 weeks)12 of 25 participants completed all sessionsNR
Burton et al, 2016 [28]Daily intervention over 4 weeks10.5 (IQR NR)134 (IQR NR)
Suganuma et al, 2018 [34]At least 15 times over 1 month45% (191/427) completed >15 days of interventionNR
Pinto et al, 2013 [31]3 sessions (baseline, 4 weeks, and 8 weeks)NRNR
Ring et al, 2015 [22]Unlimited access over 1 week15.9 (SD 8.1) interactions2.3 (SD 0.038) each

aNR: not reported.

Feasibility and Engagement

One study reported low uptake as they aimed to recruit 52 participants but closed the study at 28 participants [28]. Attrition rates between pre- and postmeasures were reported in 11 studies and varied widely from no attrition [23,24,33] to 74.1% (1978/2668) of participants [34]. Reasons reported for dropout included difficulties attending the university to take part because of financial difficulties [30], technical problems [22,27], and mental illness [22]. One study with a high attrition rate (74.1%, 1978/2668) [34] did not report any reasons; however, it should be noted that the majority of the dropouts were from the control condition (1846/2109, 88.3%) compared with 23.6% (132/559) in the intervention condition.

Studies reported differing metrics for engagement, and reporting was inconsistent (see Table 2). Engagement with the conversational agent interventions was highly variable from a short period of interaction in 1 session (eg, a mean of 13 min; [24]) to a median interaction total of 134 min [28] or exchanging a mean of 192 messages during intervention [26]. In the study by Suganuma et al [34], 236 out of 427 (55.2%) of intervention participants did not complete 15 or more days of the intervention and were excluded from the analysis. In addition, 3 people (6%) in the study by Freeman et al [23] found the intervention sessions too difficult and did not complete the intervention. However, 44 out of 49 (90%) participants completed the intervention, with a mean total intervention time of 124 min. One study [31] did not report any measures of engagement.

Psychological Outcomes

Primary outcome measures were all validated but varied (see Multimedia Appendix 1 for details); therefore, the term psychological distress will be used to facilitate a summary. Of the 13 studies included, 5 controlled studies reported significant posttreatment improvements in psychological distress in the intervention group compared with a no treatment or information control group [23,25,26,31,34]. Significant improvements were observed on measures of depression [25,26,31], psychological distress [34], anxiety [26], fear of heights [23] and positive affect [26,34]. Effects ranged from small (d=−0.24 [34]) to very large (d=2.0 [23]). In addition, 2 pilot trials with active control groups found significantly higher ratings of problem resolution in the intervention group compared with the control group [24,27].

Furthermore, 4 controlled studies reported no significant posttreatment differences on measures of psychological distress between the intervention and control groups [24,27,32,33] with both intervention and control conditions demonstrating reduced distress [24,27,33] or increased uptake of stress management techniques [32]. Despite significant reductions in depression observed in the intervention group compared with the control group in the intention-to-treat analysis by Fitzpatrick et al [25], no significant posttreatment differences in anxiety were observed between groups.

Finally, the 2 uncontrolled studies included in the review [22,29] and 2 studies that did not test for between-group effects [28,30] reported reductions in depression [28-30] and loneliness [22] postintervention. Generally, greater engagement with the conversational agent resulted in greater reductions in psychological distress [22,26,28,29,33]. Only 3 studies included a follow-up period [23,24,27].

User Experience Outcomes

Generally, from studies that reported user experience outcomes (n=11), participants reported being satisfied with the conversational agent interventions offered [22,23,25,26,29,30,32]. In addition, 3 studies reported that participants found the conversational agent interventions available and accessible [26,32,33]. Participants reported that they found the agent empathic [26], that they liked the interactivity [30], the agent’s personality [22,25], the agent’s ability to form a relationship [28,33], and the agent’s ability to learn from input [26]. Participants reported that they liked the ability to customize the gender and appearance of ECAs [28] and the option to tailor the session length to their own needs [28]. Participants in the study by Fitzpatrick et al [25] reported that they liked the daily check-ins and information provided. Furthermore, 2 studies reported that participants indicated that they would recommend the conversational agent intervention to other people [18,24] (the proactive version).

The predominant challenges to intervention with a conversational agent included repetitive content [22,25,26,28,29,33], limitations in the agents ability to understand or respond appropriately [22,25,26,29], a shallow or superficial relationship [28,33], the sound and quality of the agents voice [32], and specific intervention tools or content [25,29]. Some participants in the study by Pinto et al [30] reported that they would like more frequent, longer intervention sessions, and greater freedom to tailor content and responses to their needs.

Principal Findings

The use of conversational agents for treating mental health problems appears to be limited but is growing quickly, with 5 of the included studies published in 2018 alone [23,24,26,29,34]. Furthermore, despite the heterogeneity in evaluation methods, there is an increasing emphasis on fully powered RCTs testing efficacy. Included interventions were generally brief, allowed participants to control the intensity of intervention, and drew from a wide variety of psychological approaches. All included studies reported reduced psychological distress postintervention with a conversational agent. In addition, 5 controlled studies demonstrated significant reductions in psychological distress compared with an information or no treatment control group with small-to-large effects. This provides some support for the utility of conversational agents in treating mild-to-moderate psychological distress in adults [23,25,26,31,34]. However, their broader utility in promoting positive well-being in nonclinical populations appears uncertain [32,33]. Controlled studies with active control conditions (eg, another conversational agent or human psychological therapy) failed to demonstrate superior effects [24,27,28]. However, it is important to highlight that these studies assessed relative rather than absolute treatment efficacy, and thus, we cannot conclude an absolute lack of treatment efficacy [35].

Studies managed to recruit participants through several different methods. Remarkably, the only study that reported difficulties in recruiting participants relied on clinicians to refer patients to the study [28]. Studies that used more flexible recruitment routes such as online adverts [34] and app stores [29] recruited greater numbers of participants. It is possible that clinician apprehension about digital treatment for mental health problems affected recruitment rates. This is supported by research indicating that clinicians are perhaps more reluctant to recommend digital interventions without clinician input or support [36,37]. Our findings illustrate that conversational agents are generally an acceptable format of intervention for participants. Interestingly, participants valued aspects of agents usually seen as unique to therapy with a human, such as empathic responses, personality, the ability to build a relationship, and an interactive, conversational approach. This is consistent with research demonstrating that people relate to conversational agents as if they were human despite knowing that they are computer programs [38]. Participants also valued the ability of the agent to learn from their input, perhaps emulating the learning of a human therapist over time. Participants found intervention with conversational agents difficult or frustrating when the agent did not understand, became confused, or was repetitive. This perhaps mirrors expectations around core relationship factors such as feeling understood. Control was also important for participants especially regarding tailoring session length and content to their own needs and engaging with interventions in their own words (eg, free-text rather than fixed response options). The accessibility of the interventions was a key strength for many participants and where accessibility was limited, participants highlighted this and suggested ways to improve accessibility (eg, online access [30]).

Limitations of Included Studies

The studies described have several limitations. The methodological quality of the included studies varied, and sample sizes were mainly small and self-selected, which reduces the ability to draw firm conclusions about the reliability and validity of the findings. Furthermore, because of short or absent follow-up, conclusions about the sustainability of treatment gains cannot be made. Psychological comorbidity was not assessed in any of the studies despite comorbidities being prevalent in individuals with common mental health problems [39]. Safety was only explicitly evaluated and reported in 1 study [30]. Safety is a vital consideration in mental health interventions that use free-text, natural language input either written [24-27,29,33,34] or spoken [23,28]. Studies have demonstrated that these types of conversational agents are often not able to respond appropriately to risk information such as suicidal ideation [40,41] and have the potential to result in harm. Furthermore, users can expect a level of understanding beyond what is currently technologically possible [41]. Engagement with interventions was not reported consistently and appeared highly variable, and the reasons for this remain unexplored. Furthermore, the impact of the design or features of the conversational agents (eg, embodiment and speech or text based) on engagement or outcomes was not explicitly assessed or compared; therefore, conclusions cannot be drawn as to the most effective or acceptable modality. No studies evaluated therapeutic equivalence or superiority to other treatment modalities such as face-to-face therapy. Finally, a large proportion of agents were eclectic interventions comprising a variety of strategies and psychoeducation drawing on a range of therapeutic orientations [22,26,29,33]. Therefore, it is difficult to ascertain what the active ingredients of the interventions are.

Strengths and Limitations

Owing to the lack of standardized terminology in this area, we conducted a comprehensive search that prioritized sensitivity over specificity. We also reviewed reference lists for additional papers not identified through the database searches. Published abstracts commonly presented in technology conferences were also included as they typically provide enough detail for decisions to be made about inclusion. The review was also registered on PROSPERO before commencing. We also included a broad range of formats for conversational agents, including VR and embodied and/or text and speech input. Cohen kappa showed substantial agreement in full-text screening, and there was a high percentage of agreement overall. This is despite inconsistencies in the reporting of interventions which made the process of eligibility assessment more complicated and reflected the heterogeneity and complexity in the field. Owing to the heterogeneity of the included studies, a meta-analysis was not undertaken. Furthermore, some potentially relevant conversational agents developed for the treatment of mental health problems were excluded from this review because of not reporting a mental health outcome measure (eg, ELIZA [2,42-44]).

Future Directions

Continued growth in the use of conversational agents in mental health treatment is expected. Considering the findings, several priority areas for further research are apparent. First, addressing technical deficits such as repetition and confusion, which were reported in half of the included studies [22,25,26,28,29,33], may help to overcome barriers to engagement. Increased interdisciplinary working between computer science and mental health may facilitate this and help to drive innovations forward. Given that only 1 included study explicitly reported on safety [30], demonstrating safety will also be key to developing patient and public trust [40]. Furthermore, given the range of differing modalities of conversational agents and lack of direct comparisons between them found in this review, it will be important to compare modalities, for example, embodied or nonembodied or speech or text or offer increased choice to individuals. This would enable further insight into what works and for whom. Our review found that a large proportion of conversational agents use an eclectic mix of psychological interventions with often limited theoretical basis [22]. Only 1 included study reported on the process of psychological change [27] with conversational agent Manage Your Life Online (MYLO). Identifying and demonstrating the key mechanisms of action of conversational agent interventions has the potential to increase treatment efficiency, reduce unnecessary burden on users, and increase transparency. Given the diversity of mental health problems (eg, depression, anxiety, and phobias) appearing potentially amenable to treatment with conversational agent interventions, consideration of transdiagnostic approaches to intervention would further increase applicability and reach (eg, to people with comorbidities or difficulties that do not easily fit into prespecified diagnostic categories). Finally, in line with guidance on research priorities for digital interventions [45], it will be important to demonstrate efficacy and/or superiority compared with alternative conversational agent interventions and other treatment modalities such as face-to-face therapy to develop patient and clinician confidence in this type of intervention.


This systematic review provides an assessment of conversational agent interventions used for the treatment of mental health problems. On the basis of the current evidence, the efficacy and acceptability of conversational agent interventions appears promising compared with no treatment or information control. However, studies failed to demonstrate superiority when compared with other active, conversational interventions, and their broader utility in promoting well-being in nonclinical populations is unclear.

Therefore, whether conversational agent interventions are an adequate substitute to other therapy modalities remains unclear. Future studies should strive to demonstrate efficacy, equivalence (or superiority), and cost-effectiveness through RCTs with comparisons with other forms of treatment. Studies that can demonstrate exactly how interventions achieve psychological change and for whom will be important in streamlining bloated interventions to increase acceptability. Finally, transdiagnostic approaches to treatment may provide further opportunity to maximize the reach and simplicity of conversational agent interventions.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Characteristics of included studies.

PDF File (Adobe PDF File), 319 KB


  1. McTear M, Callejas Z, Griol D. The Conversational Interface Talking to Smart Devices. Switzerland: Springer Publishing Company; 2016.
  2. Weizenbaum J. ELIZA---a computer program for the study of natural language communication between man and machine. Commun ACM 1994 Jun;9(1):36-45. [CrossRef]
  3. Fast E, Horvitz E. Long-Term Trends in the Public Perception of Artificial Intelligence. In: Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence. 2017 Presented at: AAAI'17; February 4-9, 2017; San Francisco, California, USA   URL:
  4. Gentsch P. AI in Marketing, Sales and Service: How Marketers Without a Data Science Degree Can Use AI, Big Data and Bots. Switzerland: Palgrave Macmillan; 2019.
  5. Campbell RH, Grimshaw MN, Green GM. Relational agents: a critical review. Open Virtual Real J 2009 Apr 10;1(1):1-7. [CrossRef]
  6. Laranjo L, Dunn AG, Tong HL, Kocaballi AB, Chen J, Bashir R, et al. Conversational agents in healthcare: a systematic review. J Am Med Inform Assoc 2018 Sep 1;25(9):1248-1258 [FREE Full text] [CrossRef] [Medline]
  7. Bickmore T, Gruber A. Relational agents in clinical psychiatry. Harv Rev Psychiatry 2010;18(2):119-130. [CrossRef] [Medline]
  8. Hoermann S, McCabe KL, Milne DN, Calvo RA. Application of synchronous text-based dialogue systems in mental health interventions: systematic review. J Med Internet Res 2017 Jul 21;19(8):e267 [FREE Full text] [CrossRef] [Medline]
  9. Scholten MR, Kelders SM, van Gemert-Pijnen JE. Self-guided web-based interventions: scoping review on user needs and the potential of embodied conversational agents to address them. J Med Internet Res 2017 Nov 16;19(11):e383 [FREE Full text] [CrossRef] [Medline]
  10. Provoost S, Lau HM, Ruwaard J, Riper H. Embodied conversational agents in clinical psychology: a scoping review. J Med Internet Res 2017 May 9;19(5):e151 [FREE Full text] [CrossRef] [Medline]
  11. Luxton DD, editor. Artificial Intelligence in Behavioral and Mental Health Care. Cambridge, Massachusetts: Academic Press; 2016.
  12. Duarte A, Walker S, Littlewood E, Brabyn S, Hewitt C, Gilbody S, et al. Cost-effectiveness of computerized cognitive-behavioural therapy for the treatment of depression in primary care: findings from the randomised evaluation of the effectiveness and acceptability of computerised therapy (REEACT) trial. Psychol Med 2017 Jul;47(10):1825-1835. [CrossRef] [Medline]
  13. Gilbody S, Brabyn S, Lovell K, Kessler D, Devlin T, Smith L, REEACT Collaborative. Telephone-supported computerised cognitive-behavioural therapy: REEACT-2 large-scale pragmatic randomised controlled trial. Br J Psychiatry 2017 May;210(5):362-367. [CrossRef] [Medline]
  14. Lucas GM, Gratch J, King A, Morency LP. It’s only a computer: virtual humans increase willingness to disclose. Comput Hum Behav 2014 Aug;37:94-100. [CrossRef]
  15. Bickmore TW, Picard RW. Establishing and maintaining long-term human-computer relationships. ACM Trans Comput Hum Interact 2005 Jun 1;12(2):293-327. [CrossRef]
  16. Lovejoy CA, Buch V, Maruthappu M. Technology and mental health: the role of artificial intelligence. Eur Psychiatry 2019 Jan;55:1-3. [CrossRef] [Medline]
  17. Shamseer L, Moher D, Clarke M, Ghersi D, Liberati A, Petticrew M, PRISMA-P Group. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015: elaboration and explanation. Br Med J 2015 Jan 2;350:g7647 [FREE Full text] [CrossRef] [Medline]
  18. Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann Intern Med 2009 Aug 18;151(4):264-9, W64. [CrossRef] [Medline]
  19. American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders: DSM-5. Fifth Editiob. Arlington, VA: American Psychiatric Publishing; 2013.
  20. Popay J, Roberts H, Sowden A, Petticrew M, Aria L, Rodgers M, et al. Semantic Scholar. 2006. Guidance on the Conduct of Narrative Synthesis in Systematic Reviews: A Product from the ESRC Methods Programme   URL: [accessed 2019-09-17]
  21. Sirriyeh R, Lawton R, Gardner P, Armitage G. Reviewing studies with diverse designs: the development and evaluation of a new tool. J Eval Clin Pract 2012 Aug;18(4):746-752. [CrossRef] [Medline]
  22. Ring L, Shi L, Totzke K, Bickmore T. Social support agents for older adults: longitudinal affective computing in the home. J Multimodal User Interfaces 2014 Jun 18;9(1):79-88. [CrossRef]
  23. Freeman D, Haselton P, Freeman J, Spanlang B, Kishore S, Albery E, et al. Automated psychological therapy using immersive virtual reality for treatment of fear of heights: a single-blind, parallel-group, randomised controlled trial. Lancet Psychiatry 2018 Aug;5(8):625-632 [FREE Full text] [CrossRef] [Medline]
  24. Bird T, Mansell W, Wright J, Gaffney H, Tai S. Manage your life online: a web-based randomized controlled trial evaluating the effectiveness of a problem-solving intervention in a student sample. Behav Cogn Psychother 2018 Sep;46(5):570-582. [CrossRef] [Medline]
  25. Fitzpatrick KK, Darcy A, Vierhile M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Ment Health 2017 Jun 6;4(2):e19 [FREE Full text] [CrossRef] [Medline]
  26. Fulmer R, Joerin A, Gentile B, Lakerink L, Rauws M. Using psychological artificial intelligence (Tess) to relieve symptoms of depression and anxiety: randomized controlled trial. JMIR Ment Health 2018 Dec 13;5(4):e64 [FREE Full text] [CrossRef] [Medline]
  27. Gaffney H, Mansell W, Edwards R, Wright J. Manage Your Life Online (MYLO): a pilot trial of a conversational computer-based intervention for problem solving in a student sample. Behav Cogn Psychother 2014 Nov;42(6):731-746. [CrossRef] [Medline]
  28. Burton C, Tatar AS, McKinstry B, Matheson C, Matu S, Moldovan R, Help4Mood Consortium. Pilot randomised controlled trial of Help4Mood, an embodied virtual agent-based system to support treatment of depression. J Telemed Telecare 2016 Sep;22(6):348-355. [CrossRef] [Medline]
  29. Inkster B, Sarda S, Subramanian V. An empathy-driven, conversational artificial intelligence agent (Wysa) for digital mental well-being: real-world data evaluation mixed-methods study. J Mhealth Uhealth 2018 Nov 23;6(11):e12106 [FREE Full text] [CrossRef] [Medline]
  30. Pinto MD, Greenblatt AM, Hickman RL, Rice HM, Thomas TL, Clochesy JM. Assessing the critical parameters of eSMART-MH: a promising avatar-based digital therapeutic intervention to reduce depressive symptoms. Perspect Psychiatr Care 2016 Jul;52(3):157-168. [CrossRef] [Medline]
  31. Pinto MD, Hickman RL, Clochesy J, Buchner M. Avatar-based depression self-management technology: promising approach to improve depressive symptoms among young adults. Appl Nurs Res 2013 Feb;26(1):45-48 [FREE Full text] [CrossRef] [Medline]
  32. Gardiner PM, McCue KD, Negash LM, Cheng T, White LF, Yinusa-Nyahkoon L, et al. Engaging women with an embodied conversational agent to deliver mindfulness and lifestyle recommendations: a feasibility randomized control trial. Patient Educ Couns 2017 Sep;100(9):1720-1729 [FREE Full text] [CrossRef] [Medline]
  33. Ly KH, Ly AM, Andersson G. A fully automated conversational agent for promoting mental well-being: a pilot RCT using mixed methods. Internet Interv 2017 Dec;10:39-46 [FREE Full text] [CrossRef] [Medline]
  34. Suganuma S, Sakamoto D, Shimoyama H. An embodied conversational agent for unguided internet-based cognitive behavior therapy in preventative mental health: feasibility and acceptability pilot trial. JMIR Ment Health 2018 Jul 31;5(3):e10454 [FREE Full text] [CrossRef] [Medline]
  35. Karlsson P, Bergmark A. Compared with what? An analysis of control-group types in Cochrane and Campbell reviews of psychosocial treatment efficacy with substance use disorders. Addiction 2015 Mar;110(3):420-428 [FREE Full text] [CrossRef] [Medline]
  36. Topooco N, Riper H, Araya R, Berking M, Brunn M, Chevreul K, E-COMPARED Consortium. Attitudes towards digital treatment for depression: a European stakeholder survey. Internet Interv 2017 Jun;8:1-9 [FREE Full text] [CrossRef] [Medline]
  37. Dunne N. Electronic Theses and Dissertations (ETD) Center. 2017. Evaluation of Psychology Clinicians’ Attitudes Towards Computerized Cognitive Behavior Therapy, for Use In Their Future Clinical Practice, With Regard to Treating Those Suffering From Anxiety and Depression   URL:!etd.send_file?accession=antioch1503328670275243&disposition=inline [accessed 2019-09-17]
  38. Waytz A, Cacioppo J, Epley N. Who sees human? The stability and importance of individual differences in anthropomorphism. Perspect Psychol Sci 2010 May;5(3):219-232 [FREE Full text] [CrossRef] [Medline]
  39. Kessler RC, Wai TC, Demler O, Merikangas KR, Walters EE. Prevalence, severity, and comorbidity of 12-month DSM-IV disorders in the National Comorbidity Survey Replication. Arch Gen Psychiatry 2005 Jun;62(6):617-627 [FREE Full text] [CrossRef] [Medline]
  40. Miner AS, Milstein A, Schueller S, Hegde R, Mangurian C, Linos E. Smartphone-based conversational agents and responses to questions about mental health, interpersonal violence, and physical health. JAMA Intern Med 2016 May 1;176(5):619-625 [FREE Full text] [CrossRef] [Medline]
  41. Bickmore TW, Trinh H, Olafsson S, O'Leary TK, Asadi R, Rickles NM, et al. Patient and consumer safety risks when using conversational assistants for medical information: an observational study of Siri, Alexa, and Google assistant. J Med Internet Res 2018 Sep 4;20(9):e11510 [FREE Full text] [CrossRef] [Medline]
  42. Shinozaki T, Yamamoto Y, Tsuruta S, Knauf R. Validation of Context Respectful Counseling Agent. In: Proceedings of the International Conference on Systems, Man, and Cybernetics. 2015 Presented at: SMC'15; October 9-12, 2015; Kowloon, China p. 993-998. [CrossRef]
  43. Hudlicka E. Virtual training and coaching of health behavior: example from mindfulness meditation training. Patient Educ Couns 2013 Aug;92(2):160-166 [FREE Full text] [CrossRef] [Medline]
  44. Tielman ML, Neerincx MA, van Meggelen M, Franken I, Brinkman W. How should a virtual agent present psychoeducation? Influence of verbal and textual presentation on adherence. Technol Health Care 2017 Dec 4;25(6):1081-1096 [FREE Full text] [CrossRef] [Medline]
  45. Hollis C, Sampson S, Simons L, Davies EB, Churchill R, Betton V, et al. Identifying research priorities for digital technology in mental health care: results of the James Lind Alliance Priority Setting Partnership. Lancet Psychiatry 2018 Oct;5(10):845-854. [CrossRef] [Medline]

ECA: embodied conversational agent
MOL: method of levels
QATSDD: quality assessment tool for studies with diverse designs
RCT: randomized controlled trial
TAU: treatment as usual
VR: virtual reality

Edited by S Ondersma; submitted 27.03.19; peer-reviewed by E Grekin, T Bickmore, D Morrow; comments to author 07.05.19; revised version received 30.06.19; accepted 18.07.19; published 18.10.19


©Hannah Gaffney, Warren Mansell, Sara Tai. Originally published in JMIR Mental Health (, 18.10.2019.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on, as well as this copyright and license information must be included.