JMIR Mental Health
Internet interventions, technologies, and digital innovations for mental health and behavior change.
JMIR Mental Health is the official journal of the Society of Digital Psychiatry.
Editor-in-Chief:
John Torous, MD, MBI, Harvard Medical School, USA
Impact Factor 5.8 CiteScore 10.2
Recent Articles

Adolescence is a crucial developmental period characterized by elevated stress and significant mental health challenges, including depression and anxiety. With barriers like stigma, accessibility, and cost hindering effective treatment, leveraging school systems for mental health interventions offers a strategic advantage due to their reach and potential for scalability.


Intrusive memories are a core symptom of posttraumatic stress disorder (PTSD), yet their retrospective assessment is prone to biases, making real-time methods such as e-diaries essential. While trauma-focused treatments target intrusive symptoms, their efficacy has not yet been evaluated using real-time assessments.

Innovative, scalable mental health tools are needed to address systemic provider shortages and accessibility barriers. Large language model (LLM)-based tools can provide real-time, tailored feedback to help users engage in cognitive reappraisal outside traditional therapy sessions. Socrates 2.0 is a multi-agent artificial intelligence (AI) tool that guides users through Socratic dialogue.

Stopping smoking can improve mental health, with effect sizes similar to antidepressant treatment. Internet-based cognitive behavioral therapy (iCBT) provides evidence-based treatment for depression and anxiety, and digital interventions can support smoking cessation. However, combined digital smoking and mental health support is not currently available in UK health services.

Artificial intelligence (AI) applications in mental health have expanded rapidly, and consumers are already using freely available generative AI models for self-guided mental health support despite limited clinical validation. In August 2025, Illinois enacted Public Act 104-0054, the first state statute in the United States to explicitly define and regulate the use of AI in psychotherapy services, establishing boundaries around administrative support, supplementary support, and therapeutic communication. While the Act clarifies several aspects of AI use in therapy, it also leaves important gray areas, such as whether AI-generated session summaries, psychoeducation, or risk-flagging functions should be considered therapeutic communication. Drawing on the history of empirically supported treatments in psychology, we argue that a framework of evidence, safety, fidelity, and legal compliance could help determine when AI tools should be integrated into clinical care. This approach provides a concrete pathway for balancing patient protection with responsible innovation in the rapidly evolving field of mental health AI tools.


Cognitive behavioral therapy (CBT) is recommended as the first-line treatment for insomnia; however, few patients have access to it. A new class of Food and Drug Administration (FDA)–regulated digital CBT treatments has the potential to address this unmet need. These treatments are ordered or prescribed by health care providers and are fully automated, delivering CBT directly to patients without human coaches. This trial builds upon promising earlier digital cognitive behavioral therapy for insomnia (CBT-I) research by using a decentralized design to recruit a sample with greater representation of the US general population, including individuals from lower socioeconomic status groups who often face greater barriers to care.

The integration of artificial intelligence (AI) into daily life has introduced unprecedented forms of human-machine interaction, prompting psychiatry to reconsider the boundaries between environment, cognition, and technology. This Viewpoint reviews the concept of “AI psychosis,” which is a framework to understand how sustained engagement with conversational AI systems might trigger, amplify, or reshape psychotic experiences in vulnerable individuals. Drawing from phenomenological psychopathology, the stress-vulnerability model, cognitive theory, and digital mental health research, the paper situates AI psychosis at the intersection of predisposition and algorithmic environment. Rather than defining a new diagnostic entity, it examines how immersive and anthropomorphic AI technologies may modulate perception, belief, and affect, altering the prereflective sense of reality that grounds human experience. The argument unfolds through 4 complementary lenses. First, within the stress-vulnerability model, AI acts as a novel psychosocial stressor. Its 24-hour availability and emotional responsiveness may increase allostatic load, disturb sleep, and reinforce maladaptive appraisals. Second, the digital therapeutic alliance, a construct describing relational engagement with digital systems, is conceptualized as a double-edged mediator. While empathic design can enhance adherence and support, uncritical validation by AI systems may entrench delusional conviction or cognitive perseveration, reversing the corrective principles of cognitive-behavioral therapy for psychosis. Third, disturbances in theory of mind offer a cognitive pathway: individuals with impaired or hyperactive mentalization may project intentionality or empathy onto AI, perceiving chatbots as sentient interlocutors. This dyadic misattribution may form a “digital folie à deux,” where the AI becomes a reinforcing partner in delusional elaboration. Fourth, emerging risk factors, including loneliness, trauma history, schizotypal traits, nocturnal or solitary AI use, and algorithmic reinforcement of belief-confirming content may play roles at the individual and environmental levels. Building on this synthesis, we advance a translational research agenda and five domains of action: (1) empirical studies using longitudinal and digital-phenotyping designs to quantify dose-response relationships between AI exposure, stress physiology, and psychotic symptomatology; (2) integration of digital phenomenology into clinical assessment and training; (3) embedding therapeutic design safeguards into AI systems, such as reflective prompts and “reality-testing” nudges; (4) creation of ethical and governance frameworks for AI-related psychiatric events, modeled on pharmacovigilance; and (5) development of environmental cognitive remediation, a preventive intervention aimed at strengthening contextual awareness and reanchoring experience in the physical and social world. By applying empirical rigor and therapeutic ethics to this emerging interface, clinicians, researchers, patients, and developers can transform a potential hazard into an opportunity to deepen understanding of human cognition, safeguard mental health, and promote responsible AI integration within society.

Cognitive behavioral therapy (CBT)-based chatbots, many of which incorporate artificial intelligence (AI) techniques such as natural language processing and machine learning, are increasingly evaluated as scalable solutions for addressing mental health issues such as depression and anxiety. These fully automated or minimally supported interventions offer novel pathways for psychological support, especially for individuals with limited access to traditional therapy.

Digital mental health tools such as apps, virtual reality, and AI hold great promise but continue to face barriers to widespread clinical adoption. The Society of Digital Psychiatry, in partnership with JMIR Mental Health, presents a three-pronged roadmap to accelerate their safe, effective, and equitable implementation. First, Education: integrate digital psychiatry into core training and professional development through a global webinar series, annual symposium, newsletter, and an updated open-access curriculum addressing AI and the evolving Digital Navigator role. Second, AI Standards: develop transparent, actionable benchmarks and consensus guidance, through initiatives like MindBench.ai, to assess reasoning, safety, and representativeness across populations. Third, Digital Navigators: expand structured, train-the-trainer programs that enhance digital literacy, engagement, and workflow integration across diverse care settings, including low- and middle-income countries. Together, these pillars bridge research and practice, advancing digital psychiatry grounded in inclusivity, accountability, and measurable clinical impact.
Preprints Open for Peer-Review
Open Peer Review Period:
-
Open Peer Review Period:
-








