Published on in Vol 11 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/48974, first published .
Toward Tailoring Just-in-Time Adaptive Intervention Systems for Workplace Stress Reduction: Exploratory Analysis of Intervention Implementation

Toward Tailoring Just-in-Time Adaptive Intervention Systems for Workplace Stress Reduction: Exploratory Analysis of Intervention Implementation

Toward Tailoring Just-in-Time Adaptive Intervention Systems for Workplace Stress Reduction: Exploratory Analysis of Intervention Implementation

Original Paper

1Microsoft Research, Redmond, WA, United States

2Idiographic Dynamics Lab, Department of Psychology, University of California, Berkeley, CA, United States

3MIT Media Lab, Massachusetts Institute of Technology, Cambridge, MA, United States

4Siebel School of Computing and Data Science, University of Illinois Urbana-Champaign, Urbana, IL, United States

5Paul G Allen School of Computer Science & Engineering, University of Washington, Seattle, WA, United States

6Human-Centered Design and Engineering, University of Washington, Seattle, WA, United States

Corresponding Author:

Jina Suh, PhD

Microsoft Research

1 Microsoft Way

Redmond, WA, 98052

United States

Phone: 1 4254211584

Email: jinsuh@microsoft.com


Background: Integrating stress-reduction interventions into the workplace may improve the health and well-being of employees, and there is an opportunity to leverage ubiquitous everyday work technologies to understand dynamic work contexts and facilitate stress reduction wherever work happens. Sensing-powered just-in-time adaptive intervention (JITAI) systems have the potential to adapt and deliver tailored interventions, but such adaptation requires a comprehensive analysis of contextual and individual-level variables that may influence intervention outcomes and be leveraged to drive the system’s decision-making.

Objective: This study aims to identify key tailoring variables that influence momentary engagement in digital stress reduction microinterventions to inform the design of similar JITAI systems.

Methods: To inform the design of such dynamic adaptation, we analyzed data from the implementation and deployment of a system that incorporates passively sensed data across everyday work devices to send just-in-time stress reduction microinterventions in the workplace to 43 participants during a 4-week deployment. We evaluated 27 trait-based factors (ie, individual characteristics), state-based factors (ie, workplace contextual and behavioral signals and momentary stress), and intervention-related factors (ie, location and function) across 1585 system-initiated interventions. We built logistical regression models to identify the factors contributing to momentary engagement, the choice of interventions, the engagement given an intervention choice, the user rating of interventions engaged, and the stress reduction from the engagement.

Results: We found that women (odds ratio [OR] 0.41, 95% CI 0.21-0.77; P=.03), those with higher neuroticism (OR 0.57, 95% CI 0.39-0.81; P=.01), those with higher cognitive reappraisal skills (OR 0.69, 95% CI 0.52-0.91; P=.04), and those that chose calm interventions (OR 0.43, 95% CI 0.23-0.78; P=.03) were significantly less likely to experience stress reduction, while those with higher agreeableness (OR 1.73, 95% CI 1.10-2.76; P=.06) and those that chose prompt-based (OR 6.65, 95% CI 1.53-36.45; P=.06) or video-based (OR 5.62, 95% CI 1.12-34.10; P=.12) interventions were substantially more likely to experience stress reduction. We also found that work-related contextual signals such as higher meeting counts (OR 0.62, 95% CI 0.49-0.78; P<.001) and higher engagement skewness (OR 0.64, 95% CI 0.51-0.79; P<.001) were associated with a lower likelihood of engagement, indicating that state-based contextual factors such as being in a meeting or the time of the day may matter more for engagement than efficacy. In addition, a just-in-time intervention that was explicitly rescheduled to a later time was more likely to be engaged with (OR 1.77, 95% CI 1.32-2.38; P<.001).

Conclusions: JITAI systems have the potential to integrate timely support into the workplace. On the basis of our findings, we recommend that individual, contextual, and content-based factors be incorporated into the system for tailoring as well as for monitoring ineffective engagements across subgroups and contexts.

JMIR Ment Health 2024;11:e48974

doi:10.2196/48974

Keywords



Background

Work is a major source of stress in the United States, affecting more than half of Americans throughout most of the day [1]. Workplace stress leads to increased risk of mental and physical health disorders, decreased productivity and job satisfaction, and higher rates of accidents and employee costs [2-5]. Integrating stress reduction strategies directly into the workplace has proven to be effective and is widely recommended [6,7]. However, incorporating these individual-based techniques (eg, cognitive behavioral skills training, meditation, and exercise) [7,8] into the workday can be challenging due to work culture [9,10] or psychological barriers [9-11].

Information workers who frequently use computing technology face challenges, such as prolonged desk-bound and sedentary behaviors, that contribute to chronic physical and mental health issues [12,13]. High computer use without adequate breaks, high levels of multitasking, and constant connectivity demanded by information and communications technology has been found to be associated with increased stress and burnout [14-17]. Despite information and communications technology being associated with increased stress, workplace computing tools can also be leveraged to understand and reduce stress [18,19].

Passive sensing capabilities via ubiquitous devices have shown potential in health and well-being domains through monitoring and assessing individuals over time [20], and such data may be harnessed to provide precision mental health support. However, introducing new devices (eg, wearables) to an organization can be costly (eg, ≥US $100 per worker) and could be impractical for real-world, daily functional use with compliance and quality issues [21,22]. In contrast, everyday technologies commonly used at work (eg, webcam, keyboard, and software telemetry) can be harnessed for passing sensing and behavioral analysis, offering a more feasible approach to infer affect [23], physiology [24], attention [25], or stress [18,26].

Recently, technology-mediated support for mental health has generated interest for its ability to provide flexible and always-available access. However, these systems often provide generalized support that does not account for individual variabilities or contexts [27]. Despite an abundance of mental health apps [28,29], digital interventions that fit the specific workplace context are still highly sought after [27,30].

Digital mental health interventions (DMHIs), known as digital microinterventions, leverage technology to adapt existing evidence-based psychosocial interventions and leverage technology affordances to provide individual components of traditional psychotherapy focused on managing proximal symptoms (eg, relaxation for stress) in the hopes of achieving broad, distal objectives (eg, overcoming depression) [31]. Just-in-time adaptive interventions (JITAIs) have been introduced to deliver personalized, contextualized, and adaptable interventions using dynamic human behavior data captured through ubiquitous sensing technologies [27,32-34]. This concept has been explored in various health contexts, such as promoting physical activity, stress management, and weight management, with recent interest in applying them to positive coping skill use [35].

Despite their promises, JITAI systems are not yet pervasive with many applications still relying on ecological momentary assessments (EMAs [36]). Recent developments in algorithmic and machine learning approaches to dynamic adaptation have begun to show improvements in timing, receptivity, and engagement [37-40]. Operationalizing the adaptation of the system requires choosing appropriate tailoring variables and intervention options to drive the system’s decision-making regarding intervention timing and content. Prior research has examined factors associated with engagement in DMHIs [41,42], including personal-related (eg, demographics and personalities), content-related (eg, perceived fit and usefulness), and technology-related factors (eg, technical issues and privacy). However, most prior studies investigate study-long engagement, rather than examining “in-the-moment” engagement factors, which are crucial for improving the usability of interventions in real-world contexts [31], particularly because the integration of intervention use into life is a core facilitator for engagement [42].

A recent meta-analysis [36] has also shown that tailoring is an important aspect of JITAI design associated with greater efficacy. Despite JITAI systems’ potential to provide precision support [27], few studies demonstrate the value of just-in-time (JIT) support in improving user engagement [43,44]. Although the use of passively sensed data for contextual understanding and system adaptation is often recommended for the design of JITAI systems [27], many still rely mostly on EMAs, app use, or simple temporal features primarily from a single modality (ie, mobile devices) or without considering individual or intervention-related factors [36,44]. Passive sensing technologies offer numerous sets of contextual variables (eg, location, calendar, movement, and activity), presenting a challenge in designing sensing-capable JITAI systems: identifying a core set of tailoring variables among all possible variables that the system should consider for optimizing effective engagement [45]. Therefore, our primary goal and key contribution is to identify crucial tailoring variables that influence momentary engagement in digital stress-reduction microinterventions to inform the design of similar JITAI systems.

Summary

In this study, we analyze engagement data from a 4-week longitudinal deployment of a workplace stress-reduction intervention system [19]. We leverage everyday workplace devices as unobtrusive and passive sensors to gain a glimpse into participants’ daily work activities (eg, emails, meetings, and computer activity). This system leverages the Cloud to integrate passively collected data and EMAs to deliver JIT nudges to engage in digital microinterventions across devices. Unlike laboratory experiments or controlled studies, our study allows observing users’ moment-by-moment interactions with the system in naturalistic work environments. We combine passively sensed work contexts, the system use, including which intervention participants chose and liked, individual demographics, personality traits, and coping styles to understand how the work context influences engagement patterns and to understand the appropriate conditions that lead to momentary intervention engagement and positive outcomes. We leverage such data from 43 participants to contextualize 1585 system-initiated interventions. From statistical modeling of the impact of individual, contextual, and intervention-related factors on engagement outcomes, we confirmed that individual factors (eg, age, gender, personality traits, and coping skills), as well as contextual and content-related factors (eg, availability and intervention modality), significantly influenced momentary intervention engagement, intervention choice, user ratings, and stress reduction outcomes. These findings suggest tailoring guidelines for JITAI systems whereby contextual and personalized factors can be used to find a positive balance between user preferences and maximal intervention efficacy.

In preparing this paper, we referred to the Guidelines and Checklist for the Reporting on Digital Health Implementations (iCHECK-DH) [46], which can be found in Multimedia Appendix 1.


Objectives

There are multiple considerations that influence JITAI systems’ decision points (ie, “a time at which an intervention decision is made” [32]), such as the right timing for a prompt, the right intervention for the moment, or the intervention likely to be engaged in. Therefore, this study aimed to identify factors that may contribute to improving participant engagement for JITAI systems. The study builds on a pilot implementation and deployment of a JIT stress-reduction microinterventions in a real-world workplace setting, targeting information workers who spend most of their working hours at processing information with computing devices [47]. The deployed system did not adapt the intervention content based on changing state or user context and, therefore, is not a full JITAI system. Instead, the study conducted a retrospective analysis of the deployment data to understand factors that influence the engagement and efficacy of workplace stress-reduction JITAI systems for future development.

Participants

Information workers were recruited from a large technology organization via randomly distributed email advertisements. We recruited primarily US-based workers for the ease of system troubleshooting and to minimize any country-specific organizational factors. We enrolled participants on a first-come, first-served basis as long as we could satisfy system compatibility. To ensure system compatibility, interested participants completed a brief screener survey about their work setup (eg, primary device specification and operating system and web camera availability). Eligible participants, whose primary device specification met our sensing software requirements, were asked to install and run the study system on their primary desktop for 30 minutes. Only 43 participants that could run the sensing software on their desktop for ≥30 minutes were selected to participate in the study. The sample size of approximately 40 participants was determined for statistical power within a larger experimental study [19] that compared the JIT condition with a baseline condition for intervention effectiveness.

Our intake survey included several demographics measures such as age, gender, and role. Of the 43 participants, 29 (67%) identified as man. This distribution of gender closely aligns with the current industry demographics for large technology companies [48], and therefore, we consider this gender representation acceptable for our analysis. Of the 43 participants, 3 (7%) self-reported being aged 18 to 25 years old, 11 (25%) were 26 to 35 years old, 18 (42%) were 36 to 45 years old, 8 (19%) were 46 to 55 years old, 2 (5%) were 56 to 65 years old, and 1 (2%) was >66 years old. Moreover, 24 (56%) of 43 participants reported as being in engineering or development role, and 14 (33%) reported as being in sales or business strategy.

Other intake measures included the Depression, Anxiety, and Stress Scale-21 (DASS-21) [49]), the brief Big Five Personality Inventory-10 [50], the Emotional Regulation Questionnaire [51] designed to measure the tendency to regulate emotions through cognitive reappraisal and expressive suppression, and the 6-item brief resilience scale [52] that measures the ability to bounce back from stress. The average stress level of the stress subscale of DASS-21 reported by participants was 5 out of 21 (SD 3.8) which is within normal ranges. Our participants scored an average of 3.8 for agreeableness, 4.1 for conscientiousness, 2.6 for extraversion, 2.8 for neuroticism, and 3.5 for openness out of 5. Cognitive reappraisal was scored at 4.7 out of 7 (SD 1.1), expressive suppression was at 3.7 out of 7 (SD 1.3), and resilience was at 3.5 out of 5 (SD 0.9) on average.

Study Implementation and Procedure

We deployed our system on the 43 participants who consented to the study, for a period of 4 weeks. During every workday of the 4-week study period, the system asked users to complete 5 EMAs per day during their reported work hours to capture their subjective stress ratings from the past 30 minutes (ie, “How would you rate your level of stress during the last 30 minutes?”). When the system determined that the user’s stress level may be high, the system sent JIT nudges via a chatbot, asking users to engage in a stress reduction microintervention. In the background, the system captured use data as well as passively sensed contextual data. Detailed description of the system architecture is provided in section 1 of Multimedia Appendix 2 [5,15,19,33,53-72].

JIT Heuristics

The system determined higher-than-baseline stress levels based on our JIT heuristics informed by computed stress scores and self-reported stress levels. Stress scores were computed in real time per individual as an average of 5 components ranging between 0 and 1, each representing 5 components that previous work has identified as sources of stress: (1) the number of emails received [15,53], (2) the total number of meetings in a given day [54], (3) the percentage of time into the day [55], (4) the amount of facial expressions (via the Facial Action Coding System [56]) from corrugator (ie, brow furrowing) and lip depressor (ie, frowning) minus zygomatic major (ie, smiling) [55,57,58], and (5) heart rate [59,60]. Self-reported stress levels were obtained from EMAs.

In our JIT heuristics, first, we compute each user’s baselines as the average of the computed stress scores and self-reported stress levels based on the data from the first week of using the system. These individualized baselines (captured at week 1 of the 4-week study) are used as thresholds for delineating high stress from low stress during subsequent weeks (weeks 2 to 4). During the first week, we use the default baseline at the middle of the score range. Then, we send intervention nudges only if it is during the working hours that the users have stated at intake, if they have not explicitly scheduled an intervention at a later time that day, if they have not completed an intervention in the past hour, if there has not been an intervention nudge in the past 2 hours, and if there have not been ≥4 intervention nudges that day. Detailed description of our stress inference and heuristics is provided in section 1 of Multimedia Appendix 2.

Microinterventions

Microinterventions [31] used in our study were translated from components of cognitive behavioral therapy and dialectical behavioral therapy, 2 empirically supported and widely used psychotherapy modalities [61,62]). These were under 5-minute interventions that were either a short video, a single-turn text prompt, or a brief therapeutic conversation with the chatbot. The microinterventions used in our study can be categorized by (1) the function served for users, (2) the modality in which the intervention was delivered, and (3) the intended location to perform the intervention.

Microinterventions are primarily categorized into 3 functional categories that align approximately with the amount of effort required. “Get my mind off work” interventions are low-effort interventions designed to help users take their mind away from work with positive activities that promote emotion regulation [63]. “Feel calm and present” interventions are medium-effort interventions that help users feel calm and present by drawing inspiration from the mindfulness practices. “Think through my stress” interventions are high-effort interventions that help users think through their stress and directly address and resolve stress-inducing components of their lives. For simplicity, we refer to these 3 intervention categories as “distract,” “calm,” and “address,” respectively, in the rest of the paper. Overall, there were 18 interventions per functional category. Section 2 of Multimedia Appendix 2 provides more details about the microinterventions, including all categories and examples for each category.

User Engagement Flow

Figure 1 illustrates a series of user engagement steps when a system nudge is sent. When the system sends a nudge to perform an intervention (Figure 1A), users can choose to delay it to a later time that day (Figure 1B). The system does not send another nudge until that time. Users may ignore the nudge, and the system expires the nudge after 30 minutes of inactivity. If the user decides to engage in an intervention, they can choose from 3 intervention categories (Figure 1C), and then the system randomly selects a new intervention within the category. Before beginning the intervention, the users are asked to subjectively rate their current stress level (Figure 1D). Then, the intervention content is shown to the user (Figure 1E). Once the intervention is done, the users are asked to rate the intervention (Figure 1F) and rate their stress level again (Figure 1G) before concluding the intervention flow. An example screenshot of the full user engagement flow can be found in section 3 of Multimedia Appendix 2.

Figure 1. (A) System sends a nudge to users to perform an intervention. (B) Users can opt to postpone the intervention at a later time. (C) If users choose to do an intervention, they can select from 1 of 3 intervention categories. (D) Users first self-report their current stress level. (E) The system shows the intervention content for users to interact with. (F) User rates the intervention. (G) Users self-report their stress levels after the intervention.

Other Study Tasks

Beyond JIT-based interactions with the system, participants could also access interventions on demand, where they could perform the intervention at that moment or schedule it to a later time that day. Participants were also asked to complete morning surveys to log their sleep quality and evening surveys to log their food and beverage intake. We additionally asked participants to rate their sleep quality the night before via a morning survey and log their food and drink consumptions throughout the day via an evening survey. They also completed weekly surveys including the DASS-21 and the brief resilience scales. The exit survey solicited feedback about the usability of the system and the perceived helpfulness and impact of the interventions.

Poststudy Analysis

This section describes how we processed and analyzed the data collected from the above study to understand the factors contributing to improving participant engagement in a system-initiated intervention and the effectiveness of interventions at a given moment. Table 1 describes the full set of variables selected for our analysis and their descriptive statistics. Section 3 of Multimedia Appendix 2 includes additional detail, including descriptive statistics, inclusion or exclusion of data points, and correlational analyses conducted between variables.

Table 1. List of the per-participant, per-half hour, per-nudge, and per-intervention variables and their descriptive statistics.
VariableDistribution
Per-participant

Age (y)18-35 (n=14), 36-45 (n=18), >46 (n=11)

GenderMen (n=29), women (n=14)

Cognitive reappraisalMean 4.69 (SD 1.09); range 2-7

Expressive suppressionMean 3.74 (SD 1.26); range 2-6

ResilienceMean 3.51 (SD 0.89); range 2-5

AgreeablenessMean 3.79 (SD 0.74); range 2-5

ConscientiousnessMean 4.12 (SD 0.83); range 2-5

ExtraversionMean 2.62 (SD 0.86); range 1-4

NeuroticismMean 2.85 (SD 1.04); range 1-5

OpennessMean 3.48 (SD 0.79); range 2-5

Engagement skewnessMean –0.09 (SD 0.58); range –1.41 to 1.41
Per-half hour

Nudge probabilityMean 0.06 (SD 0.04); range 0.02-0.30
Per-nudge

Meeting countsMean 0.30 (SD 0.51); range 0-3

No meeting minutesMean 5.43 (SD 11.79); range 0-50

Self-event countsMean 0.12 (SD 0.37); range 0-3

Email messages sentMean 0.39 (SD 0.88); range 0-8

Email messages readMean 3.22 (SD 5.09); range 0-54

Chat messages countMean 3.88 (SD 5.90); range 0-64

Ad hoc call countMean 0.04 (SD 0.21); range 0-2

Number of attention signalsMean 1442.64 (SD 1316.68); range 0-5705

Nudge sourceJITa algorithm (n=1337), rescheduled (n=248)

EngagedTrue (n=563), False (n=1022)
Per-intervention (chosen)

CategoryAddress (n=112), calm (n=338), distract (n=200)

EngagedTrue (n=563), false (n=87)

ModalityVideo (n=113), prompt (n=384), conversation (n=24)

LocationAt desk (n=422), inside (n=84), outside (n=15)
Per-intervention (completed)

Stress reductionMean 0.29 (SD 0.53); range –1 to 3

RatingMean 3.61 (SD 1.02); range 1-5

ImprovedTrue (n=150), false (n=371)

LikedTrue (n=289), false (n=232)

Stress beforeMean 2.14 (SD 0.97); range 1-5

CategoryAddress (n=77), calm (n=275), distract (n=72)

ModalityVideo (n=60), prompt (n=340), conversation (n=24)

LocationAt desk (n=325), inside (n=84), outside (n=15)

aJIT: just-in-time.

Engagement Outcome Variables

From participants’ interactions with the system, we reconstructed each participant’s step-by-step interaction with the system as represented in Figure 1 to extract several outcome variables, such as whether a system-initiated intervention was completed and the effectiveness and rating of interventions once engaged. We note that the number of data points varies based on where in the engagement flow the participant exits.

Engagement Label

We labeled each system-initiated intervention as “engaged” in an intervention (ie, engaged=true) if the participant explicitly marked the intervention as done (ie, clicking on “I’m done” button in Figure 1E), regardless of whether they completed any subsequent prompts (ie, Figures 1F and G). Any ignored, incomplete, or timed-out nudges were considered not engaged. Each system-initiated intervention that was triggered at a postponed time (Figure 1B) was categorized as “rescheduled” (ie, nudge source=rescheduled). Although these rescheduled nudges looked identical to JIT nudges, we hypothesized that the participants would be more likely to engage in a system-initiated intervention if they postponed the intervention to a time that is more suitable for engagement.

Intervention Choice

The characteristics of interventions, such as location, modality, and effort level, are important to consider because certain interventions might not be feasible in certain situations (eg, the participant cannot go outside) [73,74]. If a participant chose an intervention category (Figure 1C), we marked that nudge with binary labels of distract chosen, calm chosen, or address chosen. Each of these interventions was further labeled with modality and location based on the specific intervention that the system chose within the category.

Intervention Effectiveness and Rating

We looked at 2 outcome metrics—momentary stress reduction and intervention rating. We further binarized these outcome metrics to determine if a certain intervention engagement improved stress (ie, self-reported stress rating was lowered after the intervention use) and if the participant liked the intervention (ie, rated as “good” or “very good”).

Tailoring Variables for Postdeployment Analysis

We used per-participant demographics and validated scale responses (coping skills and personality traits) to explain individual differences. We harnessed passively sensed data streams to explain the context surrounding the system initiation and intervention engagement. On the basis of the participants’ interactions with the system, we also account for the probability distribution of system nudges throughout the day and the participant’s likelihood of engaging in interventions in the morning or the afternoon.

Participant Characteristics

We included demographic variables, such as age and gender, and personality traits that have been shown to impact engagement [42]. Because of the small sample sizes on either end of the age groups, we combined the lower 2 and upper 3 age groups to create more balanced age groups (ie, 18-35, 36-45, and >46 years). We included emotion regulation [51] and resilience [52] as measures of a person’s ability to cope with stressors. We also included the Big Five personality traits [75-77] because they are known to impact stress [78,79] and engagement in mental health treatment [42].

Passively Sensed Context

To understand the context surrounding a system-initiated intervention, we leveraged 2 sources of passively sensed data: custom sensing software built by Microsoft Research and Viva Insights by Microsoft.

The custom sensing software ran on participants’ desktops and can capture activities that may not be associated with work, such as browsing the internet or using non–work related software. It captured data from the 5 components used to compute stress scores in real time, as described earlier. The sensing software also captured general user computer activity events such as mouse and keyboard interactions into a single metric, number of attention signals, which could be an important indicator for presence. We hypothesized that presence at the computer could lead to higher engagement in interventions as the nudges were designed to grab the attention of participants at work. The range of values for number of attention signals was fairly large (maximum=5705) compared with other variables, so we divided the values by 2000 to estimate a comparable coefficient and CIs during modeling (ie, to have odds ratios [ORs] within 2 decimal points). When interpreting the effect sizes, we corrected for this factor of 2000. We also hypothesized that the likelihood of engaging in a stress-reduction intervention during active participation in a meeting is low. Therefore, we included no meeting minutes to represent the total number of minutes without a scheduled meeting with others and self-event count as the total number of calendar events with only the participant as the attendee.

Viva Insights captures deidentified activity aggregates in 30-minute windows for Microsoft tools across all devices associated with an individual’s work account. From Viva Insights, we included meeting counts, ad hoc call count, and email messages sent or read. We excluded chat messages count from our analysis because the nudges were delivered through Microsoft Teams, and our data source cannot be used to discern if the messages were coming from the bot. Because Viva Insights data are limited to half-hourly windows, we associated the contextual metrics with each system-initiated intervention by taking the half-hour window that holds the nudge time stamp.

Both data sources had several overlaps or similar metrics, such as ones related to meetings or emails. Correlational analysis between 2 data sources is described in section 3 of Multimedia Appendix 2. It is important to note that contextual data at work can be noisy due to individual differences in the use of work tools for nonwork purposes (eg, personal use and subscriptions). We do not differentiate work versus nonwork data because it is challenging, for example, to isolate work-related emails as the only source of stress.

System-Initiated Intervention Probability

Although each of the system nudges could be considered as an independent, repeated observation, the timing of the system-initiated interventions was sometimes dependent on when EMAs were administered. Because the system’s JIT heuristic runs every 5 minutes to check if a nudge needs to be sent based on the stress score and the EMA stress ratings, the most likely hours for receiving a system nudge is shortly after the EMA, leading to each participant receiving more nudges during certain hours of the day than others. To account for such variability in receiving system-initiated interventions, we incorporate the momentary nudge probability in our analysis. Because the range of nudge probabilities is small (mean 0.06, SD 0.04), we multiply the measure by 100 to represent it in percentages.

Temporal Engagement Skewness

Prior research has found that different hours of the day were seen as good or not-so-good timing for stress interventions [33]. To examine if a certain participant has a temporal tendency to engage, we computed the Fisher-Pearson coefficient of skewness, or engagement skewness, on the hourly intervention engagements per participant. A positive engagement skewness means that the participants tend to engage at the beginning of their workday, and a negative engagement skewness means that the participants tend to engage toward the end of their workday. We used the skewness metric instead of simply looking at the engagement during the morning and the afternoon to account for individual differences in working hours. We incorporated this skewness per participant in our analysis.

Analysis

Taking all the tailoring and outcome variables into account, our analysis focused on estimating the effects of contextual, individual, and intervention characteristics on binary outcome variables (eg, engaged, distract chosen, and liked). Thus, we built a logistic regression model predicting each outcome based on a combination of per-participant characteristics, per-half-hour nudge probability, per-nudge contextual metrics, or per-intervention metrics as fixed effects. The outputs of the logistical regression models are presented as ORs, representing a ratio of odds (eg, probability of engaging vs probability of not engaging) under 2 different conditions (eg, being a woman vs not being a woman). The data processing was conducted using Python packages (eg, numpy, pandas, scipy, and seaborn) and the models were tested using R libraries (eg, lme4, car, and performance).

We determined the significance of the fitted model against the null hypothesis model using the analysis of deviance. We conducted ANOVA to estimate the significance of fixed effects. Because the data are unbalanced (ie, unequal number of observations for each level of a factor), we obtain ANOVA type II sums of squares [80]. For categorical variables of ≥3 levels (eg, age group and category), we estimated pairwise differences using Tukey HSD (honestly significant difference) procedure. Multicollinearity in fixed effects was tested using the variance of inflation factor, and none of our models exhibited a multicollinearity issue. Results for ANOVA, multicollinearity, and Tukey analyses are presented in section 4 of Multimedia Appendix 2.

Ethical Considerations

The study was reviewed by the Microsoft Research Institutional Review Board (OHRP IORG #0008066, IRB #IRB00009672) before the research activities and was formally approved. In addition to the ethics review, our study obtained approvals from Microsoft’s privacy, security, and legal review officers before conducting the study. All participants provided consent as part of the onboarding process and agreed for their deidentified data to be used for research purposes. The consent described the installation of our system, interaction with interventions, surveys, and joining the study data with cloud-based, device-independent telemetry data. All data, collected anonymously or otherwise, were collected and stored in a secured and access-controlled location. All data were joined and deidentified before analysis by the research team. For their participation and data, each participant was compensated with a US $400 Amazon gift card.

Implementation Considerations

The implementation of the system used in the deployment study has dependencies that restrict interoperability and sustainability. The system requires that the participants use Microsoft platforms (eg, M365, Windows, Viva Insights, and Teams) and own a decent desktop that can perform vision-based data processing on device. Our current implementation focused on the understanding of individual, contextual, and intervention-related factors to drive the tailoring of JITAI systems and does not allow generalizability beyond the supported architecture. However, we believe that recent advancements in generative artificial intelligence technologies and the implementation of interoperability layers across different technology ecosystems can enable more accessible implementation of the system. Because this study was a pilot implementation, any budget planning, sustainability model, or interoperability for sustained deployment were not in scope for this paper.


Intervention Engagement

To see which factors influenced engagement, we modeled engaged as a function of per-participant characteristics, per-half hour nudge probability, and per-nudge contextual metrics. The logistic regression model of engaged with all fixed effects was significantly different from the null hypothesis model (ie, engage ∼1; χ221=115.5; P<.001). Table 2 outlines the ORs and CIs for each predictor. Predictors engagement skewness, meeting count, number of attention signals, and nudge source remained significant after applying the Benjamini-Hochberg correction for multiple comparisons. Reviewing the coefficients of our fixed effects, we found that being aged >46 years, being a woman, higher cognitive reappraisal, higher number of attention signals, and receiving a rescheduled nudge were associated with a higher likelihood of intervention engagement. In contrast, higher engagement skewness and meeting count were associated with to a lower likelihood of intervention engagement.

Table 2. Odds ratios (ORs) and CIs for each predictor of engaged for all system-initiated nudgesa.
PredictorsEngaged, OR (95% CI)
Intercept0.1b (0.03-0.30)
Age group (y), reference (18-35)

36-451.17 (0.88-1.56)

>461.47c (1.07-2.02)
Gender reference (man)

Gender (woman)1.37c (1.05-1.79)
Cognitive reappraisal1.14c (1.01-1.29)
Expressive suppression0.92 (0.83-1.02)
Resilience1.05 (0.88-1.25)
Agreeableness0.87 (0.72-1.05)
Conscientiousness1.13 (0.96-1.33)
Extraversion1.03 (0.89-1.21)
Neuroticism0.97 (0.83-1.12)
Openness1.15 (0.99-1.34)
Engagement skewness0.64b (0.51-0.79)
Nudge probability1.02 (1.00-1.05)
Meeting counts0.62b (0.49-0.78)
No meeting minutes1 (0.99-1.01)
Self-event counts1.15 (0.85-1.54)
Email messages sent1.05 (0.92-1.20)
Email messages read1.01 (0.99-1.03)
Ad hoc call count0.89 (0.52-1.49)
Number of attention signals1.39b (1.17-1.66)
Trigger source reference (system)

Trigger source (rescheduled)1.77b (1.32-2.38)

aObservations=1585; Tjur R2=0.07.

bStatistically significant (P<.05) after Benjamini-Hochberg correction.

cStatistically significant (P<.05) before any correction.

Next, we modeled each of the 3 choice outcome measures—distract chosen, calm chosen, and address chosen—as a function of per-participant characteristics, per-half hour nudge probability, and per-nudge contextual metrics. Table 3 outlines the ORs and CIs for all models.

Table 3. Odds ratios (ORs) and CIs for each predictor of distract chosen, calm chosen, and address chosen for all participant-chosen interventions.
PredictorsDistract chosena, OR (95% CI)Calm chosenb, OR (95% CI)Address chosenc, OR (95% CI)
Intercept0.11d (0.01-0.73)1.84 (0.32-10.62)0.44 (0.04-4.52)
Age group (y), reference (18-35)

36-451.65d (1.03-2.67)0.62d (0.40-0.96)0.88 (0.46-1.65)

>460.71 (0.40-1.26)0.95 (0.57-1.57)1.6 (0.85-3.07)
Gender reference (man)

Gender (woman)0.62d (0.39-0.98)1.31 (0.87-1.96)1.29 (0.73-2.26)
Cognitive reappraisal1.01 (0.82-1.25)1.27d (1.06-1.53)0.67e (0.53-0.84)
Expressive suppression0.99 (0.83-1.19)1.06 (0.90-1.25)0.95 (0.75-1.18)
Resilience0.93 (0.69-1.25)0.87 (0.67-1.13)1.42d (1.00-2.02)
Agreeableness0.75 (0.55-1.03)1.02 (0.77-1.36)1.46 (0.99-2.20)
Conscientiousness1.30 (0.98-1.75)0.92 (0.71-1.18)0.74 (0.51-1.05)
Extraversion1.18 (0.91-1.53)1.00 (0.79-1.27)0.79 (0.57-1.11)
Neuroticism0.88 (0.68-1.14)1.00 (0.79-1.26)1.36 (0.99-1.89)
Openness1.67e (1.30-2.16)0.71d (0.57-0.88)0.83 (0.60-1.13)
Engagement skewness1.12 (0.74-1.68)1.11 (0.78-1.59)0.86 (0.53-1.38)
Nudge probability1.00 (0.96-1.05)1.01 (0.97-1.05)0.97 (0.92-1.03)
Meeting counts0.84 (0.55-1.25)1.07 (0.76-1.52)1.14 (0.71-1.77)
No meeting minutes1.00 (0.99-1.02)1.00 (0.99-1.01)1.00 (0.98-1.02)
Self-event counts1.08 (0.63-1.79)0.83 (0.53-1.30)1.19 (0.65-2.05)
Email messages sent1.05 (0.84-1.30)1.00 (0.82-1.22)0.99 (0.74-1.30)
Email messages read1.03 (0.99-1.07)0.99 (0.96-1.03)0.96 (0.90-1.01)
Ad hoc call count0.88 (0.36-1.94)0.93 (0.45-1.93)1.26 (0.45-3.05)
Number of attention signals0.65e (0.48-0.87)1.18 (0.92-1.53)1.33 (0.94-1.87)
Trigger source reference (system)

Trigger source (rescheduled)1.01 (0.61-1.63)0.76 (0.49-1.16)1.39 (0.81-2.35)

aObservations=650; Tjur R2=0.115.

bObservations=650; Tjur R2=0.059.

cObservations=650; Tjur R2=0.089.

dStatistically significant (P<.05) before any correction.

eStatistically significant (P<.05) after Benjamini-Hochberg correction.

For choosing distract interventions, we found that higher openness was associated with a higher likelihood. In contrast, being a woman and higher number of attention signals were associated with a lower likelihood of choosing distract interventions. We found that higher cognitive reappraisal was associated with a higher likelihood and lower openness was associated with a lower likelihood of choosing calm interventions. We also found that higher resilience was associated with a higher likelihood and lower cognitive reappraisal was associated with a lower likelihood of choosing address interventions.

To understand the effect of intervention choice on engagement, we modeled engaged as a function of per-participant characteristics, per-half hour nudge probability, per-nudge contextual metrics, and per-intervention characteristics. The logistic regression model of engaged was significant (χ227=69.6; P<.001). Table 4 outlines the ORs and CIs for each predictor. We found that a higher resilience, choosing a prompted-based intervention, and choosing a video-based intervention were associated with a higher likelihood of engagement. In contrast, choosing an intervention that could be performed inside was associated with a lower likelihood of engagement.

Table 4. Odds ratios (ORs), CI, and P values for each predictor of engaged after chosen, that is, engaged for all interventions after participants chose a categorya.
PredictorsEngaged after chosen, OR (95% CI)
Intercept0.05b (0.00-0.964)
Age group (y), reference (18-35)

36-451.06 (0.54-2.05)

>461.65 (0.71-4.02)
Gender (woman)1.65 (0.88-3.16)
Cognitive reappraisal1.11 (0.83-1.48)
Expressive suppression1.25 (0.95-1.57)
Resilience1.63b (1.05-2.59)
Agreeableness1.24 (0.77-2.00)
Conscientiousness0.81 (0.55-1.17)
Extraversion0.72 (0.48-1.08)
Neuroticism1.39 (0.96-2.07)
Openness1.24 (0.88-1.76)
Engagement skewness1.01 (0.60-1.73)
Nudge probability1.01 (0.95-1.07)
Meeting counts0.65 (0.41-1.06)
No meeting minutes1 (0.98-1.02)
Self-event counts1.94 (0.86-1.67)
Email messages sent1.16 (0.84-1.67)
Email messages read0.97 (0.93-1.03)
Ad hoc call count0.54 (0.23-1.33)
Number of attention signals0.9 (0.61-1.33)
Trigger source reference (system)

Trigger source (rescheduled)1.79 (0.85-4.10)
Engagement skewness1.01 (0.60-1.73)
Category reference (distract)

Category (calm)1.14 (0.60-2.14)

Category (address)0.55 (0.22-1.43)
Modality reference (conversation)

Modality (prompt)3.53b (1.35-9.60)

Modality (video)5.86b (1.69-21.68)
Location reference (at desk)

Location (inside)0.43b (0.23-0.83)

Location (outside)0.34 (0.11-1.18)

aObservations=650; Tjur R2=0.12.

bStatistically significant (P<.05) before any correction.

Intervention Effectiveness

Finally, we analyzed the factors associated with higher intervention rating and intervention effectiveness. The logistic regression model of liked as a function of per-participant characteristics, per-half hour nudge probability, per-nudge contextual metrics, and per-intervention metrics was significant (χ228=92.4; P<.001). The ORs and CIs for each predictor can be seen in Table 5. We found that being a woman, higher cognitive reappraisal, higher extraversion, and higher stress reduction were associated with a higher likelihood of liking the intervention. In contrast, being aged 36 to 45 years, higher expressive suppression and higher nudge probability were associated with a lower likelihood of liking the intervention.

Table 5. Odds ratios (ORs), CIs, and P values for each predictor of liked for all interventions that participants engaged ina.
PredictorsLiked, OR (95% CI)
Intercept5.95 (0.51-70.91)
Age group (y), reference (18-35)

36-450.51b (0.30-0.86)

>461.01 (0.55-1.83)
Gender reference (man)

Gender (woman)2.51c (1.51-4.20)
Cognitive reappraisal1.35b (1.08-1.69)
Expressive suppression0.82b (0.68-0.99)
Resilience1.10 (0.80-1.51)
Agreeableness0.72 (0.50-1.02)
Conscientiousness0.80 (0.58-1.09)
Extraversion1.46b (1.10-1.95)
Neuroticism0.84 (0.64-1.11)
Openness0.99 (0.75-1.30)
Engagement skewness0.73 (0.47-1.12)
Nudge probability0.94b (0.90-0.99)
Meeting counts0.93 (0.61-1.42)
No meeting minutes1.00 (0.99-1.02)
Self-event counts0.66 (0.40-1.10)
Email messages sent1.01 (0.79-1.28)
Email messages read1.00 (0.96-1.04)
Ad hoc call count0.66 (0.24-1.69)
Number of attention signals1.16 (0.84-1.59)
Trigger source reference (system)

Trigger source (rescheduled)0.78 (0.47-1.30)
Category reference (distract)

Category (calm)0.8 (0.49-1.29)

Category (address)0.94 (0.44-2.01)

Stress reduction2.36c (1.60-3.54)
Modality reference (conversation)

Modality (prompt)0.52 (0.17-1.53)

Modality (video)0.75 (0.22-2.47)
Location reference (at desk)

Location (inside)0.69 (0.40-1.19)

Location (outside)2.56 (0.79-9.27)

aObservations=521; Tjur R2=0.166.

bStatistically significant (P<.05) before any correction.

cStatistically significant (P<.05) after Benjamini-Hochberg correction.

When we examine the momentary stress ratings surrounding the system nudges and the intervention use (Figure 2A), we see that the distributions of EMA and preintervention and postintervention stress ratings sit higher than the individual thresholds used for sending the nudges. The average stress rating measured just before the intervention use (mean 2.14) is higher than the average stress rating used for system nudges (mean 1.73), indicating that the participants were more likely to engage in interventions when their stress ratings were higher than average (Figure 2B).

Figure 2. (A) Distribution of momentary stress ratings. A kernel density estimate plot of subjective stress ratings shows that the distribution of stress ratings when the system sent intervention nudges is higher than the individual thresholds. Stress ratings before intervention are distributed at a higher rating than stress ratings after intervention and the ratings when the nudges were sent. (B) Comparison of momentary stress ratings. Average momentary stress ratings at different points in time with 95% CIs.

We then examined the factors associated with stress reduction. The logistic regression model of improved was significant (χ230=186.7; P<.001), and the resulting ORs and CIs per predictor can be found in Table 6. We found that higher agreeableness, higher nudge probability, higher stress before, higher rating, and getting prompt-based or video interventions were associated with a higher likelihood of improvement on their stress rating. In contrast, being a woman, higher cognitive reappraisal, higher neuroticism, and choosing calm or address interventions were associated with a lower likelihood of improvement on their stress rating.

Table 6. Odds ratios (ORs), CIs, and P values for each predictor of improved for all interventions that participants engaged ina.
PredictorsImproved, OR (95% CI)
Intercept0.00b (0.00-0.00)
Age group (y), reference (18-35)

36-450.58 (0.30-1.13)

>460.98 (0.49-1.98)
Gender reference (man)

Gender (woman)0.41b (0.21-0.77)
Cognitive reappraisal0.69b (0.52-0.91)
Expressive suppression1.01 (0.79-1.28)
Resilience0.84 (0.55-1.25)
Agreeableness1.73c (1.10-2.76)
Conscientiousness1.27 (0.86-1.89)
Extraversion0.89 (0.62-1.29)
Neuroticism0.57b (0.39-0.81)
Openness1.09 (0.78-1.55)
Engagement skewness1.18 (0.68-2.03)
Nudge probability1.09b (1.03-1.16)
Meeting counts0.85 (0.48-1.42)
No meeting minutes1.01 (0.99-1.03)
Self-event counts1.29 (0.69-2.36)
Email messages sent1.04 (0.78-1.39)
Email messages read0.96 (0.90-1.01)
Ad hoc call count0.94 (0.28-2.63)
Number of attention signals1.4 (0.95-2.05)
Trigger source reference (system)

Trigger source (rescheduled)1.08 (0.57-2.02)
Category reference (distract)

Category (calm)0.43b (0.23-0.78)

Category (address)0.40c (0.16-0.97)

Stress before5.76b (3.98-8.64)

Rating2.47b (1.84-3.39)
Modality reference (conversation)

Modality (prompt)6.65c (1.53-36.45)

Modality (video)5.62c (1.12-34.10)
Location reference (at desk)

Location (inside)1.17 (0.60-2.25)

Location (outside)0.23 (0.03-1.10)

aObservations=521; Tjur R2=0.338.

bStatistically significant (P<.05) after Benjamini-Hochberg correction.

cStatistically significant (P<.05) before any correction.


Principal Findings

In this paper, we presented a comprehensive and systematic study that simultaneously encompassed trait-based factors (ie, individual characteristics), state-based factors (ie, workplace contextual and behavioral signals and momentary stress), and intervention-related factors (ie, location and function) to identify what drives JIT intervention engagement and efficacy. This study particularly focuses on momentary outcomes to inform the design of dynamic tailoring, which is a key component of JITAI systems. In this study, we leveraged surveys, EMA, and passively-sensed data from the deployment of the JIT stress-reduction intervention system to identify significant factors that influence the momentary engagement, the choice of interventions, the engagement given an intervention choice, the user rating of interventions engaged, and the stress reduction from the engagement.

We found that stress ratings immediately after the JIT interventions were significantly lower than those reported in the moments immediately before the interventions (Figure 2B). While keeping in mind the study sample, we found that women, those with higher neuroticism, those with higher cognitive reappraisal skills, and those that chose calm and address interventions were significantly less likely to experience stress reduction, while those with higher agreeableness and those that chose prompt-based or video-based interventions were significantly more likely to experience stress reduction. Surprisingly, contextual signals, such as meeting or email load, did not lead to a significant increase or decrease in stress ratings, which may indicate that trait-based or intervention-related factors matter more for efficacy or that the contextual signals may need finer granularity.

In contrast, we found that contextual signals such as lower meeting counts and lower engagement skewness were associated with a higher likelihood of engagement, indicating that state-based contextual factors such as being in a meeting or time of the day may matter more for engagement than efficacy. In addition, a JIT intervention that was explicitly rescheduled to a later time was more likely to be engaged. This implies that intervention engagement can be improved by giving some level of control to the users over complete automation.

With respect to the choice of interventions and liking the interventions, factors that significantly influenced the choice were primarily trait-based. Higher openness was associated with a higher likelihood of choosing distract but a lower likelihood of choosing calm, while higher cognitive reappraisal was associated with a higher likelihood of choosing calm but lower likelihood of choosing address. Higher resilience was associated with a higher likelihood of choosing address as well as the subsequent engagement after the choice. Interventions were liked more by women, those with higher cognitive reappraisal, and those with higher extraversion, and when the interventions led to stress reduction. One state-based factor that stood out is the number of attention signals, which negatively impacted the likelihood of choosing distract interventions. Higher number of attention signals indicates higher desk-bound work activity (eg, keyboard and mouse) where getting the mind off work (ie, how distract interventions were communicated) during intense work activities may not be feasible or desired.

In summarizing our findings, we first categorize these factors into (1) nonmodifiable individual factors, (2) modifiable individual factors, (3) contextual factors, and (4) content factors.

The distinction between modifiable and nonmodifiable individual factors is important for intervention design. Once modifiable factors are identified, strategies can be deployed to directly influence those factors if those strategies can lead to a greater impact on the engagement or the efficacy of the interventions. Nonmodifiable factors are also important to determine which subset of populations can benefit from additional targeted support through organizational or policy-level changes [81]. In our analysis, nonmodifiable individual factors include gender and age, and modifiable individual factors, although debatable [82], include emotion regulation skills and personality traits. Our findings revealed that these individual factors not only influence study-long engagement [42] but also influence momentary engagement and can be useful for tailoring JITAI systems. Although our findings for individual factors corroborate with prior studies, it is important to highlight that our study evaluates instance-level engagement metrics through play-by-play analysis of app-use behaviors rather than study-long engagement metrics. Contextual factors such as workload (eg, meeting counts and email counts) and availability (eg, presence or activity at the computer) are helpful in the JITAI system’s decision-making process for when to interrupt the user. Content factors include intervention-related information such as the amount of effort required, the modality of intervention delivery, and the ideal location for intervention engagement, and these factors can inform JITAI systems in determining which intervention to present to the user given the understanding of the current context.

In this section, we summarize and discuss the findings organized by these 4 categories of factors with recommendations for design and future research.

Nonmodifiable Individual Factors

In our analysis, we found several significant effects of gender and age on our outcome measures. Participants who self-reported as being a woman had more than twice the likelihood of liking interventions than being a man, despite having less likelihood of improving from engaging in interventions. Being a woman also had 38% less likelihood of choosing distract interventions than being a man, suggesting that there may be an unobserved motivational factor. For example, although not statistically significant, women reported higher momentary stress on average compared with men (2.15 for women vs 1.84 for men; t20.756=–1.545; P=.14), which may contribute to choosing more calm or address interventions to reduce their stress. Although the general findings from DMHI studies that women are more likely to engage in digital interventions than men are corroborated by our analysis [42], the fact that women improve less despite engaging more, liking interventions more, and choosing more address (ie, high reward) interventions is a concern for systems design that only take engagement metrics into account.

Participants aged >46 years were 47% more likely to engage in interventions while being aged 36 to 45 years increased the likelihood of choosing distract interventions by 65% and decreased the likelihood of liking interventions by 49% compared with those aged 18 to 35 years. Although the effects of different age groups on engagement have mixed results across prior studies, the higher engagement rate for participants aged >46 years in the study could be explained by a higher rate of interest in digital interventions for older populations [83]. Participants in the 36 to 45 years age group choosing distract interventions and not liking them highlight an opportunity for finding different types of interventions that they might enjoy.

Tailoring JITAI systems to individuals has been suggested to improve engagement and efficacy of interventions in prior research [36]. Our findings suggest that certain age or gender groups may benefit more than others from our system. The lack of engagement or the lack of improvement despite engagement for some groups highlights opportunities for targeting research and design efforts to further understand unobserved barriers to engagement and effective responses. Although it is generally recommended that intervention content be tailored based on individuals, we also recommend that the efficacy and user rating of interventions be closely monitored to ensure that certain groups enjoy the same level of benefits as other groups. In addition, new intervention content could be codesigned with groups that may not be benefiting as much and added to the system on a regular basis to equalize outcomes across subgroups.

Modifiable Individual Factors

Our findings corroborate prior research that personality trait is a strong factor in the engagement of DMHIs [42]. A prior study found that openness to experience is associated with better adherence and lower odds of attrition [84]. In our analysis, we did not find a significant effect on engagement, but we found that openness significantly increases the likelihood of choosing distract but decreases the likelihood of choosing calm interventions. One possible explanation may be the variety in the intervention content, which people with high openness would prefer [85]. Distract interventions tend to offer more variety in content with videos of travel destinations and humor as well as opportunities to explore social connections, whereas calm interventions were mostly introspective activities such as breathing or focused observations.

Although the effects were only moderately significant (P<.06), we found that an increase by one point in neuroticism and agreeableness scales increases the likelihood of choosing address interventions, which were designed to help users directly address resolve the stress-inducing components of their lives. For participants with high scores in neuroticism and agreeableness scales, it is possible that interventions that help them directly address their stress were more appealing than others that were designed to distract from stress and refocus on the present. Prior study has also found neuroticism and agreeableness to associate with a stronger interest in the use of stress management apps [86]. We also found that one point increase in agreeableness scale was associated with an increased likelihood of improvement by 73%, whereas the same point increase in neuroticism scale was associated with a lower likelihood of improvement by 43%, despite both having higher tendency to choose address interventions. Because agreeableness is known to be positively associated with the therapeutic alliance in mental health treatments [87], it is possible that the prosocial and cooperative nature of those with high agreeableness [88] allowed them to fully engage in the address interventions that were more action-oriented, leading to a greater improvement. In contrast, neuroticism has been known to negatively correlate with adherence to mental health recommendations [89] and to a wide variety of mental health treatment outcomes [87].

Prior work has associated extraversion with lower interest in using web-based mental health services over face-to-face interactions with a provider [83]. Although our interventions were delivered solely through technology, we found that one point increase in extraversion scale increases the likelihood of liking the intervention by 46%. We found no significant association between the intervention content (category, modality, or location) and liking them to explain this behavior. Future research should explore how extraversion facilitates momentary engagement in interventions.

Corroborating with prior study-long findings that associate personalities with DMHI engagement, our study found that personality traits also influence momentary engagement and efficacy of interventions. We recommend that JITAI systems carefully monitor potentially unhelpful use behaviors by incorporating personality traits in the system adaptation algorithm because they may impact the choice of interventions that may lead to negative downstream effects on outcomes. For example, it may be beneficial to offer a variety of more effortful interventions for people who report higher scores in openness. For people who report higher scores in neuroticism, the system could suggest less effortful interventions.

Across the board, emotion regulation styles had significant effects. One point increase in the cognitive reappraisal scale was associated with a 14% increase in the likelihood of engaging, a 27% increase in the likelihood of choosing calm interventions, a 35% increase in the likelihood of liking the intervention, a 33% decrease in the likelihood of choosing address interventions, and a 31% decrease in the likelihood of improvement after engagement. It is important to note the differences in the range of point scales. An increase of 14% for a 7-point scale is equivalent to an increase of 20% for a 5-point scale. In contrast, one point increase in the expressive suppression scale was associated with an 18% decrease in the likelihood of liking the intervention. We also found that one point increase in the resilience scale increases the likelihood of choosing address interventions by 42% and increases the likelihood of subsequently engaging in interventions by 63%.

Prior research has explored the role of emotion regulation in stress coping. For example, emotion regulation skills help assess stressful situations and determine the appropriate emotional response [90] or act as buffers against the negative effects of stress [91]. Emotion regulation has also been theorized as a moderator for increased resilience after encountering a stressful situation [92]. Although our analysis cannot claim the causal direction between coping skills and engagement, our findings suggest that emotion regulation and resilience may play a role in not only the stress-coping process but also in choosing different interventions or deciding to engage in an intervention. While most research has argued for increasing coping skills as an outcome measure or a treatment target [93], our findings suggest promising new research directions in understanding how coping skills could impact our decisions to engage in therapeutic interventions in the moment.

The role of emotion regulation strategies in altering our decisions and choices in various contexts has been empirically studied in highly controlled laboratory settings [94]. Prior research has also studied personality traits [82,95] and coping skills [93] as “states” that exhibit intrapersonal variations, and such states can be modified through therapeutic strategies (eg, cognitive behavioral therapy). Although our study has assumed personality traits, emotion regulation skills, and resilience to be stable over the course of the study, it remains to be seen how shifting the perspectives of these characteristics to be more dynamic would inform the JITAI systems. Because of the potential mediating role of coping skills in perceived stress [96] and the role of perceived stress in outcome improvement (ie, one point increase in stress rating before the intervention leads to being >5 times more likely to improve in our findings), coping skills should be measured periodically and incorporated into the decision-making process of JITAI systems. Therefore, we recommend further research to explore how taking a dynamic approach to personality and coping skills would inform the design of JITAI systems.

Contextual Factors

We found that the more the participant tends to engage at the end of the day, the more likely they would engage in interventions and that the system-initiated interventions that were rescheduled to a later time increased the likelihood of engagement. These findings may suggest a tendency to defer interventions to later in the day. Prior work that applied the self-determination theory [97] to JITAI systems suggested that perceived competence and self-regulation abilities may deplete throughout the day, potentially leading to unhealthy choices (eg, unhealthy food and alcohol) toward the end of the day [34]. This has a serious consequence for those who tend to procrastinate or postpone healthy behaviors (eg, exercise, and stress intervention) toward the end of the day. In contrast, allowing people to defer an intervention to a specific time may increase self-efficacy and perception of control, which might lead to an increased chance of behavior change [98]. In fact, Howe et al [19] reported that many participants liked the ability to perform the interventions when they wanted. We also found that the increase in nudge probability decreases the likelihood of liking the intervention but increases the likelihood of improvement, revealing that a JIT intervention might be “a bitter pill to swallow” but a useful pill, nonetheless. However, a relentless reminder could lead to distraction [99] and eventual system abandonment [44].

Our findings show encouragement that intelligent timing based on contextual information could improve the engagement and effectiveness of interventions, but perhaps at the cost of a lowered sense of user agency and control, negative perceptions toward the interventions, or leading to unhealthy choices toward the end of the day. It also highlights that relying solely on what people like may not be helpful for stress reduction and that there needs to be an additional investigation into how momentary factors could influence user ratings. Therefore, future JITAI systems should carefully balance user ratings (eg, likes) with intervention efficacy (eg, stress reduction and engagement) and help users discover what works best for them. Considering that higher user ratings may not always reflect the effectiveness of the interventions, such systems should aim to simultaneously explore the rating and the improvement in determining the timing of interventions.

In evaluating the work context, as expected, we found that a nudge sent at a time when the user is less likely to be in a meeting but more active at the desk improves engagement. Contrary to our hypothesis, we found no significant associations with no meeting minutes or self-event counts. It is possible that there is high variability in the level of focus and attention needed during times carved out for self. For example, our data sources cannot discern if the times carved out for self were work-related (eg, focus time for reading and writing) or non–work related (eg, running errands, child pick up, and exercise). This study’s data sources cannot achieve automatic detection of activities beyond basic work activities, such as meetings, emails, chats, calls, or computer activities.

Although tailoring to the activity context is the defining promise of JITAI systems, automatically detecting the activities performed within a time window is not an easy task. In addition, how much tailoring and the level of system intelligence are really necessary to maximize engagement and outcomes is unknown, especially given the cost and risks associated with the invasion of privacy in passive sensing. Further research is necessary to understand the cost and benefit of accurate activity detection and intelligent timing on the engagement and effectiveness of JITAI systems.

Content Factors

In our analysis, intervention categories, modalities, and locations showed pronounced effects on engagement and improvement, suggesting the importance of the intervention content in the design of JITAI systems. We found that having an intervention that could be done at the desk more than doubled the likelihood of engagement compared with an intervention that could be done indoors, but not at the desk. It is possible that leaving the desk at the moment of the nudge was not appropriate given the situation or there was an unobserved motivational barrier. In these scenarios, suggesting a different activity, rescheduling the activity, or waiting until the next appropriate transition time might have been beneficial. To improve engagement, intervention designers could provide additional desk-based stress reduction techniques to minimize the burden of leaving the desk.

In contrast, we found that having a prompt- or video-based intervention increased the likelihood of engagement by more than 3-fold and improvement by more than 5-fold, compared with a conversation-based intervention. Although prompt- and video-based interventions were typically less effortful than conversation-based interventions that require many turn-taking interactions with the bot, conversation-based interventions were designed to address the sources of the stress with the hope of creating a longer-lasting impact. It is possible that the conversations were not usable for participants to fully engage in the content. This finding suggests that quick, effortless interventions could be useful at the moment, but complex, turn-taking interventions need more thoughtful redesign.

Our findings revealed that the effects of intervention content types were significant and large. To investigate the impact of choices on improving outcomes and user ratings, future digital intervention systems that provide a catalog of interventions would benefit from characterizing each intervention on multiple dimensions, such as the level of effort, location, modality, etc. Such characterization would help understand the interplay among personality traits, contextual cues, and intervention types that users are likely to choose and benefit from. Therefore, we recommend that JITAI systems provide a variety of interventions to fine-tune its recommendations based on contexts but also to identify interventions that may need redesign.

Limitations

Our system design and analysis setup have several limitations. The retrospective analysis of the inferred stress scores indicates that there is room for improvement. Our stress scores were computed from a generalized algorithm, but recent studies have shown that stress can be idiosyncratic and, therefore, needs to be modeled at an individual level [18]. Our user engagement flow cannot differentiate the dismissal of the nudge due to bad timing or low stress or both. Our analysis setup does not allow for determining the causal relationship between the individual, contextual, and content factors with engagement, stress reduction, and intervention ratings. Even though the deployment study collected long-term stress measures via the DASS-21 scale, our analysis was limited to momentary stress ratings because microinterventions are more appropriate for proximal outcomes over distal outcomes [31]. Microrandomized trials are a promising research direction for JITAI systems to quantify the impact of tailored interventions on both short- and long-term outcomes [100] with careful considerations for the appropriate sample size [101]. Our data were also limited by a small sample population that exhibited low-stress levels, focusing on US-based information workers from a large technology organization, skewed toward engineers and those that identified as man. Prior work has suggested that, when evaluating engagement (or attrition), the severity of symptoms should also be considered [41]. Therefore, future research should evaluate the system through microrandomized trials with a sample population exhibiting high severity of stress symptoms while also expanding into other information worker roles from different sectors.

The deployment of the system used in the study is limited by the system requirements. Our system assumes that a single organization may offer a standardized set of software tools to workers for several reasons (eg, security, privacy, productivity, compliance, and scalability) and does not support interoperability across different software tools used by workers within an organization or across organizations. Beyond this challenge, it is important to note that the data required to describe an individual worker’s stress necessarily crosses work–non-work boundaries, and such boundaries are increasingly blurred by technologies [102] and remote work [103]. Our contextual data captured at work were noisy and were aggregated without peeking into the content (eg, of email or documents) to preserve privacy; in addition, it is challenging to separate blurred contexts. Such blurred boundaries raise concerns surrounding boundary preferences, data ownership, values and incentives, well-being definitions, and power dynamics that may undermine the successful deployment of personal stress support systems [104]. Perhaps more important than the interoperability of a system or clean data is a human-centered implementation process that allows for flexibility and adaptation of the system based on multistakeholder perspectives and use over time. This paper piloted one aspect of such adaptation. However, much work is needed in realizing the full potential of JITAI systems that take sociotechnical considerations into account and where users can participate in drawing the boundary and providing the right level and fidelity of data that meets their needs.

Conclusions

JITAI systems have the potential to integrate timely support into the workplace. To identify key factors that should be incorporated into the system’s adaptable decision-making process, we analyzed data from a 4-week deployment of a JIT workplace stress reduction microintervention system. On the basis of our findings, we recommend that individual, contextual, and content-based factors be incorporated into the system for tailoring as well as for monitoring unhelpful use behaviors across subgroups and contexts. Future work should explore careful balancing of individual preferences, intervention efficacy, and system accuracy to help users discover what works best for them and to continuously improve system recommendations.

Acknowledgments

This research was funded by Microsoft Research. TA was supported in part by the National Science Foundation (NSF) (grant CAREER IIS-2142794; NSF grant IIS-1901386) and the National Institutes of Health (grant R01MH125179). The authors are deeply grateful to the participants for their time and willingness to install our system, engage in our interventions, and provide valuable feedback and data. This study would not have been possible without the guidance and support of Dr Patricia Areán, Dr James Fogarty, Dr Eric Horvitz, and the Microsoft Research Human Understanding and Empathy group. We thank Ann Paradiso for generating the title image for this article.

Data Availability

The data sets generated during or analyzed during this study are not publicly available due to privacy and legal restrictions but are available from the corresponding author on reasonable request with a clear justification and a license agreement. The request will be reviewed and approved case by case by Microsoft Research Release and Compliance team, at which point a license agreement will be drafted and shared.

Conflicts of Interest

The research was conducted when JS, EH, RL, JH, KS, and MC were employed at Microsoft Research. JS, EH, JH, and MC were involved in the design, development, and deployment of the innovations and software evaluated in this study.

Multimedia Appendix 1

Completed Checklist of Guidelines and Checklist for the Reporting on Digital Health Implementations (iCHECK-DH).

PDF File (Adobe PDF File), 122 KB

Multimedia Appendix 2

(1) Detailed system architecture and stress inference methods, (2) intervention design and example interaction flows, (3) detailed description of how interaction and context data were extracted, and (4) additional results from ANOVA type II and multicollinearity analyses conducted for each logistic regression model.

PDF File (Adobe PDF File), 2820 KB

  1. Ray J. Americans' stress, worry and anger intensified in 2018. Gallup. 2019. URL: https://news.gallup.com/poll/249098/americans-stress-worry-anger-intensified2018.aspx [accessed 2021-08-20]
  2. Boudreaux E, Mandry C, Brantley PJ. Stress, job satisfaction, coping, and psychological distress among emergency medical technicians. Prehosp Disaster Med. Jun 28, 2012;12(4):9-16. [CrossRef]
  3. Elizabeth B, Kahaema B, Peter JG, Liu C, Prather AA, Thomas K, et al. Stress and health disparities: contexts, mechanisms, and interventions among racial/ethnic minority and low-socioeconomic status populations. American Psychiatric Association. 2017. URL: http://www.apa.org/pi/health-disparities/resources/stress-report.aspx, [accessed 2024-04-29]
  4. Colligan TW, Higgins EM. Workplace stress: etiology and consequences. J Workplace Behav Health. Jul 25, 2006;21(2):89-97. [CrossRef]
  5. Westman M, Etzion D. The impact of vacation and job stress on burnout and absenteeism. Psychol Health. Sep 2001;16(5):595-606. [CrossRef] [Medline]
  6. Cooper CL, Cartwright S. An intervention strategy for workplace stress. J Psychosom Res. Jul 1997;43(1):7-16. [CrossRef] [Medline]
  7. Richardson KM, Rothstein HR. Effects of occupational stress management intervention programs: a meta-analysis. J Occup Health Psychol. Jan 2008;13(1):69-93. [CrossRef] [Medline]
  8. Kinman G, Jones F. Lay representations of workplace stress: what do people really mean when they say they are stressed? Work Stress. Feb 23, 2007;19(2):101-120. [CrossRef]
  9. Dollard MF. Psychosocial safety climate: a lead indicator of workplace psychological health and engagement and a precursor to intervention success. In: Biron C, Karanika-Murray M, Cooper C, editors. Improving Organizational Interventions for Stress and Well-Being: Addressing Process and Context. New York, NY. Routledge; 2012:77-110.
  10. Herrero SG, Saldaña MÁ, Rodriguez JG, Ritzel DO. Influence of task demands on occupational stress: gender differences. J Safety Res. Dec 2012;43(5-6):365-374. [CrossRef] [Medline]
  11. Helbig S, Fehm L. Problems with homework in CBT: rare exception or rather frequent? Behav Cogn Psychother. Jun 14, 2004;32(3):291-301. [CrossRef]
  12. Lee E, Kim Y. Effect of university students' sedentary behavior on stress, anxiety, and depression. Perspect Psychiatr Care. Apr 24, 2019;55(2):164-169. [FREE Full text] [CrossRef] [Medline]
  13. Yuhan L, Bongshin L, Donghee YW, Amanda LR, David EC, Eun KC. Time for break: understanding information workers' sedentary behavior through a break prompting system. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 2018. Presented at: CHI '18; April 21-26, 2018:1-14; Montreal, QC. URL: https://dl.acm.org/doi/10.1145/3173574.3173701 [CrossRef]
  14. Thomée S, Härenstam A, Hagberg M. Computer use and stress, sleep disturbances, and symptoms of depression among young adults – a prospective cohort study. BMC Psychiatry. Oct 22, 2012;12(1):176. [CrossRef]
  15. Gloria M, Yiran W, Melissa N. Stress and multitasking in everyday college life: an empirical study of online activity. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 2014. Presented at: CHI '14; April 26-May1, 2014:41-50; Toronto, ON. URL: https://dl.acm.org/doi/10.1145/2556288.2557361 [CrossRef]
  16. Day A, Paquet S, Scott N, Hambley L. Perceived information and communication technology (ICT) demands on employee outcomes: the moderating effect of organizational ICT support. J Occup Health Psychol. Oct 2012;17(4):473-491. [CrossRef] [Medline]
  17. William CM. Who?s running the machine? A theoretical exploration of work stress and burnout of technologically tethered workers. J individ Employ Rights. 2007;12(3):263. [CrossRef]
  18. Mehrab BM, Javier H, Daniel M, Jina S, Esther H, Kael R, et al. Advancing the understanding and measurement of workplace stress in remote information workers from passive sensors and behavioral data. In: Proceedings of the 10th International Conference on Affective Computing and Intelligent Interaction. 2022. Presented at: ACII '22; October 18-21, 2022:1-8; Nara, Japan. URL: https://ieeexplore.ieee.org/document/9953824 [CrossRef]
  19. Howe G, Suh J, Bin Morshed M, McDuff D, Rowan K, Hernandez J, et al. Design of digital workplace stress-reduction intervention systems: effects of intervention type and timing. In: Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 2022. Presented at: CHI '22; April 29-May 5, 2022:1-16; New Orleans, LA. URL: https://dl.acm.org/doi/10.1145/3491102.3502027 [CrossRef]
  20. Cornet VP, Holden RJ. Systematic review of smartphone-based passive sensing for health and wellbeing. J Biomed Inform. Jan 2018;77:120-132. [FREE Full text] [CrossRef] [Medline]
  21. Martinez GJ, Mattingly SM, Robles-Granda P, Saha K, Sirigiri A, Young J, et al. Predicting participant compliance with fitness tracker wearing and ecological momentary assessment protocols in information workers: observational study. JMIR Mhealth Uhealth. Nov 12, 2021;9(11):e22218. [FREE Full text] [CrossRef] [Medline]
  22. Gonzalo JM, Stephen MM, Shayan M, Subigya KN, Andrew TC, Anind KD, et al. On the quality of real-world wearable data in a longitudinal study of information workers. In: Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops. 2020. Presented at: PerComWorkshops '20; March 23-27, 2020:1-6; Austin, TX. URL: https://ieeexplore.ieee.org/document/9156113 [CrossRef]
  23. Daniel M, Kael R, Piali C, Jessica W, ThuVan P, Mary C. A multimodal emotion sensing platform for building emotion-aware applications. arXiv. Preprint posted online March 28, 2019. [FREE Full text]
  24. Poh M, McDuff DJ, Picard RW. Advancements in noncontact, multiparameter physiological measurements using a webcam. IEEE Trans Biomed Eng. Jan 2011;58(1):7-11. [CrossRef]
  25. Gloria M, Shamsi TI, Mary C, Paul J. Bored Mondays and focused afternoons: the rhythm of attention and online activity in the workplace. In: Proceedings of the 2014 SIGCHI Conference on Human Factors in Computing Systems. 2014. Presented at: CHI '14; April 26-May 1, 2014:3025-3034; Toronto, ON. URL: https://dl.acm.org/doi/10.1145/2556288.2557204 [CrossRef]
  26. Javier H, Pablo P, Asta R, Mary C. Under pressure: sensing stress of computer users. In: Proceedings of the 2014 SIGCHI Conference on Human Factors in Computing Systems. 2014. Presented at: CHI '14; April 26-May 1, 2014:51-60; Toronto, ON. URL: https:/​/dl.​acm.org/​doi/​10.1145/​2556288.​2557165#:~:text=During%20the%20stressful%20conditions%2C%20the,75%25%20of%20the%20participants) [CrossRef]
  27. Bidargaddi N, Schrader G, Klasnja P, Licinio J, Murphy S. Designing m-Health interventions for precision mental health support. Transl Psychiatry. Jul 07, 2020;10(1):222. [FREE Full text] [CrossRef] [Medline]
  28. Huguet A, Rao S, McGrath PJ, Wozney L, Wheaton M, Conrod J, et al. A systematic review of cognitive behavioral therapy and behavioral activation apps for depression. PLoS One. May 2, 2016;11(5):e0154248. [FREE Full text] [CrossRef] [Medline]
  29. Neary M, Schueller SM. State of the field of mental health apps. Cogn Behav Pract. Nov 2018;25(4):531-537. [FREE Full text] [CrossRef] [Medline]
  30. de Korte EM, Wiezer N, Janssen JH, Vink P, Kraaij W. Evaluating an mHealth app for health and well-being at work: mixed-method qualitative study. JMIR Mhealth Uhealth. Mar 28, 2018;6(3):e72. [FREE Full text] [CrossRef] [Medline]
  31. Baumel A, Fleming T, Schueller SM. Digital micro interventions for behavioral and mental health gains: core components and conceptualization of digital micro intervention care. J Med Internet Res. Oct 29, 2020;22(10):e20631. [FREE Full text] [CrossRef] [Medline]
  32. Nahum-Shani I, Smith SN, Spring BJ, Collins LM, Witkiewitz K, Tewari A, et al. Just-in-time adaptive interventions (JITAIs) in mobile health: key components and design principles for ongoing health behavior support just-in-time adaptive interventions (JITAIs) in mobile health: key components and design principles for ongoing health behavior support. Ann Behav Med. May 18, 2018;52(6):446-462. [FREE Full text] [CrossRef] [Medline]
  33. Akane S, Paul J, Mary C. Designing opportune stress intervention delivery timing using multi-modal data. In: Proceedings of the 7th International Conference on Affective Computing and Intelligent Interaction. 2017. Presented at: ACII '17; October 23-26, 2017:346-353; San Antonio, TX. URL: https://ieeexplore.ieee.org/document/8273623 [CrossRef]
  34. Donna SN. Dynamic models of behavior for just-in-time adaptive interventions. IEEE Pervasive Comput. 2014;13(3):17.
  35. Juarascio AS, Parker MN, Lagacey MA, Godfrey KM. Just-in-time adaptive interventions: a novel approach for enhancing skill utilization and acquisition in cognitive behavioral therapy for eating disorders. Int J Eat Disord. Aug 26, 2018;51(8):826-830. [FREE Full text] [CrossRef] [Medline]
  36. Wang L, Miller LC. Just-in-the-moment adaptive interventions (JITAI): a meta-analytical review. Health Commun. Nov 05, 2020;35(12):1531-1544. [CrossRef] [Medline]
  37. Liao P, Dempsey W, Sarker H, Hossain SM, Al'absi M, Klasnja P, et al. Just-in-time but not too much: determining treatment timing in mobile health. Proc ACM Interact Mob Wearable Ubiquitous Technol. Dec 27, 2018;2(4):1-21. [CrossRef] [Medline]
  38. Liao P, Greenewald K, Klasnja P, Murphy S. Personalized HeartSteps: a reinforcement learning algorithm for optimizing physical activity. Proc ACM Interact Mob Wearable Ubiquitous Technol. Mar 18, 2020;4(1):1-22. [FREE Full text] [CrossRef] [Medline]
  39. Mishra V, Künzler F, Kramer J, Fleisch E, Kowatsch T, Kotz D. Detecting receptivity for mHealth interventions in the natural environment. Proc ACM Interact Mob Wearable Ubiquitous Technol. Jun 24, 2021;5(2):1-24. [FREE Full text] [CrossRef] [Medline]
  40. Künzler F, Mishra V, Kramer J, Kotz D, Fleisch E, Kowatsch T. Exploring the state-of-receptivity for mHealth interventions. Proc ACM Interact Mob Wearable Ubiquitous Technol. Dec 14, 2019;3(4):1-27. [FREE Full text] [CrossRef] [Medline]
  41. Gavin D, David C, John S. Engagement with online mental health interventions: an exploratory clinical study of a treatment for depression. In: Proceedings of the 2012 SIGCHI Conference on Human Factors in Computing Systems. 2012. Presented at: CHI '12; May 5-10, 2012:1421-1430; Austin, TX. URL: https://dl.acm.org/doi/10.1145/2207676.2208602 [CrossRef]
  42. Borghouts J, Eikey E, Mark G, De Leon C, Schueller SM, Schneider M, et al. Barriers to and facilitators of user engagement with digital mental health interventions: systematic review. J Med Internet Res. Mar 24, 2021;23(3):e24387. [FREE Full text] [CrossRef] [Medline]
  43. Hardeman W, Houghton J, Lane K, Jones A, Naughton F. A systematic review of just-in-time adaptive interventions (JITAIs) to promote physical activity. Int J Behav Nutr Phys Act. Apr 03, 2019;16(1):31. [FREE Full text] [CrossRef] [Medline]
  44. Bidargaddi N, Almirall D, Murphy S, Nahum-Shani I, Kovalcik M, Pituch T, et al. To prompt or not to prompt? A microrandomized trial of time-varying push notifications to increase proximal engagement with a mobile health app. JMIR Mhealth Uhealth. Nov 29, 2018;6(11):e10123. [FREE Full text] [CrossRef] [Medline]
  45. Yardley L, Spring BJ, Riper H, Morrison LG, Crane DH, Curtis K, et al. Understanding and promoting effective engagement with digital behavior change interventions. Am J Prev Med. Nov 2016;51(5):833-842. [CrossRef] [Medline]
  46. Perrin Franck C, Babington-Ashaye A, Dietrich D, Bediang G, Veltsos P, Gupta PP, et al. iCHECK-DH: guidelines and checklist for the reporting on digital health implementations. J Med Internet Res. May 10, 2023;25:e46694. [FREE Full text] [CrossRef] [Medline]
  47. Perrin Franck C, Babington-Ashaye A, Dietrich D, Bediang G, Veltsos P, Gupta PP, et al. iCHECK-DH: guidelines and checklist for the reporting on digital health implementations. J Med Internet Res. May 10, 2023;25:e46694. [FREE Full text] [CrossRef] [Medline]
  48. Susanne H, Sayantani M, Ariane B, Gillian C. Women in the tech industry: gaining ground, but facing new headwinds. Deloitte Insights. Dec 1, 2021. URL: https:/​/www2.​deloitte.com/​us/​en/​insights/​industry/​technology/​technology-media-and-telecom-predictions/​2022/​statistics-show-women-in-technology-are-facing-new-headwinds.​html [accessed 2024-02-18]
  49. Lovibond P, Lovibond S. The structure of negative emotional states: comparison of the Depression Anxiety Stress Scales (DASS) with the Beck Depression and Anxiety Inventories. Behav Res Ther. Mar 1995;33(3):335-343. [CrossRef] [Medline]
  50. Rammstedt B, John OP. Measuring personality in one minute or less: a 10-item short version of the Big Five Inventory in English and German. J Res Pers. Feb 2007;41(1):203-212. [CrossRef]
  51. Gross JJ, John OP. Individual differences in two emotion regulation processes: implications for affect, relationships, and well-being. J Pers Soc Psychol. Aug 2003;85(2):348-362. [CrossRef] [Medline]
  52. Smith BW, Dalen J, Wiggins K, Tooley E, Christopher P, Bernard J. The brief resilience scale: assessing the ability to bounce back. Int J Behav Med. Sep 2008;15(3):194-200. [CrossRef] [Medline]
  53. Gloria M, Shamsi TI, Mary C, Paul J, Akane S, Yuliya L. Email duration, batching and self-interruption: patterns of email use on productivity and stress. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. 2016. Presented at: CHI '16; May 7-12, 2016:1717-1728; San Jose, CA. URL: https://dl.acm.org/doi/10.1145/2858036.2858262 [CrossRef]
  54. Gloria M, Shamsi I, Mary C. How blocking distractions affects workplace focus and productivity. In: Proceedings of the 2017 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2017 ACM International Symposium on Wearable Computers. 2017. Presented at: UbiComp '17; September 11-15, 2017:928-934; Maui, HI. URL: https://dl.acm.org/doi/abs/10.1145/3123024.3124558 [CrossRef]
  55. McDuff D, Jun E, Rowan K, Czerwinski M. Longitudinal observational evidence of the impact of emotion regulation strategies on affective expression. IEEE Trans Affective Comput. Jul 1, 2021;12(3):636-647. [CrossRef]
  56. Rosenberg EL, Ekman P. What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS) (Series in Affective Science). Oxford, UK. Oxford University Press; 1998.
  57. Mayo LM, Heilig M. In the face of stress: interpreting individual differences in stress-induced facial expressions. Neurobiol Stress. Feb 2019;10:100166. [FREE Full text] [CrossRef] [Medline]
  58. Karim Sadik K. Assessment of emotional experience through facial expression. ProQuest Dissertation & Theses, Harvard University. 2010. URL: https://www.proquest.com/openview/bb4208bb09f6a82b0247e66d69cba33a/ [accessed 2024-06-18]
  59. Vrijkotte TG, van Doornen LJ, de Geus EJ. Effects of work stress on ambulatory blood pressure, heart rate, and heart rate variability. Hypertension. Apr 2000;35(4):880-886. [FREE Full text] [CrossRef] [Medline]
  60. Hjortskov N, Rissén D, Blangsted AK, Fallentin N, Lundberg U, Søgaard K. The effect of mental stress on heart rate variability and blood pressure during computer work. Eur J Appl Physiol. Jun 1, 2004;92(1-2):84-89. [CrossRef] [Medline]
  61. Butler AC, Chapman JE, Forman EM, Beck AT. The empirical status of cognitive-behavioral therapy: a review of meta-analyses. Clin Psychol Rev. Jan 2006;26(1):17-31. [CrossRef] [Medline]
  62. Anita L, Linehan MM. Dialectical behavior therapy: a comprehensive multi- and transdiagnostic intervention. In: Nezu CM, Nezu AM, editors. The Oxford Handbook of Cognitive and Behavioral Therapies. Oxford, UK. Oxford University Press; 2016:200-214.
  63. Linehan MM. DBT Skills Training Manual. 2nd edition. New York, NY. Guilford Publications; 2014.
  64. Overview of Microsoft graph. Microsoft Learn. 2023. URL: https://learn.microsoft.com/en-us/graph/overview [accessed 2024-05-25]
  65. Advanced insights metrics description. Microsoft Learn. 2023. URL: https://learn.microsoft.com/en-us/previous-versions/viva/insights/use/metric-definitions [accessed 2024-05-25]
  66. Liu X, Fromm J, Patel S, McDuff D. Multi-task temporal shift attention networks for on-device contactless vitals measurement. arXiv. Preprint posted online February 28, 2021. [FREE Full text]
  67. Microsoft bot framework. Microsoft. 2024. URL: https://dev.botframework.com [accessed 2024-05-25]
  68. Adaptive cards. Microsoft. 2024. URL: https://adaptivecards.io [accessed 2024-05-25]
  69. Smyth JM, Heron KE. Is providing mobile interventions "just-in-time" helpful? An experimental proof of concept study of just-in-time intervention for stress management. In: Proceedings of the 2016 IEEE Wireless Health. 2016. Presented at: WH '16; October 25-28, 2016:1-7; Bethesda, MD. URL: https://ieeexplore.ieee.org/document/7764561 [CrossRef]
  70. Klasnja P, Harrison BL, LeGrand L, LaMarca A, Froehlich JE, Hudson S. Using wearable sensors and real time inference to understand human recall of routine activities. In: Proceedings of the 10th international conference on Ubiquitous computing. 2008. Presented at: UbiComp '08; September 21-24, 2008:154-163; Seoul, Republic of Korea. URL: https://dl.acm.org/doi/10.1145/1409635.1409656 [CrossRef]
  71. Horvitz E. Principles of mixed-initiative user interfaces. In: Proceedings of 1999 the SIGCHI conference on Human Factors in Computing Systems. 1999. Presented at: CHI '99; May 15-20, 1999:159-166; Pittsburgh, PA. URL: https://dl.acm.org/doi/10.1145/302979.303030 [CrossRef]
  72. Schroeder J, Suh J, Wilks C, Czerwinski M, Munson SA, Fogarty J, et al. Data-driven implications for translating evidence-based psychotherapies into technology-delivered interventions. Int Conf Pervasive Comput Technol Healthc. May 2020;2020:274-287. [FREE Full text] [CrossRef] [Medline]
  73. Schroeder J, Suh J, Wilks C, Czerwinski M, Munson SA, Fogarty J, et al. Data-driven implications for translating evidence-based psychotherapies into technology-delivered interventions. Int Conf Pervasive Comput Technol Healthc. May 2020;2020:274-287. [FREE Full text] [CrossRef] [Medline]
  74. Choi W, Park S, Kim D, Lim Y, Lee U. Multi-stage receptivity model for mobile just-in-time health intervention. Proc ACM Interact Mob Wearable Ubiquitous Technol. Jun 21, 2019;3(2):1-26. [CrossRef]
  75. McCrae RR, Costa Jr PT. A contemplated revision of the NEO five-factor inventory. Pers Individ Dif. Feb 2004;36(3):587-596. [CrossRef]
  76. McCrae RR, John OP. An introduction to the five-factor model and its applications. J Pers. Jun 28, 1992;60(2):175-215. [CrossRef] [Medline]
  77. John OP, Srivastava S. The big five trait taxonomy: history, measurement, and theoretical perspectives. In: Pervin LA, John OP, editors. Handbook of Personality: Theory and Research. 2nd edition. New York, NY. Guilford Press; 1999:102-138.
  78. Connor-Smith JK, Flachsbart C. Relations between personality and coping: a meta-analysis. J Pers Soc Psychol. Dec 2007;93(6):1080-1107. [CrossRef] [Medline]
  79. Bolger N, Zuckerman A. A framework for studying personality in the stress process. J Pers Soc Psychol. Nov 1995;69(5):890-902. [CrossRef] [Medline]
  80. Langsrud Ø. ANOVA for unbalanced data: Use type ii instead of type iii sums of squares. Stat Comput. 2003;13(2):167. [FREE Full text] [CrossRef]
  81. Ho FK, Celis-Morales CA, Gray SR, Katikireddi SV, Niedzwiedz CL, Hastie C, et al. Modifiable and non-modifiable risk factors for COVID-19, and comparison to risk factors for influenza and pneumonia: results from a UK Biobank prospective cohort study. BMJ Open. Nov 19, 2020;10(11):e040402. [FREE Full text] [CrossRef] [Medline]
  82. Roberts BW, Luo J, Briley DA, Chow PI, Su R, Hill PL. A systematic review of personality trait change through intervention. Psychol Bull. Feb 2017;143(2):117-141. [CrossRef] [Medline]
  83. March S, Day J, Ritchie G, Rowe A, Gough J, Hall T, et al. Attitudes toward e-Mental health services in a community sample of adults: online survey. J Med Internet Res. Feb 19, 2018;20(2):e59. [FREE Full text] [CrossRef] [Medline]
  84. Mikolasek M, Witt CM, Barth J. Adherence to a mindfulness and relaxation self-care app for cancer patients: mixed-methods feasibility study. JMIR Mhealth Uhealth. Dec 06, 2018;6(12):e11271. [FREE Full text] [CrossRef] [Medline]
  85. McCrae RR, Costa PT. Conceptions and correlates of openness to experience. In: Hogan R, Johnson JA, Briggs SR, editors. Handbook of Personality Psychology. Cambridge, MA. Academic Press; 1997:825-847.
  86. Ervasti M, Kallio J, Määttänen I, Mäntyjärvi J, Jokela M. Influence of personality and differences in stress processing among Finnish students on interest to use a mobile stress management app: survey study. JMIR Ment Health. May 13, 2019;6(5):e10039. [FREE Full text] [CrossRef] [Medline]
  87. Bucher MA, Suzuki T, Samuel DB. A meta-analytic review of personality traits and their associations with mental health treatment outcomes. Clin Psychol Rev. Jun 2019;70:51-63. [CrossRef] [Medline]
  88. Graziano WG, Eisenberg N. Agreeableness: a dimension of personality. In: Hogan R, Johnson JA, Briggs SR, editors. Handbook of Personality Psychology. Cambridge, MA. Academic Press; 1997:795-824.
  89. Pereira A, Cabacos C, Soares S, Araújo A, Manão A, Amaral A, et al. Relationship between personality, adherence to (mental) health behaviours and psychological distress during the COVID-19 pandemic. Eur Psychiatr. Sep 01, 2022;65(S1):S499-S500. [CrossRef]
  90. Myruski S, Denefrio S, Dennis-Tiwary TA. Stress and emotion regulation: the dynamic fit model. In: Harkness KL, Hayden EP, editors. The Oxford Handbook of Stress and Mental Health. Oxford, UK. Oxford University Press; 2019:415-434.
  91. Wang M, Saudino KJ. Emotion regulation and stress. J Adult Dev. Jan 11, 2011;18(2):95-103. [CrossRef]
  92. Troy AS, Mauss IB. Resilience in the face of stress: emotion regulation as a protective factor. In: Southwick SM, Litz BT, Charney D, Friedman MJ, editors. Resilience and Mental Health: Challenges Across the Lifespan. Cambridge, MA. Cambridge University Press; 2011:30-44.
  93. Berking M, Wupperman P, Reichardt A, Pejic T, Dippel A, Znoj H. Emotion-regulation skills as a treatment target in psychotherapy. Behav Res Ther. Nov 2008;46(11):1230-1237. [CrossRef] [Medline]
  94. Grecucci A, Sanfey AG. Emotion regulation and decision making. In: Gross JJ, editor. Handbook of Emotion Regulation. 2nd edition. New York, NY. The Guilford Press; 2014:140-153.
  95. Glinski K, Page AC. Modifiability of neuroticism, extraversion, and agreeableness by group cognitive behaviour therapy for social anxiety disorder. Behav Change. Feb 22, 2012;27(1):42-52. [CrossRef]
  96. Boyle CC, Stanton AL, Ganz PA, Crespi CM, Bower JE. Improvements in emotion regulation following mindfulness meditation: effects on depressive symptoms and perceived stress in younger breast cancer survivors. J Consult Clin Psychol. Apr 2017;85(4):397-402. [FREE Full text] [CrossRef] [Medline]
  97. Deci EL, Ryan RM. Intrinsic Motivation and Self-Determination in Human Behavior. Berlin, Germany. Springer Science & Business Media; 2013.
  98. Ajzen I. The theory of planned behavior. Organ Behav Hum Decis Process. Dec 1991;50(2):179-211. [CrossRef]
  99. Kostadin K, Jason P, Elizabeth WD. "Silence your phones": smartphone notifications increase inattention and hyperactivity symptoms. In: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems. 2016. Presented at: CHI '16; May 7-12, 2016:1011-1020; San Jose, CA. URL: https://dl.acm.org/doi/10.1145/2858036.2858359 [CrossRef]
  100. Klasnja P, Smith S, Seewald NJ, Lee A, Hall K, Luers B, et al. Efficacy of contextually tailored suggestions for physical activity: a micro-randomized optimization trial of HeartSteps. Ann Behav Med. May 03, 2019;53(6):573-582. [FREE Full text] [CrossRef] [Medline]
  101. Liao P, Klasnja P, Tewari A, Murphy SA. Sample size calculations for micro-randomized trials in mHealth. Stat Med. May 30, 2016;35(12):1944-1971. [FREE Full text] [CrossRef] [Medline]
  102. Olson-Buchanan JB, Boswell WR. Blurring boundaries: correlates of integration and segmentation between work and nonwork. J Vocat Behav. Jun 2006;68(3):432-445. [CrossRef]
  103. Allen TD, Merlo K, Lawrence RC, Slutsky J, Gray CE. Boundary management and work‐nonwork balance while working from home. Applied Psychology. Jan 20, 2021;70(1):60-84. [CrossRef]
  104. Jina S, Javier H, Koustuv S, Kathy D, Mehrab BM, Esther H, et al. Towards successful deployment of wellbeing sensing technologies: identifying misalignments across contextual boundaries. In: Proceedings of the 11th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos. 2023. Presented at: ACIIW '23; September 10-13, 2023:1-8; Cambridge, MA. URL: https://ieeexplore.ieee.org/document/10388088 [CrossRef]


DASS-21: Depression, Anxiety, and Stress Scale-21
DMHI: digital mental health intervention
EMA: ecological momentary assessment
iCHECK-DH: Guidelines and Checklist for the Reporting on Digital Health Implementations
JIT: just-in-time
JITAI: just-in-time adaptive intervention
OR: odds ratio


Edited by I Lefter; submitted 17.05.23; peer-reviewed by M Rabbi, H Wen; comments to author 15.01.24; revised version received 05.06.24; accepted 17.07.24; published 12.09.24.

Copyright

©Jina Suh, Esther Howe, Robert Lewis, Javier Hernandez, Koustuv Saha, Tim Althoff, Mary Czerwinski. Originally published in JMIR Mental Health (https://mental.jmir.org), 12.09.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on https://mental.jmir.org/, as well as this copyright and license information must be included.