<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Ment Health</journal-id><journal-id journal-id-type="publisher-id">mental</journal-id><journal-id journal-id-type="index">16</journal-id><journal-title>JMIR Mental Health</journal-title><abbrev-journal-title>JMIR Ment Health</abbrev-journal-title><issn pub-type="epub">2368-7959</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v11i1e60003</article-id><article-id pub-id-type="doi">10.2196/60003</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Balancing Between Privacy and Utility for Affect Recognition Using Multitask Learning in Differential Privacy&#x2013;Added Federated Learning Settings: Quantitative Study</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Benouis</surname><given-names>Mohamed</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author"><name name-style="western"><surname>Andre</surname><given-names>Elisabeth</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Can</surname><given-names>Yekta Said</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff1"/></contrib></contrib-group><aff id="aff1"><institution>Faculty of Applied Computer Science, Augsburg University</institution><addr-line>Augsburg</addr-line><country>Germany</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Torous</surname><given-names>John</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Jaafar</surname><given-names>Eman</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Washington</surname><given-names>Peter</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Gupta</surname><given-names>Puneet</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Yekta Said Can, PhD, Faculty of Applied Computer Science, Augsburg University, Universit&#x00E4;tsstra&#x00DF;e 6a, Augsburg, 86159, Germany, 49 015221351388; <email>yekta.can@uni-a.de</email></corresp></author-notes><pub-date pub-type="collection"><year>2024</year></pub-date><pub-date pub-type="epub"><day>23</day><month>12</month><year>2024</year></pub-date><volume>11</volume><elocation-id>e60003</elocation-id><history><date date-type="received"><day>28</day><month>04</month><year>2024</year></date><date date-type="rev-recd"><day>21</day><month>08</month><year>2024</year></date><date date-type="accepted"><day>22</day><month>08</month><year>2024</year></date></history><copyright-statement>&#x00A9; Mohamed Benouis, Elisabeth Andre, Yekta Said Can. Originally published in JMIR Mental Health (<ext-link ext-link-type="uri" xlink:href="https://mental.jmir.org">https://mental.jmir.org</ext-link>), 23.12.2024. </copyright-statement><copyright-year>2024</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Mental Health, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://mental.jmir.org/">https://mental.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://mental.jmir.org/2024/1/e60003"/><abstract><sec><title>Background</title><p>The rise of wearable sensors marks a significant development in the era of affective computing. Their popularity is continuously increasing, and they have the potential to improve our understanding of human stress. A fundamental aspect within this domain is the ability to recognize perceived stress through these unobtrusive devices.</p></sec><sec><title>Objective</title><p>This study aims to enhance the performance of emotion recognition using multitask learning (MTL), a technique extensively explored across various machine learning tasks, including affective computing. By leveraging the shared information among related tasks, we seek to augment the accuracy of emotion recognition while confronting the privacy threats inherent in the physiological data captured by these sensors.</p></sec><sec sec-type="methods"><title>Methods</title><p>To address the privacy concerns associated with the sensitive data collected by wearable sensors, we proposed a novel framework that integrates differential privacy and federated learning approaches with MTL. This framework was designed to efficiently identify mental stress while preserving private identity information. Through this approach, we aimed to enhance the performance of emotion recognition tasks while preserving user privacy.</p></sec><sec sec-type="results"><title>Results</title><p>Comprehensive evaluations of our framework were conducted using 2 prominent public datasets. The results demonstrate a significant improvement in emotion recognition accuracy, achieving a rate of 90%. Furthermore, our approach effectively mitigates privacy risks, as evidenced by limiting reidentification accuracies to 47%.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>This study presents a promising approach to advancing emotion recognition capabilities while addressing privacy concerns in the context of empathetic sensors. By integrating MTL with differential privacy and federated learning, we have demonstrated the potential to achieve high levels of accuracy in emotion recognition while ensuring the protection of user privacy. This research contributes to the ongoing efforts to use affective computing in a privacy-aware and ethical manner.</p></sec></abstract><kwd-group><kwd>privacy preservation</kwd><kwd>multitask learning</kwd><kwd>federated learning</kwd><kwd>privacy</kwd><kwd>physiological signals</kwd><kwd>affective computing</kwd><kwd>wearable sensors</kwd><kwd>sensitive data</kwd><kwd>empathetic sensors</kwd><kwd>data privacy</kwd><kwd>digital mental health</kwd><kwd>wearables</kwd><kwd>ethics</kwd><kwd>emotional well-being</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><sec id="s1-1"><title>Background</title><p>Imagine awakening in the morning to find yourself dealing with stress, perhaps stemming from a disagreement with a friend, a pressing financial concern, or an unresolved work&#x2013;related matter. In this scenario, an advanced digital assistant, akin to Siri or Alexa, seamlessly detects the user&#x2019;s stress level through the analysis of physiological signals captured by wearable technology. Leveraging this data, the digital assistant dynamically adjusts its language and tone to suit the user&#x2019;s mood. Moreover, it proactively recommends personalized relaxation techniques, drawing from past successful interventions such as yoga, mindfulness practices, listening to favorite music, and watching uplifting videos. Such a sophisticated stress recognition and intervention system holds promise for enhancing the well-being and emotional resilience of individuals. Although most of the studies use facial expressions [<xref ref-type="bibr" rid="ref1">1</xref>] and speech [<xref ref-type="bibr" rid="ref2">2</xref>] for recognizing stress and emotions (ie, affects as an umbrella term), physiology-based methodologies also emerged as an alternative since they can offer promise for seamless, continuous monitoring within everyday contexts. Notably, 330 million smartwatches, fitness trackers, and wearables were sold in 2020, which are capable of gathering quantitative physiological data [<xref ref-type="bibr" rid="ref3">3</xref>]. They are promising tools for affect recognition due to their unobtrusive data collection capabilities.</p><p>Multitask learning (MTL) is a method proposed to tackle training multiple related tasks at the same time. It works by sharing knowledge between these tasks to make each model perform better [<xref ref-type="bibr" rid="ref4">4</xref>]. Essentially, MTL acts like a behind-the-scenes helper, enhancing the ability of machine learning (ML) models to generalize different types of data [<xref ref-type="bibr" rid="ref5">5</xref>]. This technique has been particularly useful in improving ML models across various fields, including affective computing [<xref ref-type="bibr" rid="ref6">6</xref>]. For instance, if we want to improve how a system recognizes emotions, we can also train it to recognize gender [<xref ref-type="bibr" rid="ref7">7</xref>]. By doing this, the system can learn from both tasks simultaneously, making it more effective in recognizing emotions and enhancing performance. The motivation for using identity and emotion has also a psychological basis. Connolly et al [<xref ref-type="bibr" rid="ref8">8</xref>] examined whether facial responses are shared between identity and emotion. They showed that there is a strong positive correlation between face emotion recognition and face identity recognition tasks, and the 2 recognition tasks share a common processing mechanism. Sheng and Li [<xref ref-type="bibr" rid="ref9">9</xref>] also confirmed the dependence of identity and emotion tasks by using a gait signal. Inspired by the MTL idea and previous studies, we used MTL to perform each task separately, stress recognition and identity recognition, respectively, by using physiological signals. However, there is a concern when third parties gain access to additional information, like gender, identity, or age, which could be misused for targeted advertising or even discrimination in things like job opportunities [<xref ref-type="bibr" rid="ref10">10</xref>]. Moreover, if sensitive information like biometric data gets leaked, it could lead to cybercrimes such as identity theft. Therefore, it is crucial for MTL systems to incorporate privacy-preserving methods.</p><p>Researchers have introduced federated learning (FL) as a way to address privacy issues. FL allows users to train their data locally while sharing only the trained model parameters rather than the original data itself. The philosophy behind FL is to bring code to the data instead of moving the data to the code [<xref ref-type="bibr" rid="ref11">11</xref>]. This method complies with privacy regulations such as the European Union&#x2019;s General Data Protection Regulation [<xref ref-type="bibr" rid="ref12">12</xref>] and the California Consumer Privacy Act [<xref ref-type="bibr" rid="ref13">13</xref>]. FL has shown promise in safeguarding user privacy in Internet of Things networks.</p><p>Although FL is a significant step forward in protecting user privacy, it is not invincible. There is still a risk of sensitive information being uncovered by reverse engineering the local model parameters. To further enhance privacy, researchers have integrated privacy-preserving techniques into ML for physiological data. One such method is differential privacy (DP), which adds random noise to each model in the client or server, disrupting updates and limiting the leakage of sensitive information between nodes [<xref ref-type="bibr" rid="ref14">14</xref>]. However, it is worth noting that DP can reduce the performance of ML models.</p><p>In practice, protecting the privacy of users without degradation in model utility is still an open problem. In this study, we first implemented an MTL architecture for recognizing stress and identity from multimodal physiological signals. We further added FL and DP mechanisms to preserve privacy. To obtain a robust performance, we separated the 2 tasks and added noise to only the privacy task, which is biometric identity recognition. In this way, we were able to improve the stress recognition performance with the help of MTL and hide identity information by adding noise to the identity task model by using DP. To the best of our knowledge, this study is the first MTL-based affect recognition study using FL and DP to preserve privacy at the same time.</p><p>In the next section, we mentioned the related works using MTL and FL for affect recognition. We then presented our approach in the <italic>Methods</italic> section. In the <italic>Results</italic> section, we compared multitask centralized, decentralized FL, and decentralized FL with DP approaches on selected public datasets for recognizing stress levels and identity of users. We concluded the study with the lessons learned, limitations, and future works in the <italic>Discussion</italic> section.</p></sec><sec id="s1-2"><title>Related Works</title><p>To develop a robust affective computing system that can be used in practical applications, researchers tested various modalities with state-of-the-art deep learning techniques. MTL has also raised significant attention from various domains over the past few years, including affective computing. Chen et al [<xref ref-type="bibr" rid="ref14">14</xref>] applied it to audiovisual signals in the Audio/Visual Emotion Challenge and achieved a significant performance increase compared to baseline ML algorithms by detecting arousal and valence levels simultaneously. Since the human face source can be used for several tasks, such as gender recognition or age estimation, Sang et al [<xref ref-type="bibr" rid="ref7">7</xref>] used MTL with convolutional neural network (CNN) for smile detection, emotion recognition, and gender classification.</p><p>They outperformed the state-of-the-art techniques in selected 3 benchmark datasets (Internet Movie Database and Wiki dataset, GENKI-4K dataset [a dataset containing images labeled with facial expressions, commonly used in emotion recognition research], and Facial Expression Recognition Challenge-2013 dataset). MTL-based centralized learning (CL) for affective computing is beneficial for simultaneously sharing features and learning auxiliary tasks. However, when private tasks (ie, face, gender, and person detection) are included in MTL to improve affect recognition performances, the models create the risk to reveal this sensitive information to possibly malicious parties especially while uploading the features to the central server for both utility and auxiliary tasks for learning.</p><p>Data privacy has become an issue of great concern in affect recognition using either verbal or nonverbal data, as gender, age, and identity of the user can be revealed in the process. FL is proposed to preserve privacy while taking advantage of ML and has attracted significant attention from various domains over the past few years, but affective computing research and applications on emotion-related tasks are rarely discussed. Most existing works are conducted on private datasets or in limited scenarios, making it difficult for researchers to compare their methods fairly [<xref ref-type="bibr" rid="ref15">15</xref>].</p><p>FL has been widely applied for facial features for affect recognition. Somandepalli et al [<xref ref-type="bibr" rid="ref16">16</xref>] investigated FL for 2 affective computing tasks: classifying self-reports and perception ratings in audio, video, and text datasets. Using the speech modality, Latif et al [<xref ref-type="bibr" rid="ref17">17</xref>] investigated an FL approach in emotion recognition tasks while sharing only the model among clients. They implemented a long short-term memory classifier and tested it on the interactive emotional dyadic motion capture dataset with 4 emotions: happy, sad, angry, and neutral. Face and speech modalities were also often combined to get a more robust performance [<xref ref-type="bibr" rid="ref18">18</xref>]. Chhikara et al [<xref ref-type="bibr" rid="ref19">19</xref>] combined emotion recognition from facial expressions and speech with an FL approach. For the face modality, they used a combination of CNN and support vector machine models, whereas for the audio modality, they applied a 2D CNN model to extract spectrogram images. Their proposed framework has been validated and tested on 2 datasets, Facial Expression Recognition 2013 for facial emotion recognition and Ryerson Audio-Visual Database of Emotional Speech and Song for speech emotion recognition, respectively.</p><p>On the contrary, FL has been seldom used in the context of affect recognition from physiological signals, as indicated in <xref ref-type="table" rid="table1">Table 1</xref>. Can and Ersoy [<xref ref-type="bibr" rid="ref15">15</xref>] implemented an FL learning model to forecast perceived stress using physiological data. Each subclient uses an multilayer perceptron classifier to locally train its data on the edge, and the sharing of individual updating parameters of multilayer perceptron is facilitated using the federated averaging (FedAvg) algorithm. FL has also been extended to handle multimodal physiological signals. Nandi and Xhafa [<xref ref-type="bibr" rid="ref20">20</xref>] proposed Federated Representation Learning for Multi-Modal Emotion Classification System, an FL framework-based ML model for emotion recognition. They applied wavelet feature extraction and a neural network to electrodermal activity (EDA) and respiration data from the Dataset for Emotion Analysis using Physiological signals to recognize valence arousal levels, validating their approach. These studies demonstrate the successful application of FL without compromising affect recognition performance.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Studies using MTL<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup> or privacy-preserving approaches for various applications.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Study</td><td align="left" valign="bottom">Application</td><td align="left" valign="bottom">Signal</td><td align="left" valign="bottom">MTL</td><td align="left" valign="bottom">FL<sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></td><td align="left" valign="bottom">DP<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup></td><td align="left" valign="bottom">Multimodality</td></tr></thead><tbody><tr><td align="left" valign="top">Zhao et al [<xref ref-type="bibr" rid="ref21">21</xref>]</td><td align="left" valign="top">Network anomaly detection</td><td align="left" valign="top">Network signals</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Smith et al [<xref ref-type="bibr" rid="ref22">22</xref>]</td><td align="left" valign="top">Activity recognition</td><td align="left" valign="top">Acceleration, gyroscope</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td></tr><tr><td align="left" valign="top">Somandepalli et al [<xref ref-type="bibr" rid="ref16">16</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">Speech, video, text</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td></tr><tr><td align="left" valign="top">Sang et al [<xref ref-type="bibr" rid="ref7">7</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">Face images</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Shome and Kar [<xref ref-type="bibr" rid="ref23">23</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">Face images</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Chhikara et al [<xref ref-type="bibr" rid="ref19">19</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">Face and speech</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td></tr><tr><td align="left" valign="top">Feng et al [<xref ref-type="bibr" rid="ref24">24</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">Speech</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Can and Ersoy [<xref ref-type="bibr" rid="ref15">15</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">PPG<sup><xref ref-type="table-fn" rid="table1fn4">d</xref></sup></td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top"/></tr><tr><td align="left" valign="top">Nandi and Xhafa [<xref ref-type="bibr" rid="ref20">20</xref>]</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">EEG<sup><xref ref-type="table-fn" rid="table1fn5">e</xref></sup>, EDA<sup><xref ref-type="table-fn" rid="table1fn6">f</xref></sup></td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td></tr><tr><td align="left" valign="top">Wang et al [<xref ref-type="bibr" rid="ref18">18</xref>]</td><td align="left" valign="top">Biometric</td><td align="left" valign="top">Face<break/>speaker</td><td align="left" valign="top"/><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td></tr><tr><td align="left" valign="top">This study</td><td align="left" valign="top">Affect recognition</td><td align="left" valign="top">PPG, EDA, ACC<sup><xref ref-type="table-fn" rid="table1fn7">g</xref></sup>, ST<sup><xref ref-type="table-fn" rid="table1fn8">h</xref></sup></td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td><td align="left" valign="top">&#x2713;</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>MTL: multitask learning.</p></fn><fn id="table1fn2"><p><sup>b</sup>FL: federated learning.</p></fn><fn id="table1fn3"><p><sup>c</sup>DP: differential privacy.</p></fn><fn id="table1fn4"><p><sup>d</sup>PPG: photoplethysmography.</p></fn><fn id="table1fn5"><p><sup>e</sup>EEG: electroencephalography.</p></fn><fn id="table1fn6"><p><sup>f</sup>EDA: electrodermal activity.</p></fn><fn id="table1fn7"><p><sup>g</sup>ACC: acceleration.</p></fn><fn id="table1fn8"><p><sup>h</sup>ST: skin temperature.</p></fn></table-wrap-foot></table-wrap><p>While FL has been introduced to enhance model training in terms of privacy, the privacy vulnerabilities inherent in the stochastic gradient descent (SGD) algorithm remain unresolved. DP mechanisms have primarily been discussed in FL settings, involving the injection of noise into each model client or server to perturb updates and limit gradient leakage shared among nodes (ie, client and server) [<xref ref-type="bibr" rid="ref25">25</xref>]. In one of the initial applications, authors introduced a new private training method termed differential private SGD, which reduces local and global gradient information leakage between the client and server. Instead of using the standard composition theorem to calculate the final distribution of overall noise clients, they used a moments accountant metric to adaptively monitor the overall privacy loss. Recognizing that servers are often curious or untrustworthy, Wei et al [<xref ref-type="bibr" rid="ref26">26</xref>] proposed a local DP mechanism algorithm by introducing Gaussian noise distribution into user models before uploading them to servers. To address the communication overhead required for optimal convergence upper bound for DP, they introduced a novel approach known as communication rounds discounting method, which achieves a better trade-off between the computational complexity of searching and convergence performance. DP has also been leveraged for affect recognition. Feng et al [<xref ref-type="bibr" rid="ref24">24</xref>] used user-level DP to mitigate privacy leaks in FL for speech emotion recognition. Recently, Smith et al [<xref ref-type="bibr" rid="ref22">22</xref>] showcased promising performance in preserving privacy via multitask federated learning (MFL) for activity recognition.</p><p>When examining the affective computing literature, we find studies using MTL to improve performance and others using FL or DP to preserve privacy. However, there are no studies that leverage MTL for performance enhancement while also preserving privacy in this literature. The overall idea of our approach is to separate the utility and the privacy tasks preventing DP from compromising the performance of the utility task (affect recognition) and introducing noise exclusively to the privacy task (biometric identity recognition). Consequently, our work represents the first MTL-based approach to affect recognition using FL and DP to preserve privacy at the same time.</p></sec></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Overview</title><p>Our proposed framework is divided into 3 main substeps: feature extraction, FL model, and FL with DP settings (see <xref ref-type="fig" rid="figure1">Figure 1</xref> for the block diagram).</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Block diagram of the proposed MTL system. CNN: convolutional neural network; MTL: multitask learning.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig01.png"/></fig></sec><sec id="s2-2"><title>Ethical Considerations</title><p>This study did not require formal ethics approval, as it exclusively used publicly available datasets, which were obtained from open-access repositories with no personally identifiable information. According to the General Data Protection Regulation, studies involving the use of anonymized, publicly accessible data that do not engage human participants directly are exempt from ethics review. Therefore, no ethics approval was sought or required.</p></sec><sec id="s2-3"><title>Data Description and Feature Extraction</title><p>For our experiments, we use 2 datasets, wearable stress and affect detection (WESAD) [<xref ref-type="bibr" rid="ref27">27</xref>] and virtual environment for real-time biometric interaction and observation (VERBIO) [<xref ref-type="bibr" rid="ref28">28</xref>], which have been created for affective state monitoring.</p><sec id="s2-3-1"><title>WESAD Dataset</title><p>In the WESAD dataset, each participant recorded physiological signals such as blood volume pulse, electrocardiogram (ECG), EDA, electromyogram, respiration, body temperature, and 3-axis acceleration measured from the chest and wrist using Plux RespiBAN and Empatica E4 devices. In total, 15 (12 male and 3 female) people participated in this experience. There were 4 states: baseline, amusement, stress, and meditation. More details can be found in Schmidt et al [<xref ref-type="bibr" rid="ref27">27</xref>].</p></sec><sec id="s2-3-2"><title>VERBIO Dataset</title><p>The VERBIO dataset [<xref ref-type="bibr" rid="ref28">28</xref>] consists of biobehavioral responses and self-reported data during public speaking presentations, both in front of real-life and virtual audiences. In total, 55 participants conducted 10 distinct presentations, each lasting approximately 5 minutes, across the 3 segments of the study, spanning 4 days: PRE (1 session, day 1), TEST (8 sessions, days 2&#x2010;3), and POST (1 session, day 4). The PRE and POST segments involved real-life audiences, while the TEST segment featured various virtual audiences. To prevent participant exhaustion, the TEST segment was divided into 2 days, each comprising 4 sessions. In total, the study yielded 10,800 minutes of acoustic and physiological data from 82 real and 216 virtual reality presentations. In total, 35,174 segments were created. We solely used physiological data collected using the Empatica E4 device, possessing identical sampling rates and modalities as those used in the WESAD dataset.</p></sec><sec id="s2-3-3"><title>Preprocessing</title><p>Each modality of the signal is segmented using 700 sample windows size with 50% overlap, as suggested in the literature [<xref ref-type="bibr" rid="ref29">29</xref>]. To maximize the correlation among interparticipants and minimize among participants, these segments were further processed for extracting features [<xref ref-type="bibr" rid="ref15">15</xref>] such as mean, variance, root mean square, frequency domain features, average first amplitude difference, second amplitude difference, skewness, kurtosis, and entropy as a nonlinear feature.</p></sec></sec><sec id="s2-4"><title>Decentralized MFL Model</title><p>For privatizing the user identity while preserving stress recognition accuracy, we adopted an MFL approach that can effectively improve the performance of stress recognition while limiting the risk of inferring sensitive information from the training model since the client does not want to be exposed to the cloud service provider. MFL architecture-based stress recognition is developed as follows:</p><list list-type="order"><list-item><p>The dataset is partitioned into <italic>K</italic> clients. The data size of all the clients is the same. The client distribution is also assumed as independent and identically distributed (IID) and no IID [<xref ref-type="bibr" rid="ref10">10</xref>].<xref ref-type="bibr" rid="ref10">10</xref></p></list-item><list-item><p>For the local training process, there is only 1 iteration for SGD local training for each client. In particular, <bold>w</bold> is the local model parameter [<xref ref-type="bibr" rid="ref10">10</xref>] given by:</p></list-item></list><disp-formula id="E1"><label>(1)</label><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mrow><mml:mi mathvariant="normal">a</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">r</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">g</mml:mi></mml:mrow><mml:munder><mml:mrow><mml:mrow><mml:mi mathvariant="normal">m</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow></mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub></mml:munder><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mfrac><mml:mi>&#x03BC;</mml:mi><mml:mn>2</mml:mn></mml:mfrac><mml:msup><mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msup><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mi>D</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo symmetric="true">&#x2016;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mstyle></mml:math></disp-formula><list list-type="simple"><list-item><p>where <inline-formula><mml:math id="ieqn1"><mml:mi>&#x03BC;</mml:mi></mml:math></inline-formula> is the learning rate and <inline-formula><mml:math id="ieqn2"><mml:msub><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the local loss function of the <italic>k</italic>th user.</p></list-item></list><list list-type="order"><list-item><p>The local data of different clients cannot be communicated, and only the models can be shared.<xref ref-type="bibr" rid="ref10">10</xref><xref ref-type="bibr" rid="ref10">10</xref></p></list-item><list-item><p>Following the FedAvg algorithm [<xref ref-type="bibr" rid="ref10">10</xref>], the server uses a global averaging approach to aggregate all local training models to compute the final global model. Formally [<xref ref-type="bibr" rid="ref10">10</xref>], the server aggregates the weights sent from the <italic>K</italic> clients as follows:</p></list-item></list><disp-formula id="E2"><label>(2)</label><mml:math id="eqn2"><mml:mi mathvariant="bold">w</mml:mi><mml:mo>=</mml:mo><mml:mrow><mml:munderover><mml:mo stretchy="false">&#x2211;</mml:mo><mml:mrow><mml:mi>U</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi mathvariant="normal">&#x200D;</mml:mi></mml:mrow></mml:mrow><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:msubsup><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msubsup></mml:math></disp-formula><list list-type="simple"><list-item><p>where <inline-formula><mml:math id="ieqn3"><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the parameter vector trained at the <italic>k</italic>th client, <inline-formula><mml:math id="ieqn4"><mml:mi>w</mml:mi></mml:math></inline-formula> is the parameter vector after aggregating at the server, <inline-formula><mml:math id="ieqn5"><mml:mi>K</mml:mi></mml:math></inline-formula> is the number of participating clients, <inline-formula><mml:math id="ieqn6"><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the dataset size of each participating client, <inline-formula><mml:math id="ieqn7"><mml:mi>D</mml:mi><mml:mo>=</mml:mo><mml:mi>U</mml:mi><mml:msub><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the whole distributed dataset, and <italic>P<sub>i</sub></italic>=|<italic>D<sub>i</sub></italic>|/|<italic>D</italic>|.</p></list-item></list><list list-type="order"><list-item><p>The global training epoch is set to <inline-formula><mml:math id="ieqn8"><mml:mi>M</mml:mi></mml:math></inline-formula> rounds (aggregations). The server solves the optimization problem [<xref ref-type="bibr" rid="ref10">10</xref>]:</p></list-item></list><disp-formula id="E3"><label>(3)</label><mml:math id="eqn3"><mml:msup><mml:mrow><mml:mi mathvariant="bold">W</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">*</mml:mi></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">r</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:munder><mml:mrow><mml:mi mathvariant="normal">m</mml:mi><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">n</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold">U</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:munder><mml:mrow><mml:munderover><mml:mo stretchy="false">&#x2211;</mml:mo><mml:mrow><mml:mi>U</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>M</mml:mi></mml:mrow></mml:munderover><mml:mrow><mml:mi mathvariant="normal">&#x200D;</mml:mi></mml:mrow></mml:mrow><mml:msub><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold">w</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="bold">U</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="script">D</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></disp-formula><list list-type="simple"><list-item><p>where  <inline-formula><mml:math id="ieqn9"><mml:msub><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mi>U</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> is the local loss function of the <italic>k</italic>th client. Generally, the local loss function is given by local empirical risks.</p></list-item></list></sec><sec id="s2-5"><title>Decentralized FL With DP</title><p>In conventional FL, the global model is computed through averaging over model client participants, which performs better within homogeneous FL settings. However, using inference or adversarial attack, this shared model may contain sensitive and private information such as gender, age, and biometric template user. In such cases, the MFL framework is required to reduce the leakage of the black box gradient exchanged model. To overcome this limitation, researchers have used the DP scheme to protect either local or global data training FL model. However, the perturbed gradient using DP with a low budget has high variance, which leads to worse performance and slower convergence. We also compared adding noise to whole model and task-specific last layers in addition to the shared layers and demonstrated the performance of these 3 different approaches (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). Motivated by personalized FL [<xref ref-type="bibr" rid="ref26">26</xref>], our work focuses on client-level privacy, which aims at a private specific layer of the client model rather than perturbing the entire whole local model. This is because the base layers are mostly redundant information, while the most important information that holds private and public information is located in the upper layers. To meet the utility privacy trade-off, the DP mechanism is used to perturb the gradients using Gaussian noise at a specific layer or task. Here, we use all steps in the FL model except step 4, that is, before uploading the local SGD model client to the global server, we inject an amount of noise to the updated local parameters. In that sense, we perturb the local gradient training inference with two kinds of noise distributions:</p><list list-type="order"><list-item><p>An additive Gaussian noise <italic>&#x03B7;</italic>&#x223C;<italic>N</italic> (0, <italic>&#x03C3;</italic>) to each weight. This operation can be mathematically described as follows:</p></list-item></list><disp-formula id="E4"><label>(4)</label><mml:math id="eqn4"><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:mi>&#x03B7;</mml:mi></mml:math></disp-formula><list list-type="order"><list-item><p>A set of noise distributions can be sampled from the DP mechanism. A randomized mechanism <inline-formula><mml:math id="ieqn10"><mml:mi mathvariant="script">M</mml:mi></mml:math></inline-formula> on the training set with domain <italic>X</italic> and range <italic>R</italic> satisfies (<italic>&#x03F5;</italic>, <italic>&#x03B4;</italic>) &#x2212; DP for 2 small positive numbers and if the following inequality holds [<xref ref-type="bibr" rid="ref25">25</xref>]:</p></list-item></list><disp-formula id="E5"><label>(5)</label><mml:math id="eqn5"><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mo>[</mml:mo><mml:mi mathvariant="script">M</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mi mathvariant="script">S</mml:mi><mml:mo>]</mml:mo><mml:mo>&#x2264;</mml:mo><mml:msup><mml:mrow><mml:mi>e</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x03F5;</mml:mi></mml:mrow></mml:msup><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mfenced open="[" close="]" separators="|"><mml:mrow><mml:mi mathvariant="script">M</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>`</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfenced><mml:mo>&#x2208;</mml:mo><mml:mi mathvariant="script">S</mml:mi></mml:mrow></mml:mfenced><mml:mo>+</mml:mo><mml:mi>&#x03B4;</mml:mi></mml:math></disp-formula><list list-type="simple"><list-item><p>where <inline-formula><mml:math id="ieqn11"><mml:mi>x</mml:mi></mml:math></inline-formula> and <inline-formula><mml:math id="ieqn12"><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>`</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula><inline-formula><mml:math id="ieqn13"><mml:mo>&#x2208;</mml:mo><mml:mi mathvariant="script">X</mml:mi></mml:math></inline-formula> are 2 input neighbor datasets, and <inline-formula><mml:math id="ieqn14"><mml:mi mathvariant="script">S</mml:mi><mml:mo>&#x2286;</mml:mo><mml:mi>R</mml:mi></mml:math></inline-formula>, where <inline-formula><mml:math id="ieqn15"><mml:mi>R</mml:mi></mml:math></inline-formula> is the set of all possible outputs, <inline-formula><mml:math id="ieqn16"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> is privacy loss or failure probability, and <inline-formula><mml:math id="ieqn17"><mml:mi>&#x03F5;</mml:mi></mml:math></inline-formula> is privacy budget.</p></list-item></list><p>An ideal DP mechanism provides a lower value of <inline-formula><mml:math id="ieqn18"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> and a smaller value of <inline-formula><mml:math id="ieqn19"><mml:mi>&#x03F5;</mml:mi></mml:math></inline-formula>. Unfortunately, these values decrease the function utility (eg, accuracy metric), so the main question is how much DP values we must perturb its output while guaranteeing trade-off privacy-utility. Intuitively, an output perturbation mechanism takes an input <inline-formula><mml:math id="ieqn20"><mml:mi>x</mml:mi></mml:math></inline-formula> and returns a random variable <inline-formula><mml:math id="ieqn21"><mml:mi>s</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:math></inline-formula>. This operation can be modeled by:</p><disp-formula id="E6"><label>(6)</label><mml:math id="eqn6"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mrow><mml:mi mathvariant="fraktur">M</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="italic">x</mml:mi></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="italic">s</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="italic">x</mml:mi></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:msub><mml:mi>N</mml:mi><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:math></disp-formula><p>where <inline-formula><mml:math id="ieqn22"><mml:mi>N</mml:mi></mml:math></inline-formula> is scaling noise sampled from a specific distribution. In this work, we chose Laplace and Gaussian mechanisms [<xref ref-type="bibr" rid="ref25">25</xref>] that use L1 and L2 norm sensitivity, respectively. The sensitivity function can be expressed as:</p><disp-formula id="E7"><label>(7)</label><mml:math id="eqn7"><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>f</mml:mi><mml:mo>=</mml:mo><mml:munder><mml:mrow><mml:mi mathvariant="normal">m</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">x</mml:mi></mml:mrow><mml:mrow><mml:mi>D</mml:mi><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>D</mml:mi></mml:mrow><mml:mrow><mml:mi>`</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:munder><mml:mfenced open="&#x2016;" close="&#x2016;" separators="|"><mml:mrow><mml:mi>s</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo><mml:mo>-</mml:mo><mml:mi>s</mml:mi><mml:msub><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>`</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>1,2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></disp-formula><p>Scaling noise can be computed as:</p><disp-formula id="E8"><label>(8)</label><mml:math id="eqn8"><mml:mi>&#x03C3;</mml:mi><mml:mo>=</mml:mo><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>f</mml:mi><mml:mo>/</mml:mo><mml:mi>&#x03B5;</mml:mi></mml:math></disp-formula><p>Output perturbation satisfies <inline-formula><mml:math id="ieqn23"><mml:mo>(</mml:mo><mml:mi>&#x03F5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo><mml:mo>-</mml:mo><mml:mi mathvariant="normal">D</mml:mi><mml:mi mathvariant="normal">P</mml:mi></mml:math></inline-formula> when we properly select the value scaling noise. Thus, it sampled from Laplace and Gaussian distributions [<xref ref-type="bibr" rid="ref25">25</xref>] as:</p><disp-formula id="E9"><label>(9)</label><mml:math id="eqn9"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mi>M</mml:mi><mml:mrow><mml:mrow><mml:mi mathvariant="normal">L</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">a</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">p</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">l</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">a</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">c</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">e</mml:mi></mml:mrow></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>f</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mi>s</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mi mathvariant="normal">L</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">p</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mi>&#x03BC;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="E10"><label>(10)</label><mml:math id="eqn10"><mml:msub><mml:mrow><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">G</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">u</mml:mi><mml:mi mathvariant="normal">s</mml:mi><mml:mi mathvariant="normal">s</mml:mi><mml:mi mathvariant="normal">i</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">n</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>,</mml:mo><mml:mi>f</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B5;</mml:mi><mml:mo>,</mml:mo><mml:mi>&#x03B4;</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mi>f</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo><mml:mo>+</mml:mo><mml:mi mathvariant="script">N</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:mi>&#x03BC;</mml:mi><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mfenced></mml:math></disp-formula><p>The gradient information leakage can be reduced by applying a gradient thresholding or clipping algorithm. As explained by Abadi et al [<xref ref-type="bibr" rid="ref11">11</xref>], gradient clipping is crucial in ensuring the DP of FL algorithms. So, each provider or client model update needs to have a bounded norm, which is ensured by applying an operation that shrinks individual model updates when their norm exceeds a given threshold. Clipping impacts of an FL algorithm&#x2019;s convergence performance should be known to create FL algorithms that protect DP.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Overview</title><p>Three scenarios are created to tackle the aforementioned challenges with the DP learning approaches: centralized, decentralized FL, and decentralized FL with DP. Their performances are evaluated on WESAD and VERBIO datasets on 2 different tasks. The first task is identifying users from a set of registered and recorded users. The second task is perceived binary stress recognition, which tries to distinguish the user&#x2019;s stress level, stress versus nonstress. We train a multitask deep learning model for handling these tasks simultaneously. In addition, for better training and to avoid overfitting, an early stopping regularization technique is used as gradient descent. The accuracy metric is used for measuring identification and stress recognition performance. In each simulation scenario, we run 5-fold cross-validation, where each fold is tested based on the training of the other 4. Inspired from a successful architecture [<xref ref-type="bibr" rid="ref30">30</xref>] and as described in <xref ref-type="table" rid="table2">Table 2</xref>, the multitask 1D-CNN model is based on 3 convolutional layers, a pooling layer, 2 fully connected layers, and 2 linear classifiers to classify the studied tasks. The multitask model uses the cross-entropy loss function and SGD learning rate (&#x03B2;=.0005).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>Hyperparameters of the multitask 1D-convolutional neural network&#x2013;based architecture.</p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Layer</td><td align="left" valign="bottom">Type</td><td align="left" valign="bottom">Hyperparameters</td></tr></thead><tbody><tr><td align="left" valign="top">Input</td><td align="left" valign="top">Input</td><td align="left" valign="top">Features size</td></tr><tr><td align="left" valign="top">Conv1D</td><td align="left" valign="top">Convolution</td><td align="left" valign="top">Input=1, output=20, <italic>K</italic>=8, stride=1, padding</td></tr><tr><td align="left" valign="top">ReLu</td><td align="left" valign="top">Activation function</td><td align="left" valign="top">ReLu</td></tr><tr><td align="left" valign="top">Pooling</td><td align="left" valign="top">Pooling</td><td align="left" valign="top">Stride=2, max pooling</td></tr><tr><td align="left" valign="top">Conv</td><td align="left" valign="top">Convolution layer</td><td align="left" valign="top">Input=20, output=40, <italic>K</italic>=8, stride=1, padding</td></tr><tr><td align="left" valign="top">Relu</td><td align="left" valign="top">Activation function</td><td align="left" valign="top">ReLu</td></tr><tr><td align="left" valign="top">Pooling</td><td align="left" valign="top">Pooling</td><td align="left" valign="top">Stride=2, max pooling</td></tr><tr><td align="left" valign="top">Conv</td><td align="left" valign="top">Convolution layer</td><td align="left" valign="top">Input=40, output=60, <italic>K</italic>=8, stride=1, padding</td></tr><tr><td align="left" valign="top">Relu</td><td align="left" valign="top">Activation function</td><td align="left" valign="top">ReLu</td></tr><tr><td align="left" valign="top">Pooling</td><td align="left" valign="top">Pooling</td><td align="left" valign="top">Stride=2, max pooling</td></tr><tr><td align="left" valign="top">Fully connected 1</td><td align="left" valign="top">Fully connected layer</td><td align="left" valign="top">Input=360, output=100</td></tr><tr><td align="left" valign="top">Fully connected 2</td><td align="left" valign="top">Fully connected layer</td><td align="left" valign="top">Input=100, output=300</td></tr><tr><td align="left" valign="top">Linear 1</td><td align="left" valign="top">Input: 100, output: 2</td><td align="left" valign="top">Output=2 affect classes, activation function:linear</td></tr><tr><td align="left" valign="top">Linear 2</td><td align="left" valign="top">Input: 300, output: N</td><td align="left" valign="top">Output=N identity classes, activation function:linear</td></tr></tbody></table></table-wrap><p>For each target task, the individual loss is determined by the cross-entropy for both stress recognition (Loss<sub>1</sub>) and identification tasks (Loss<sub>2</sub>). The individual losses are summed and form the total cost function (Loss<sub>T</sub>).</p></sec><sec id="s3-2"><title>CL Approach</title><p>Most existing stress recognition studies in conventional centralized ML settings use WESAD to position their works against state of the art [<xref ref-type="bibr" rid="ref31">31</xref>-<xref ref-type="bibr" rid="ref33">33</xref>] and achieve over 95% accuracy for binary stress recognition. Similarly, we used the CL approach on the WESAD dataset as a baseline experiment. We measured the performance of stress levels for each participant independently and then computed the average performance accuracy. Each individual participant model is trained using 5-fold cross-validation.</p><p><xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref> shows the results of our CL approach using a 1D-CNN multitask model for stress and identity recognition tasks on 2 datasets: WESAD and VERBIO. After training, the output layers are used to infer the stress level and the identity of the user, resulting in quite similar average scores of 99%. The results reveal a potential information leakage in this case, wherein model accuracy is preserved at the expense of the user&#x2019;s privacy. As a result, this approach fails to ensure the users&#x2019; privacy since their data are transmitted to the server for training purposes.</p></sec><sec id="s3-3"><title>Multitask Decentralized FL Approach (MFL)</title><p>Here, <italic>K</italic> (ie, participating clients) training models were created to train the whole dataset and the size of local samples <italic>D<sub>i</sub></italic>=1000. We set the number of training epochs (communication round) to <italic>T</italic>=40 and local training epochs to 1. Overall, the average accuracy result achieved is 97% and 95% for stress mood recognition and 93% and 90% for user identification on WESAD and VERBIO, respectively.</p><p>To examine the effect of client participation within the MFL model, we tested different numbers of clients, that is, <italic>K</italic>=5, <italic>K</italic>=10, and <italic>K</italic>=20. As reported by Wang et al [<xref ref-type="bibr" rid="ref34">34</xref>] and confirmed in <xref ref-type="fig" rid="figure2">Figure 2</xref>, an increasing number of clients and more client participation provide better performance for MFL training. The client distribution is different in assessing the MFL model in real-world conditions. We compare the convergence performance of the MFL model under IID and no IID for both tasks: stress recognition and subject identification (<xref ref-type="fig" rid="figure3">Figure 3</xref>). We note that the data distribution dramatically affects the quality of the FL training and obviously affects MFL&#x2019;s convergence performance. IID is a more idealistic case, whereas no IID is a more realistic case. Therefore, we expect the result to be lower in no IID cases. The performance difference proves the importance of data distribution.</p><p>Adjusting FL hyperparameter setting results can achieve a better performance than the CL approach. However, it may lead to a lower privacy level. As a result, SGD training may still reveal sensitive information about the client while exchanging the ML model with the global server.</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>The impact of the user participant size (various number of clients) on the multitask federated learning performance (wearable stress and affect detection).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig02.png"/></fig><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>The impact of IID versus no IID distribution on the multitask federated learning performance. IID: independent and identically distributed.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig03.png"/></fig></sec><sec id="s3-4"><title>MFL With DP Approach</title><list list-type="order"><list-item><p>To highlight the benefits of our proposed approach, we examine the impact of injecting noise into the local client training network according to these 3 scenarios: the full layers, shared layers, and task-specific layers (<xref ref-type="fig" rid="figure2">Figure 2</xref>). The used noise is sampled via the following mechanisms:</p></list-item><list-item><p>With DP technique&#x2013;based Laplace and Gaussian mechanisms, the noise scale is drawn from the output perturbation mechanism. DP parameters are computed at each local training round to generate appropriate noise injected from specific distributions (ie, Laplace and Gaussian). Besides appropriate (<inline-formula><mml:math id="ieqn24"><mml:mi>&#x03F5;</mml:mi></mml:math></inline-formula>,<inline-formula><mml:math id="ieqn25"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula>)&#x2212;DP initialization, there are a few hyperparameters to be tuned, such as the number of clients <inline-formula><mml:math id="ieqn26"><mml:mi>N</mml:mi></mml:math></inline-formula>, the number of maximum communication rounds <italic>T</italic>, and the number of chosen clients <inline-formula><mml:math id="ieqn27"><mml:mi>K</mml:mi></mml:math></inline-formula>.</p><list list-type="alpha-lower"><list-item><p>Laplace distribution [<xref ref-type="bibr" rid="ref11">11</xref>] is computed as:</p><p><disp-formula id="E11"><label>(11)</label><mml:math id="eqn11"><mml:mi>&#x03C3;</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x03F5;</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p></list-item><list-item><p>Gaussian distribution&#x2014;2 distributions are given by Abadi et al [<xref ref-type="bibr" rid="ref11">11</xref>] and Schmidt et al [<xref ref-type="bibr" rid="ref27">27</xref>], respectively:</p><p><disp-formula id="E12"><label>(12)</label><mml:math id="eqn12"><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:mfrac><mml:mrow><mml:mn>1.25</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow></mml:mfrac></mml:mrow><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:mfrac></mml:msqrt></mml:math></disp-formula></p><p><disp-formula id="E13"><label>(13)</label><mml:math id="eqn13"><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>f</mml:mi><mml:msqrt><mml:mn>2</mml:mn><mml:mi>q</mml:mi><mml:mi>T</mml:mi><mml:mi mathvariant="normal">l</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">g</mml:mi><mml:mfenced separators="|"><mml:mrow><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>&#x03B4;</mml:mi></mml:mrow></mml:mfrac></mml:mrow></mml:mfenced></mml:msqrt></mml:mrow><mml:mrow><mml:mi>&#x03B5;</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula></p></list-item></list></list-item></list><p><named-content content-type="indent">&#x2003;</named-content>where <inline-formula><mml:math id="ieqn28"><mml:mi mathvariant="normal">&#x0394;</mml:mi><mml:mi>f</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>U</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:math></inline-formula>, <inline-formula><mml:math id="ieqn29"><mml:mi>q</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>K</mml:mi></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></inline-formula>, and  <inline-formula><mml:math id="ieqn30"><mml:mi>C</mml:mi></mml:math></inline-formula> = clipping threshold. We set the clipping factor to 1 and <inline-formula><mml:math id="ieqn31"><mml:mi>&#x03B4;</mml:mi></mml:math></inline-formula> to 0.00001.</p><p><xref ref-type="fig" rid="figure4">Figures 4</xref><xref ref-type="fig" rid="figure5"/>-<xref ref-type="fig" rid="figure6">6</xref> show the accuracy comparison when adding Gaussian distribution levels into local training according to the 3 scenarios on the WESAD dataset. Compared to the baseline scenario, that is, no private mechanism, the perturbing share layer scheme with  <inline-formula><mml:math id="ieqn32"><mml:mi>&#x03C3;</mml:mi></mml:math></inline-formula>=0.1 and <inline-formula><mml:math id="ieqn33"><mml:mi>&#x03C3;</mml:mi></mml:math></inline-formula>=0.3 only provides better results for utility tasks; however, the identification task reached an accuracy of around 86%.</p><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Decentralized multitask federated learning approach results under different Gaussian noise levels (<italic>&#x03C3;</italic>=0.1).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig04.png"/></fig><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>Decentralized multitask federated learning approach results under different Gaussian noise levels (<italic>&#x03C3;</italic>=0.3).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig05.png"/></fig><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Decentralized multitask federated learning approach results under different Gaussian noise levels (<italic>&#x03C3;</italic>=0.6).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig06.png"/></fig><p>In this case, the amount of noise drawn from the Gaussian distribution is used to balance the utility and privacy and does not consider the FL settings parameters. The Gaussian levels are set to <italic>&#x03C3;</italic>=0.1, <italic>&#x03C3;</italic>=0.3, and <italic>&#x03C3;</italic>=0.6. For instance, increasing Gaussian noise leads to poor performance, as depicted in <xref ref-type="fig" rid="figure6">Figure 6</xref>.</p><p>To examine the DP impact on the utility-privacy trade-off, we assessed the performance of MFL with a DP mechanism in WESAD and VERBIO datasets under the aforementioned scenarios. The DP budgets are set as follows <italic>&#x03F5;</italic>=1, <italic>&#x03F5;</italic>=15, <italic>&#x03F5;</italic>=50 and <italic>&#x03F5;</italic>=2000 for this experiment. As depicted in <xref ref-type="fig" rid="figure7">Figures 7</xref><xref ref-type="fig" rid="figure8"/><xref ref-type="fig" rid="figure9"/><xref ref-type="fig" rid="figure10"/><xref ref-type="fig" rid="figure11"/><xref ref-type="fig" rid="figure12"/><xref ref-type="fig" rid="figure13"/>-<xref ref-type="fig" rid="figure14">14</xref> compared to the baseline scenario, that is, no privacy enhancing mechanism, adding DP noise into both shared and specific layers provides better results for utility performance; however, in terms of privacy, the perturbing specific task layer scheme provides better results than the perturbing shared layer. Results show that FL with perturbing all layers slows the convergence compared to others, although it provides better privacy (ie, decreasing identification accuracy).</p><fig position="float" id="figure7"><label>Figure 7.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (wearable stress and affect detection, <italic>&#x03F5;</italic>=1).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig07.png"/></fig><fig position="float" id="figure8"><label>Figure 8.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (virtual environment for real-time biometric interaction and observation, <italic>&#x03F5;</italic>=1).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig08.png"/></fig><fig position="float" id="figure9"><label>Figure 9.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (wearable stress and affect detection, <italic>&#x03F5;</italic>=5).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig09.png"/></fig><fig position="float" id="figure10"><label>Figure 10.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (virtual environment for real-time biometric interaction and observation, <italic>&#x03F5;</italic>=5).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig10.png"/></fig><fig position="float" id="figure11"><label>Figure 11.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (wearable stress and affect detection, <italic>&#x03F5;</italic>=15).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig11.png"/></fig><fig position="float" id="figure12"><label>Figure 12.</label><caption><p>Multitask federated learning performance under the differential privacy mechanism &#x201C;different layer perturbations&#x201D; (virtual environment for real-time biometric interaction and observation, <italic>&#x03F5;</italic>=15).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig12.png"/></fig><fig position="float" id="figure13"><label>Figure 13.</label><caption><p>The evaluation results of the multitask federated learning approach with 3 differential privacy (DP) distributions (various layer task perturbation, <italic>&#x03F5;</italic>=15).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig13.png"/></fig><fig position="float" id="figure14"><label>Figure 14.</label><caption><p>The evaluation results of the multitask federated learning approach with 3 differential privacy (DP) distributions (full perturbation, <italic>&#x03F5;</italic>=15).</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="mental_v11i1e60003_fig14.png"/></fig><p>Intuitively, our results demonstrate that adding noise to upper layers (identity recognition layers) effectively achieves a better privacy-utility trade-off. This advantage comes at the expense of a formal quantification of the relationship between learning features, that is, what we aim to share, and private variables, that is, what we aim to protect, which is rarely available in practice. We also evaluate the impact of noise distribution type on the proposed framework performance. The results from the WESAD dataset demonstrate that adding Laplace noise in our local training model preserves the stress recognition accuracy better than the Gaussian noise types (<xref ref-type="fig" rid="figure5">Figures 13</xref> and <xref ref-type="fig" rid="figure6">14</xref>). However, it also maintains the identity recognition task performance.</p><p>Nevertheless, using the Gaussian mechanism (ie, Gaussian 2) increases the privacy level of the local training model because increasing the number of global iterations will also negatively affect its global convergence performance, that is, a larger <inline-formula><mml:math id="ieqn34"><mml:mi>T</mml:mi></mml:math></inline-formula> would increase the noise level variance, dramatically decreasing the global accuracy model (<xref ref-type="disp-formula" rid="E13">equation 13</xref>). In addition, we have found that a larger <inline-formula><mml:math id="ieqn35"><mml:mi>K</mml:mi></mml:math></inline-formula> contributes to avoiding the vanishing local SGD gradient problem; however, a larger <italic>N</italic> leads to a scale-down in the variance of injected noise level to model parameters and fools the SGD training inference.</p><p>When DP is also used in FL settings, our experiments suggest that it achieves encouraging performance even on lower budget values (ie, increasing privacy requirements). The compromise of physiological data privacy can have significant consequences for a life of individuals. It provides an opening for data attackers and puts them at danger of data breaches, which can result in several threats [<xref ref-type="bibr" rid="ref35">35</xref>]. For instance, Shen et al [<xref ref-type="bibr" rid="ref36">36</xref>] provided an analytic study to estimate the leakage level of the global training FL model about user&#x2019;s information, and they generated a membership attack system on the global model to check if the target user is participated in training FL model or not. The experiences have been validated and tested on health data, called the University of Michigan Intern Health Study. The authors suggested to use DP to protect user-level privacy while maintaining the utility task performance. In similar work [<xref ref-type="bibr" rid="ref37">37</xref>], applied personalized FL in recognizing stress using mobile health data (ie, WESAD) to decrease the information leakage from the global FL model.</p><p>Besides, there are also other risks including the possibility of revealing a soft biometric (eg, gender, location, and authenticity) and a hard biometric (ie, identity). So far, most of the research in using physiological modalities in both biometric domains focused on ECG, electroencephalography, EDA, and photoplethysmography [<xref ref-type="bibr" rid="ref35">35</xref>,<xref ref-type="bibr" rid="ref38">38</xref>]. We found that the model obviously achieves a better balance between the stress and identification classification task when multimodal signals are combined and used as an input compared to the single modality. However, as aimed in our work, it is difficult to provide privacy guarantees since there is no study that provides an investigation on which modality can reveal sensitive information about the subject. In this experiment, we provide a comparative study among these used modalities for both tasks, subject identification and emotion recognition. According to <xref ref-type="table" rid="table3">Tables 3</xref> and <xref ref-type="table" rid="table4">4</xref>, we found that the ECG and acceleration modalities always led to an increase in the performance of the identification task. This fact might have contributed to their popularity in the biometric field [<xref ref-type="bibr" rid="ref38">38</xref>]. This observation hints us that discarding model parameters containing sensitive information learned from modalities associated with identity information might help to effectively achieve a good privacy-utility trade-off to a certain extent, that is, decrease information leakage. Although for these modalities (eg, ACC and ECG for WESAD), the identification accuracy is higher than binary stress recognition accuracy, in general, it is lower than stress recognition performance. We can interpret this fact by the number of labels and the complexity of the task. For WESAD, there are 15 participants and for VERBIO, there are 55 participants, which corresponds to 15 and 55 different labels for biometric identification ML systems. If we think that the signals of some people can be similar under the same experimental conditions, it can cause some confusion for the ML algorithm and can cause some performance decrease. On the other hand, for stress, we used binary labels, and although the stress response can change from individual to individual, it can be stated that it is a relatively simpler task with 2 labels.</p><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>The effect of modality on the stress and subject identification performance (wearable stress and affect detection).</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" colspan="2">Modalities and task</td><td align="left" valign="bottom" colspan="9">Approach</td></tr><tr><td align="left" valign="bottom" colspan="2"/><td align="left" valign="bottom" colspan="3">Centralized</td><td align="left" valign="bottom" colspan="3">FL<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup></td><td align="left" valign="bottom" colspan="3">FL+DP<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td></tr><tr><td align="left" valign="bottom" colspan="2"/><td align="left" valign="bottom">ACC<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup></td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup></td><td align="left" valign="bottom">ACC</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC</td><td align="left" valign="bottom">ACC</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="11"><bold>ACC</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">91.60</td><td align="left" valign="top">86.09</td><td align="left" valign="top">92.01</td><td align="left" valign="top">89.90</td><td align="left" valign="top">84.81</td><td align="left" valign="top">90.35</td><td align="left" valign="top">84.46</td><td align="left" valign="top">79.23</td><td align="left" valign="top">85.00</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">95.50</td><td align="left" valign="top">91.70</td><td align="left" valign="top">96.00</td><td align="left" valign="top">94.88</td><td align="left" valign="top">90.76</td><td align="left" valign="top">96.34</td><td align="left" valign="top">58.19</td><td align="left" valign="top">52.01</td><td align="left" valign="top">60.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>BVP<sup><xref ref-type="table-fn" rid="table3fn5">e</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">98.34</td><td align="left" valign="top">96.14</td><td align="left" valign="top">99.13</td><td align="left" valign="top">96.05</td><td align="left" valign="top">92.01</td><td align="left" valign="top">97.00</td><td align="left" valign="top">90.97</td><td align="left" valign="top">86.92</td><td align="left" valign="top">92.46</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">94.96</td><td align="left" valign="top">90.88</td><td align="left" valign="top">95.00</td><td align="left" valign="top">92.05</td><td align="left" valign="top">88.03</td><td align="left" valign="top">92.88</td><td align="left" valign="top">54.95</td><td align="left" valign="top">49.89</td><td align="left" valign="top">55.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>EDA<sup><xref ref-type="table-fn" rid="table3fn6">f</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">99.00</td><td align="left" valign="top">98.01</td><td align="left" valign="top">99.00</td><td align="left" valign="top">95.29</td><td align="left" valign="top">91.09</td><td align="left" valign="top">95.57</td><td align="left" valign="top">91.27</td><td align="left" valign="top">86.03</td><td align="left" valign="top">92.00</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">86.15</td><td align="left" valign="top">70.06</td><td align="left" valign="top">86.88</td><td align="left" valign="top">85.89</td><td align="left" valign="top">80.78</td><td align="left" valign="top">86.27</td><td align="left" valign="top">50.05</td><td align="left" valign="top">43.00</td><td align="left" valign="top">50.05</td></tr><tr><td align="left" valign="top" colspan="11"><bold>ECG<sup><xref ref-type="table-fn" rid="table3fn7">g</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">97.01</td><td align="left" valign="top">95.00</td><td align="left" valign="top">97.02</td><td align="left" valign="top">93.37</td><td align="left" valign="top">89.03</td><td align="left" valign="top">94.00</td><td align="left" valign="top">91.60</td><td align="left" valign="top">86.45</td><td align="left" valign="top">92.10</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">98.62</td><td align="left" valign="top">96.54</td><td align="left" valign="top">99.00</td><td align="left" valign="top">90.96</td><td align="left" valign="top">82.89</td><td align="left" valign="top">91.50</td><td align="left" valign="top">68.20</td><td align="left" valign="top">61.03</td><td align="left" valign="top">69.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>TEMP<sup><xref ref-type="table-fn" rid="table3fn8">h</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">99.00</td><td align="left" valign="top">98.00</td><td align="left" valign="top">99.00</td><td align="left" valign="top">94.88</td><td align="left" valign="top">90.76</td><td align="left" valign="top">95.23</td><td align="left" valign="top">90.00</td><td align="left" valign="top">85.00</td><td align="left" valign="top">90.00</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">97.88</td><td align="left" valign="top">95.78</td><td align="left" valign="top">98.00</td><td align="left" valign="top">91.22</td><td align="left" valign="top">86.01</td><td align="left" valign="top">92.00</td><td align="left" valign="top">51.23</td><td align="left" valign="top">43.00</td><td align="left" valign="top">52.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>All</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">99.86</td><td align="left" valign="top">96.72</td><td align="left" valign="top">99.86</td><td align="left" valign="top">95.56</td><td align="left" valign="top">91.38</td><td align="left" valign="top">96.13</td><td align="left" valign="top">90.00</td><td align="left" valign="top">85.00</td><td align="left" valign="top">90.00</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">98.56</td><td align="left" valign="top">96.47</td><td align="left" valign="top">99.00</td><td align="left" valign="top">92.00</td><td align="left" valign="top">88.00</td><td align="left" valign="top">92.00</td><td align="left" valign="top">47.00</td><td align="left" valign="top">40.00</td><td align="left" valign="top">47.00</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>FL: federated learning.</p></fn><fn id="table3fn2"><p><sup>b</sup>DP: differential privacy.</p></fn><fn id="table3fn3"><p><sup>c</sup>ACC: acceleration.</p></fn><fn id="table3fn4"><p><sup>d</sup>AUC: area under the curve.</p></fn><fn id="table3fn5"><p><sup>e</sup>BVP: blood volume pulse.</p></fn><fn id="table3fn6"><p><sup>f</sup>EDA: electrodermal activity.</p></fn><fn id="table3fn7"><p><sup>g</sup>ECG: electrocardiogram.</p></fn><fn id="table3fn8"><p><sup>h</sup>TEMP: skin temperature.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>The effect of modality on the stress and subject identification performance (virtual environment for real-time biometric interaction and observation).</p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" colspan="2">Modalities and task</td><td align="left" valign="bottom" colspan="9">Approach</td></tr><tr><td align="left" valign="bottom" colspan="2"/><td align="left" valign="bottom" colspan="3">Centralized</td><td align="left" valign="bottom" colspan="3">FL<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup></td><td align="left" valign="bottom" colspan="3">FL+DP<sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></td></tr><tr><td align="left" valign="bottom" colspan="2"/><td align="left" valign="bottom">ACC<sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup></td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="bottom">ACC</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC</td><td align="left" valign="bottom">ACC</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub></td><td align="left" valign="bottom">AUC</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="11"><bold>ACC</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">90.61</td><td align="left" valign="top">82.56</td><td align="left" valign="top">91.50</td><td align="left" valign="top">85.48</td><td align="left" valign="top">80.13</td><td align="left" valign="top">86.45</td><td align="left" valign="top">82.18</td><td align="left" valign="top">77.02</td><td align="left" valign="top">83.03</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject Identification</td><td align="left" valign="top">86.42</td><td align="left" valign="top">83.00</td><td align="left" valign="top">87.12</td><td align="left" valign="top">84.40</td><td align="left" valign="top">79.16</td><td align="left" valign="top">85.00</td><td align="left" valign="top">63.60</td><td align="left" valign="top">58.46</td><td align="left" valign="top">63.60</td></tr><tr><td align="left" valign="top" colspan="11"><bold>BVP<sup><xref ref-type="table-fn" rid="table4fn5">e</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">92.96</td><td align="left" valign="top">88.87</td><td align="left" valign="top">93.00</td><td align="left" valign="top">88.95</td><td align="left" valign="top">83.84</td><td align="left" valign="top">90.56</td><td align="left" valign="top">85.00</td><td align="left" valign="top">80.00</td><td align="left" valign="top">85.00</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject Identification</td><td align="left" valign="top">85.63</td><td align="left" valign="top">80.31</td><td align="left" valign="top">86.33</td><td align="left" valign="top">80.86</td><td align="left" valign="top">74.37</td><td align="left" valign="top">90.13</td><td align="left" valign="top">48.64</td><td align="left" valign="top">42.51</td><td align="left" valign="top">49.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>EDA<sup><xref ref-type="table-fn" rid="table4fn6">f</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">93.16</td><td align="left" valign="top">89.01</td><td align="left" valign="top">94.56</td><td align="left" valign="top">92.46</td><td align="left" valign="top">88.23</td><td align="left" valign="top">93.00</td><td align="left" valign="top">89.24</td><td align="left" valign="top">84.06</td><td align="left" valign="top">90.01</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">88.30</td><td align="left" valign="top">83.12</td><td align="left" valign="top">88.30</td><td align="left" valign="top">83.83</td><td align="left" valign="top">78.65</td><td align="left" valign="top">84.23</td><td align="left" valign="top">45.12</td><td align="left" valign="top">38.01</td><td align="left" valign="top">46.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>TEMP<sup><xref ref-type="table-fn" rid="table4fn7">g</xref></sup></bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">92.64</td><td align="left" valign="top">88.37</td><td align="left" valign="top">93.34</td><td align="left" valign="top">90.42</td><td align="left" valign="top">85.12</td><td align="left" valign="top">90.42</td><td align="left" valign="top">86.32</td><td align="left" valign="top">81.12</td><td align="left" valign="top">87.06</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject identification</td><td align="left" valign="top">85.34</td><td align="left" valign="top">80.07</td><td align="left" valign="top">86.17</td><td align="left" valign="top">82.56</td><td align="left" valign="top">76.33</td><td align="left" valign="top">82.56</td><td align="left" valign="top">39.59</td><td align="left" valign="top">31.42</td><td align="left" valign="top">40.00</td></tr><tr><td align="left" valign="top" colspan="11"><bold>All</bold></td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Stress</td><td align="left" valign="top">95.00</td><td align="left" valign="top">91.01</td><td align="left" valign="top">95.00</td><td align="left" valign="top">93.96</td><td align="left" valign="top">89.88</td><td align="left" valign="top">94.00</td><td align="left" valign="top">90.94</td><td align="left" valign="top">85.89</td><td align="left" valign="top">90.94</td></tr><tr><td align="left" valign="top"/><td align="left" valign="top">Subject Identification</td><td align="left" valign="top">90.00</td><td align="left" valign="top">85.01</td><td align="left" valign="top">90.00</td><td align="left" valign="top">88.92</td><td align="left" valign="top">83.84</td><td align="left" valign="top">89.10</td><td align="left" valign="top">38.46</td><td align="left" valign="top">30.18</td><td align="left" valign="top">39.35</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>FL: federated learning.</p></fn><fn id="table4fn2"><p><sup>b</sup>DP: differential privacy.</p></fn><fn id="table4fn3"><p><sup>c</sup>ACC: acceleration.</p></fn><fn id="table4fn4"><p><sup>d</sup>AUC: area under the curve.</p></fn><fn id="table4fn5"><p><sup>e</sup>BVP: blood volume pulse.</p></fn><fn id="table4fn6"><p><sup>f</sup>EDA: electrodermal activity.</p></fn><fn id="table4fn7"><p><sup>g</sup>TEMP: skin temperature.</p></fn></table-wrap-foot></table-wrap></sec><sec id="s3-5"><title>Comparison With State of the Art</title><p>The proposed approach is also compared with existing approaches used for stress recognition in FL settings, and the results are reported in <xref ref-type="table" rid="table5">Table 5</xref>. Most similar works have mainly focused on the benefits of FL in affective computing by generating data clients from unique datasets (ie, as used in the FL framework with standard benchmarks, namely Modified National Institute of Standards and Technology, Canadian Institute for Advanced Research-10, and Canadian Institute for Advanced Research-100). To the best of our knowledge, the number of studies using physiological signals in FL settings is relatively low compared to other case studies in affective computing, such as speech and facial expression&#x2013;based studies. According to <xref ref-type="table" rid="table5">Table 5</xref>, we can see that existing works used only FedAvg with the WESAD dataset and achieved similar performance, as confirmed by our previous study. To the best of our knowledge, our paper both improves privacy protection and provides slightly higher accuracies when compared with similar studies. Consistent with compared works, for both databases, WESAD and VERBIO, the obtained performance behaves similarly.</p><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>Comparison of affect recognition studies using federated learning (FL) and privacy-preserving approaches.<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Study</td><td align="left" valign="bottom">Dataset</td><td align="left" valign="bottom">FL algorithm</td><td align="left" valign="bottom">Data split</td><td align="left" valign="bottom">Accuracy</td></tr></thead><tbody><tr><td align="left" valign="top">Almadhor et al [<xref ref-type="bibr" rid="ref39">39</xref>]</td><td align="left" valign="top">WESAD<sup><xref ref-type="table-fn" rid="table5fn2">b</xref></sup></td><td align="left" valign="top">FedAvg<sup><xref ref-type="table-fn" rid="table5fn3">c</xref></sup>+logistic regression</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table5fn4">d</xref></sup></td><td align="left" valign="top">86.82</td></tr><tr><td align="left" valign="top">Fauzi et al [<xref ref-type="bibr" rid="ref40">40</xref>]</td><td align="left" valign="top">WESAD</td><td align="left" valign="top">FedAvg+DNN<sup><xref ref-type="table-fn" rid="table5fn5">e</xref></sup> network</td><td align="left" valign="top">N/A</td><td align="left" valign="top">85.75</td></tr><tr><td align="left" valign="top">Can and Ersoy [<xref ref-type="bibr" rid="ref15">15</xref>]</td><td align="left" valign="top">Private dataset</td><td align="left" valign="top">FedAvg+MLP<sup><xref ref-type="table-fn" rid="table5fn6">f</xref></sup></td><td align="left" valign="top">N/A</td><td align="left" valign="top">88.55</td></tr><tr><td align="left" valign="top">Lee et al [<xref ref-type="bibr" rid="ref29">29</xref>]</td><td align="left" valign="top">WESAD</td><td align="left" valign="top">FedAvg+MLP</td><td align="left" valign="top">LOOCV<sup><xref ref-type="table-fn" rid="table5fn7">g</xref></sup></td><td align="left" valign="top">75.00</td></tr><tr><td align="left" valign="top">Our previous study [<xref ref-type="bibr" rid="ref38">38</xref>]</td><td align="left" valign="top">WESAD</td><td align="left" valign="top">FedAvg+1 DCNN<sup><xref ref-type="table-fn" rid="table5fn8">h</xref></sup>+DP<sup><xref ref-type="table-fn" rid="table5fn9">i</xref></sup></td><td align="left" valign="top">5-fold CV<sup><xref ref-type="table-fn" rid="table5fn10">j</xref></sup></td><td align="left" valign="top">90.00</td></tr><tr><td align="left" valign="top">This study</td><td align="left" valign="top">VERBIO<sup><xref ref-type="table-fn" rid="table5fn11">k</xref></sup></td><td align="left" valign="top">FedAvg+1 DCNN+DP</td><td align="left" valign="top">5-fold CV</td><td align="left" valign="top">88.67</td></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>The table reports the main approaches have been applied for stress detection by using a physiological dataset.</p></fn><fn id="table5fn2"><p><sup>b</sup>WESAD: wearable stress and affect detection.</p></fn><fn id="table5fn3"><p><sup>c</sup>FedAvg: federated averaging.</p></fn><fn id="table5fn4"><p><sup>d</sup>N/A: not applicable.</p></fn><fn id="table5fn5"><p><sup>e</sup>DNN: deep neural network.</p></fn><fn id="table5fn6"><p><sup>f</sup>MLP: multilayer perceptron.</p></fn><fn id="table5fn7"><p><sup>g</sup>LOOCV: leave-one-out cross-validation.</p></fn><fn id="table5fn8"><p><sup>h</sup>DCNN: deep convolutional neural network.</p></fn><fn id="table5fn9"><p><sup>i</sup>DP: differential privacy.</p></fn><fn id="table5fn10"><p><sup>j</sup>CV: cross-validation.</p></fn><fn id="table5fn11"><p><sup>k</sup>VERBIO: virtual environment for real-time biometric interaction and observation.</p></fn></table-wrap-foot></table-wrap><p>Unlike the compared works [<xref ref-type="bibr" rid="ref15">15</xref>,<xref ref-type="bibr" rid="ref29">29</xref>,<xref ref-type="bibr" rid="ref38">38</xref>,<xref ref-type="bibr" rid="ref40">40</xref>], which analyze privacy concerns in stress recognition, we introduced a multitask model enabling us to strike a balance between utility and privacy. This is achieved by selectively adding noise to layers prone to subject information leakage. The comparison provided in <xref ref-type="table" rid="table5">Table 5</xref> reveals that prior studies achieved performance comparable to our case study when using FedAvg on WESAD or a private dataset. Through our experiments, we provided a comprehensive investigation of how to balance privacy preservation and stress recognition. The obtained results demonstrate that the combination of MFL and DP can maintain good performance without sacrificing the user&#x2019;s privacy.</p></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>The overall objective of our work was to present an approach to stress recognition that ensures robustness while protecting the user&#x2019;s privacy. To this end, we developed a personalized multitask federated model framework with DP. We used a user-level DP mechanism by injecting an amount of noise into personalized layers for perturbing identity while preserving task-specific utility. Using multitask and DP (<italic>&#x03B5;</italic>=15), we obtained 90% and 87% accuracy for recognizing emotions while limiting the reidentification accuracies to 47% and 38% on WESAD and VERBIO, respectively. We extensively tested different parameters including layer of neural network, privacy budget, and different noise distributions. As expected, adding noise to the upper layer decreases the affect recognition performance less when compared to the last task-specific layers. We also demonstrated the effect of data distribution on the performance. Having said that, the paper is not without limitations. Although we tested our algorithms with state-of-the-art datasets (WESAD 15 participants and VERBIO 55 participants), especially for the biometric task, larger datasets with a more heterogenous and higher number of participants should be used. Furthermore, these datasets were recorded in a controlled environment. Real-life or in-the-wild datasets will create more challenges and might require more advanced and complex architectures. However, with the available data, we managed to develop an automatic stress recognition system with around 90% accuracy by keeping biometric identification accuracy below 40%. We believe that our results will help researchers in determining suitable parameters and distributions for achieving the desired trade-off between utility and privacy for affective computing applications. Currently, new gradient-based unsupervised adversarial attackers are attacking deep neural classification models to infer the privacy of the distributed training gradient. In future works, to address this threat, we are planning to conduct additional experiments with federated differentially private generative adversarial networks that can provide better privacy protection and data diversity for widespread applications of physiological computing systems.</p></sec></body><back><ack><p>This work was carried out within the framework of the AI Production Network Augsburg. This work was partly funded by the German Research Foundation under the project "Health-relevant effects of different urban forest structures" (Learning About the Effects of Augsburg Forest). This work presents and discusses results in the context of the research project ForDigitHealth. The project is part of the Bavarian Research Association on Healthy Use of Digital Technologies and Media (ForDigitHealth), which is funded by the Bavarian Ministry of Science and Arts.</p></ack><fn-group><fn fn-type="con"><p>MB and YSC together conceptualized and designed the paper and interpreted the results. MB performed coding, data processing, and analysis. EA supervised the study, provided valuable suggestions for improvement, and critically revised the manuscript.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">CL</term><def><p>centralized learning</p></def></def-item><def-item><term id="abb2">CNN</term><def><p>convolutional neural network</p></def></def-item><def-item><term id="abb3">DP</term><def><p>differential privacy</p></def></def-item><def-item><term id="abb4">ECG</term><def><p>electrocardiogram</p></def></def-item><def-item><term id="abb5">EDA</term><def><p>electrodermal activity</p></def></def-item><def-item><term id="abb6">FedAvg</term><def><p>federated averaging</p></def></def-item><def-item><term id="abb7">FL</term><def><p>federated learning</p></def></def-item><def-item><term id="abb8">IID</term><def><p>independent and identically distributed</p></def></def-item><def-item><term id="abb9">MFL</term><def><p>multitask federated learning</p></def></def-item><def-item><term id="abb10">ML</term><def><p>machine learning</p></def></def-item><def-item><term id="abb11">MTL</term><def><p>multitask learning</p></def></def-item><def-item><term id="abb12">SGD</term><def><p>stochastic gradient descent</p></def></def-item><def-item><term id="abb13">VERBIO</term><def><p>virtual environment for real-time biometric interaction and observation</p></def></def-item><def-item><term id="abb14">WESAD</term><def><p>wearable stress and affect detection</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>De Silva</surname><given-names>LC</given-names> </name><name name-style="western"><surname>Miyasato</surname><given-names>T</given-names> </name><name name-style="western"><surname>Nakatsu</surname><given-names>R</given-names> </name></person-group><article-title>Facial emotion recognition using multi-modal information</article-title><conf-name>Proceedings of ICICS, International Conference on Information, Communications and Signal Processing Theme: Trends in Information Systems Engineering and Wireless Multimedia Communications</conf-name><conf-date>Sep 12, 1997</conf-date><conf-loc>Singapore</conf-loc><fpage>397</fpage><lpage>401</lpage><pub-id pub-id-type="doi">10.1109/ICICS.1997.647126</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Schuller</surname><given-names>B</given-names> </name><name name-style="western"><surname>Rigoll</surname><given-names>G</given-names> </name><name name-style="western"><surname>Lang</surname><given-names>M</given-names> </name></person-group><article-title>Hidden Markov model-based speech emotion recognition</article-title><conf-name>International Conference on Acoustics, Speech and Signal Processing (ICASSP&#x2019;03)</conf-name><conf-date>Jul 6-9, 2003</conf-date><conf-loc>Hong Kong, China</conf-loc><fpage>II</fpage><lpage>1</lpage><pub-id pub-id-type="doi">10.1109/ICASSP.2003.1202279</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="web"><article-title>Wearables unit shipments worldwide by vendor from 2014 to 2022 (in millions), by vendor</article-title><source>Statista</source><year>2020</year><access-date>2023-02-25</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.statista.com/statistics/515634/wearables-shipments-worldwide-by-vendor/">https://www.statista.com/statistics/515634/wearables-shipments-worldwide-by-vendor/</ext-link></comment></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>C</given-names> </name><name name-style="western"><surname>Hu</surname><given-names>X</given-names> </name><name name-style="western"><surname>Xie</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Gong</surname><given-names>M</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>B</given-names> </name></person-group><article-title>A privacy-preserving multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition</article-title><source>Front Neurorobot</source><year>2019</year><volume>13</volume><fpage>112</fpage><pub-id pub-id-type="doi">10.3389/fnbot.2019.00112</pub-id><pub-id pub-id-type="medline">31992979</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>Q</given-names> </name></person-group><article-title>A survey on multi-task learning</article-title><source>IEEE Trans Knowl Data Eng</source><year>2022</year><volume>34</volume><issue>12</issue><fpage>5586</fpage><lpage>5609</lpage><pub-id pub-id-type="doi">10.1109/TKDE.2021.3070203</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Dai</surname><given-names>W</given-names> </name><name name-style="western"><surname>Cahyawijaya</surname><given-names>S</given-names> </name><name name-style="western"><surname>Bang</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Fung</surname><given-names>P</given-names> </name></person-group><article-title>Weakly-supervised multi-task learning for multimodal affect recognition</article-title><source>arXiv</source><comment>Preprint posted online on  Apr 23, 2021</comment><pub-id pub-id-type="doi">10.48550/arXiv.2104.11560</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Sang</surname><given-names>DV</given-names> </name><name name-style="western"><surname>Cuong</surname><given-names>LTB</given-names> </name><name name-style="western"><surname>Van Thieu</surname><given-names>V</given-names> </name></person-group><article-title>Multi-task learning for smile detection, emotion recognition and gender classification</article-title><conf-name>SoICT &#x2019;17: Proceedings of the 8th International Symposium on Information and Communication Technology</conf-name><conf-date>Dec 7-8, 2017</conf-date><conf-loc>Nha Trang City, Viet Nam</conf-loc><fpage>340</fpage><lpage>347</lpage><pub-id pub-id-type="doi">10.1145/3155133.3155207</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Connolly</surname><given-names>HL</given-names> </name><name name-style="western"><surname>Young</surname><given-names>AW</given-names> </name><name name-style="western"><surname>Lewis</surname><given-names>GJ</given-names> </name></person-group><article-title>Recognition of facial expression and identity in part reflects a common ability, independent of general intelligence and visual short-term memory</article-title><source>Cogn Emot</source><year>2019</year><month>09</month><volume>33</volume><issue>6</issue><fpage>1119</fpage><lpage>1128</lpage><pub-id pub-id-type="doi">10.1080/02699931.2018.1535425</pub-id><pub-id pub-id-type="medline">30336725</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sheng</surname><given-names>W</given-names> </name><name name-style="western"><surname>Li</surname><given-names>X</given-names> </name></person-group><article-title>Multi-task learning for gait-based identity recognition and emotion recognition using attention enhanced temporal graph convolutional network</article-title><source>Pattern Recognit DAGM</source><year>2021</year><month>06</month><volume>114</volume><fpage>107868</fpage><pub-id pub-id-type="doi">10.1016/j.patcog.2021.107868</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>McMahan</surname><given-names>B</given-names> </name><name name-style="western"><surname>Moore</surname><given-names>E</given-names> </name><name name-style="western"><surname>Ramage</surname><given-names>D</given-names> </name><name name-style="western"><surname>Hampson</surname><given-names>S</given-names> </name><name name-style="western"><surname>y Arcas</surname><given-names>BA</given-names> </name></person-group><article-title>Communication-efficient learning of deep networks from decentralized data</article-title><conf-name>Proceedings of the 20th International Conference on Artificial Intelligence and Statistics</conf-name><conf-date>Apr 20-22, 2017</conf-date><conf-loc>Fort Lauderdale, FL, United States</conf-loc><fpage>1273</fpage><lpage>1282</lpage></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Abadi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Chu</surname><given-names>A</given-names> </name><name name-style="western"><surname>Goodfellow</surname><given-names>I</given-names> </name><etal/></person-group><article-title>Deep learning with differential privacy</article-title><conf-name>Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security</conf-name><conf-date>Oct 24-28, 2016</conf-date><conf-loc>Vienna, Austria</conf-loc><fpage>308</fpage><lpage>318</lpage><pub-id pub-id-type="doi">10.1145/2976749.2978318</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="report"><article-title>Regulation (EU) 2016/679 of the European Parliament and of the Council. Regulation (EU), 679</article-title><year>2016</year><access-date>2024-11-11</access-date><publisher-name>European Union</publisher-name><comment><ext-link ext-link-type="uri" xlink:href="https://eur-lex.europa.eu/eli/reg/2016/679/oj">https://eur-lex.europa.eu/eli/reg/2016/679/oj</ext-link></comment></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>de la Torre</surname><given-names>L</given-names> </name></person-group><article-title>A guide to the California Consumer Privacy Act of 2018</article-title><source>SSRN</source><comment>Preprint posted online on  Dec 21, 2018</comment><pub-id pub-id-type="doi">10.2139/ssrn.3275571</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Jin</surname><given-names>Q</given-names> </name><name name-style="western"><surname>Zhao</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wang</surname><given-names>S</given-names> </name></person-group><article-title>Multimodal multi-task learning for dimensional and continuous emotion recognition</article-title><conf-name>AVEC '17: Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge</conf-name><conf-date>Oct 23, 2017</conf-date><conf-loc>Mountain View, CA, United States</conf-loc><fpage>19</fpage><lpage>26</lpage><pub-id pub-id-type="doi">10.1145/3133944.3133949</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Can</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Ersoy</surname><given-names>C</given-names> </name></person-group><article-title>Privacy-preserving federated deep learning for wearable IoT-based biomedical monitoring</article-title><source>ACM Trans Internet Technol</source><year>2021</year><month>02</month><day>28</day><volume>21</volume><issue>1</issue><fpage>1</fpage><lpage>17</lpage><pub-id pub-id-type="doi">10.1145/3428152</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Somandepalli</surname><given-names>K</given-names> </name><name name-style="western"><surname>Qi</surname><given-names>H</given-names> </name><name name-style="western"><surname>Eoff</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Federated learning for affective computing tasks</article-title><conf-name>2022 10th International Conference on Affective Computing and Intelligent Interaction (ACII)</conf-name><conf-date>Oct 18-21, 2022</conf-date><conf-loc>Nara, Japan</conf-loc><fpage>1</fpage><lpage>8</lpage><pub-id pub-id-type="doi">10.1109/ACII55700.2022.9953856</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Latif</surname><given-names>S</given-names> </name><name name-style="western"><surname>Khalifa</surname><given-names>S</given-names> </name><name name-style="western"><surname>Rana</surname><given-names>R</given-names> </name><name name-style="western"><surname>Jurdak</surname><given-names>R</given-names> </name></person-group><article-title>Poster abstract: federated learning for speech emotion recognition applications</article-title><conf-name>2020 19th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN)</conf-name><conf-date>Apr 21-24, 2020</conf-date><conf-loc>Sydney, NSW, Australia</conf-loc><fpage>341</fpage><lpage>342</lpage><pub-id pub-id-type="doi">10.1109/IPSN48710.2020.00-16</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>G</given-names> </name><name name-style="western"><surname>Dai</surname><given-names>H</given-names> </name><name name-style="western"><surname>Bai</surname><given-names>Y</given-names> </name></person-group><article-title>DAFL: domain adaptation-based federated learning for privacy-preserving biometric recognition</article-title><source>Fut Gen Comput Syst</source><year>2024</year><month>01</month><volume>150</volume><fpage>436</fpage><lpage>450</lpage><pub-id pub-id-type="doi">10.1016/j.future.2023.09.020</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chhikara</surname><given-names>P</given-names> </name><name name-style="western"><surname>Singh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Tekchandani</surname><given-names>R</given-names> </name><name name-style="western"><surname>Kumar</surname><given-names>N</given-names> </name><name name-style="western"><surname>Guizani</surname><given-names>M</given-names> </name></person-group><article-title>Federated learning meets human emotions: a decentralized framework for human&#x2013;computer interaction for IoT applications</article-title><source>IEEE Int Things J</source><year>2020</year><volume>8</volume><issue>8</issue><fpage>6949</fpage><lpage>6962</lpage><pub-id pub-id-type="doi">10.1109/JIOT.2020.3037207</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Nandi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Xhafa</surname><given-names>F</given-names> </name></person-group><article-title>A federated learning method for real-time emotion state classification from multi-modal streaming</article-title><source>Methods</source><year>2022</year><month>08</month><volume>204</volume><fpage>340</fpage><lpage>347</lpage><pub-id pub-id-type="doi">10.1016/j.ymeth.2022.03.005</pub-id><pub-id pub-id-type="medline">35314343</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Zhao</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>J</given-names> </name><name name-style="western"><surname>Wu</surname><given-names>D</given-names> </name><name name-style="western"><surname>Teng</surname><given-names>J</given-names> </name><name name-style="western"><surname>Yu</surname><given-names>S</given-names> </name></person-group><article-title>Multi-task network anomaly detection using federated learning</article-title><conf-name>SoICT '19: Proceedings of the 10th International Symposium on Information and Communication Technology</conf-name><conf-date>Dec 4-6, 2019</conf-date><conf-loc>Hanoi, Ha Long Bay, Viet Nam</conf-loc><fpage>273</fpage><lpage>279</lpage><pub-id pub-id-type="doi">10.1145/3368926.3369705</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Smith</surname><given-names>V</given-names> </name><name name-style="western"><surname>Chiang</surname><given-names>CK</given-names> </name><name name-style="western"><surname>Sanjabi</surname><given-names>M</given-names> </name><name name-style="western"><surname>Talwalkar</surname><given-names>AS</given-names> </name></person-group><article-title>Federated MTL</article-title><conf-name>Advances in Neural Information Processing Systems</conf-name><conf-date>Dec 4-9, 2017</conf-date><conf-loc>Long Beach, CA, United States</conf-loc></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Shome</surname><given-names>D</given-names> </name><name name-style="western"><surname>Kar</surname><given-names>T</given-names> </name></person-group><article-title>FedAffect: few-shot federated learning for facial expression recognition</article-title><conf-name>Proceedings of the IEEE/CVF International Conference on Computer Vision</conf-name><conf-date>Oct 11-17, 2021</conf-date><conf-loc>Montreal, BC, Canada</conf-loc><fpage>4168</fpage><lpage>4175</lpage><pub-id pub-id-type="doi">10.1109/ICCVW54120.2021.00463</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="preprint"><person-group person-group-type="author"><name name-style="western"><surname>Feng</surname><given-names>T</given-names> </name><name name-style="western"><surname>Peri</surname><given-names>R</given-names> </name><name name-style="western"><surname>Narayanan</surname><given-names>S</given-names> </name></person-group><article-title>User-level differential privacy against attribute inference attack of speech emotion recognition on federated learning</article-title><source>arXiv</source><comment>Preprint posted online on  May 17, 2022</comment><pub-id pub-id-type="doi">10.48550/arXiv.2204.02500</pub-id></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Dwork</surname><given-names>C</given-names> </name><name name-style="western"><surname>Roth</surname><given-names>A</given-names> </name></person-group><article-title>The algorithmic foundations of differential privacy</article-title><source>FNT Theor Comput Sci</source><year>2014</year><volume>9</volume><issue>3-4</issue><fpage>211</fpage><lpage>407</lpage><pub-id pub-id-type="doi">10.1561/0400000042</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wei</surname><given-names>K</given-names> </name><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Ding</surname><given-names>M</given-names> </name><etal/></person-group><article-title>User-level privacy-preserving federated learning: analysis and performance optimization</article-title><source>IEEE Trans Mobile Comput</source><year>2021</year><volume>21</volume><issue>9</issue><fpage>3388</fpage><lpage>3401</lpage><pub-id pub-id-type="doi">10.1109/TMC.2021.3056991</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Schmidt</surname><given-names>P</given-names> </name><name name-style="western"><surname>Reiss</surname><given-names>A</given-names> </name><name name-style="western"><surname>Duerichen</surname><given-names>R</given-names> </name><name name-style="western"><surname>Marberger</surname><given-names>C</given-names> </name><name name-style="western"><surname>Van Laerhoven</surname><given-names>K</given-names> </name></person-group><article-title>Introducing WESAD, a multimodal dataset for wearable stress and affect detection</article-title><conf-name>Proceedings of the International Conference on Multimodal Interaction</conf-name><conf-date>Oct 16-20, 2018</conf-date><conf-loc>Boulder, CO, United States</conf-loc><pub-id pub-id-type="doi">10.1145/3242969.3242985</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yadav</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sakib</surname><given-names>MN</given-names> </name><name name-style="western"><surname>Nirjhar</surname><given-names>EH</given-names> </name><name name-style="western"><surname>Feng</surname><given-names>K</given-names> </name><name name-style="western"><surname>Behzadan</surname><given-names>AH</given-names> </name><name name-style="western"><surname>Chaspari</surname><given-names>T</given-names> </name></person-group><article-title>Exploring individual differences of public speaking anxiety in real-life and virtual presentations</article-title><source>IEEE Trans Affective Comput</source><year>2022</year><month>07</month><day>1</day><volume>13</volume><issue>3</issue><fpage>1168</fpage><lpage>1182</lpage><pub-id pub-id-type="doi">10.1109/TAFFC.2020.3048299</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>N</given-names> </name><name name-style="western"><surname>Pham</surname><given-names>V</given-names> </name><name name-style="western"><surname>Lee</surname><given-names>J</given-names> </name><name name-style="western"><surname>Chung</surname><given-names>TM</given-names> </name></person-group><article-title>Privacy preserving stress detection system using physiological data from wearable device</article-title><conf-name>Intelligent Human Systems Integration (IHSI 2023) Integrating People and Intelligent Systems</conf-name><conf-date>Feb 24-25, 2023</conf-date><conf-loc>Venice, Italy</conf-loc><pub-id pub-id-type="doi">10.54941/ahfe1002853</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Rezaei</surname><given-names>S</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>X</given-names> </name></person-group><article-title>Multitask learning for network traffic classification</article-title><conf-name>2020 29th International Conference on Computer Communications and Networks (ICCCN)</conf-name><conf-date>Aug 3-6, 2020</conf-date><conf-loc>Honolulu, HI, United States</conf-loc><pub-id pub-id-type="doi">10.1109/ICCCN49398.2020.9209652</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Abd-Alrazaq</surname><given-names>A</given-names> </name><name name-style="western"><surname>Alajlani</surname><given-names>M</given-names> </name><name name-style="western"><surname>Ahmad</surname><given-names>R</given-names> </name><etal/></person-group><article-title>The performance of wearable AI in detecting stress among students: systematic review and meta-analysis</article-title><source>J Med Internet Res</source><year>2024</year><month>01</month><day>31</day><volume>26</volume><fpage>e52622</fpage><pub-id pub-id-type="doi">10.2196/52622</pub-id><pub-id pub-id-type="medline">38294846</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kargarandehkordi</surname><given-names>A</given-names> </name><name name-style="western"><surname>Slade</surname><given-names>C</given-names> </name><name name-style="western"><surname>Washington</surname><given-names>P</given-names> </name></person-group><article-title>Personalized AI-driven real-time models to predict stress-induced blood pressure spikes using wearable devices: proposal for a prospective cohort study</article-title><source>JMIR Res Protoc</source><year>2024</year><month>03</month><day>25</day><volume>13</volume><issue>1</issue><fpage>e55615</fpage><pub-id pub-id-type="doi">10.2196/55615</pub-id><pub-id pub-id-type="medline">38526539</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>J</given-names> </name><name name-style="western"><surname>Washington</surname><given-names>P</given-names> </name></person-group><article-title>A comparison of personalized and generalized approaches to emotion recognition using consumer wearable devices: machine learning study</article-title><source>JMIR AI</source><year>2024</year><month>05</month><day>10</day><volume>3</volume><issue>1</issue><fpage>e52171</fpage><pub-id pub-id-type="doi">10.2196/52171</pub-id><pub-id pub-id-type="medline">38875573</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Wang</surname><given-names>M</given-names> </name><name name-style="western"><surname>Qin</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Liu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Li</surname><given-names>W</given-names> </name></person-group><article-title>Identifying personal physiological data risks to the Internet of Everything: the case of facial data breach risks</article-title><source>Humanit Soc Sci Commun</source><year>2023</year><volume>10</volume><issue>1</issue><fpage>216</fpage><pub-id pub-id-type="doi">10.1057/s41599-023-01673-3</pub-id><pub-id pub-id-type="medline">37192941</pub-id></nlm-citation></ref><ref id="ref35"><label>35</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>El Rahman</surname><given-names>SA</given-names> </name><name name-style="western"><surname>Alluhaidan</surname><given-names>AS</given-names> </name></person-group><article-title>Enhanced multimodal biometric recognition systems based on deep learning and traditional methods in smart environments</article-title><source>PLoS One</source><year>2024</year><volume>19</volume><issue>2</issue><fpage>e0291084</fpage><pub-id pub-id-type="doi">10.1371/journal.pone.0291084</pub-id><pub-id pub-id-type="medline">38358992</pub-id></nlm-citation></ref><ref id="ref36"><label>36</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Shen</surname><given-names>A</given-names> </name><name name-style="western"><surname>Francisco</surname><given-names>L</given-names> </name><name name-style="western"><surname>Sen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tewari</surname><given-names>A</given-names> </name></person-group><article-title>Exploring the relationship between privacy and utility in mobile health: algorithm development and validation via simulations of federated learning, differential privacy, and external attacks</article-title><source>J Med Internet Res</source><year>2023</year><month>04</month><day>20</day><volume>25</volume><fpage>e43664</fpage><pub-id pub-id-type="doi">10.2196/43664</pub-id><pub-id pub-id-type="medline">37079370</pub-id></nlm-citation></ref><ref id="ref37"><label>37</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Liu</surname><given-names>JC</given-names> </name><name name-style="western"><surname>Goetz</surname><given-names>J</given-names> </name><name name-style="western"><surname>Sen</surname><given-names>S</given-names> </name><name name-style="western"><surname>Tewari</surname><given-names>A</given-names> </name></person-group><article-title>Learning from others without sacrificing privacy: simulation comparing centralized and federated machine learning on mobile health data</article-title><source>JMIR Mhealth Uhealth</source><year>2021</year><month>03</month><day>30</day><volume>9</volume><issue>3</issue><fpage>e23728</fpage><pub-id pub-id-type="doi">10.2196/23728</pub-id><pub-id pub-id-type="medline">33783362</pub-id></nlm-citation></ref><ref id="ref38"><label>38</label><nlm-citation citation-type="confproc"><person-group person-group-type="author"><name name-style="western"><surname>Benouis</surname><given-names>M</given-names> </name><name name-style="western"><surname>Can</surname><given-names>YS</given-names> </name><name name-style="western"><surname>Andr&#x00E9;</surname><given-names>E</given-names> </name></person-group><article-title>A privacy-preserving multi-task learning framework for emotion and identity recognition from multimodal physiological signals</article-title><conf-name>2023 11th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)</conf-name><conf-date>Sep 10-13, 2023</conf-date><conf-loc>Cambridge, MA, United States</conf-loc><fpage>1</fpage><lpage>8</lpage><pub-id pub-id-type="doi">10.1109/ACIIW59127.2023.10388160</pub-id></nlm-citation></ref><ref id="ref39"><label>39</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Almadhor</surname><given-names>A</given-names> </name><name name-style="western"><surname>Sampedro</surname><given-names>GA</given-names> </name><name name-style="western"><surname>Abisado</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Wrist-based electrodermal activity monitoring for stress detection using federated learning</article-title><source>Sensors (Basel)</source><year>2023</year><month>04</month><day>14</day><volume>23</volume><issue>8</issue><fpage>3984</fpage><pub-id pub-id-type="doi">10.3390/s23083984</pub-id><pub-id pub-id-type="medline">37112323</pub-id></nlm-citation></ref><ref id="ref40"><label>40</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Fauzi</surname><given-names>MA</given-names> </name><name name-style="western"><surname>Yang</surname><given-names>B</given-names> </name><name name-style="western"><surname>Blobel</surname><given-names>B</given-names> </name></person-group><article-title>Comparative analysis between individual, centralized, and federated learning for smartwatch based stress detection</article-title><source>J Pers Med</source><year>2022</year><month>09</month><day>26</day><volume>12</volume><issue>10</issue><fpage>1584</fpage><pub-id pub-id-type="doi">10.3390/jpm12101584</pub-id><pub-id pub-id-type="medline">36294724</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Three different scenarios for adding noise with differential privacy: adding noise to only shared layers, task-specific layers, and full layers.</p><media xlink:href="mental_v11i1e60003_app1.png" xlink:title="PNG File, 29 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Stress versus identity recognition performance of multitask centralized learning using 2 different datasets.</p><media xlink:href="mental_v11i1e60003_app2.png" xlink:title="PNG File, 52 KB"/></supplementary-material></app-group></back></article>