<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">JMIR AI</journal-id>
      <journal-title>JMIR AI</journal-title>
      <issn pub-type="epub">2817-1705</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v3i1e47805</article-id>
      <article-id pub-id-type="pmid">38875667</article-id>
      <article-id pub-id-type="doi">10.2196/47805</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>Framework for Ranking Machine Learning Predictions of Limited, Multimodal, and Longitudinal Behavioral Passive Sensing Data: Combining User-Agnostic and Personalized Modeling</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Huo</surname>
            <given-names>Yuankai</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zheng</surname>
            <given-names>Lei</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Tomar</surname>
            <given-names>Ayush</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib id="contrib1" contrib-type="author" corresp="yes">
          <name name-style="western">
            <surname>Mullick</surname>
            <given-names>Tahsin</given-names>
          </name>
          <degrees>MEng</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Department of Systems and Information Engineering</institution>
            <institution>University of Virginia</institution>
            <addr-line>Olsson Hall, 151 Engineer's Way</addr-line>
            <addr-line>Charlottesville, VA, 22903</addr-line>
            <country>United States</country>
            <phone>1 4349245393</phone>
            <email>tum7q@virginia.edu</email>
          </address>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-2695-3809</ext-link>
        </contrib>
        <contrib id="contrib2" contrib-type="author">
          <name name-style="western">
            <surname>Shaaban</surname>
            <given-names>Sam</given-names>
          </name>
          <degrees>MBA</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0001-8273-1517</ext-link>
        </contrib>
        <contrib id="contrib3" contrib-type="author">
          <name name-style="western">
            <surname>Radovic</surname>
            <given-names>Ana</given-names>
          </name>
          <degrees>MD, MSc</degrees>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0002-4022-4151</ext-link>
        </contrib>
        <contrib id="contrib4" contrib-type="author">
          <name name-style="western">
            <surname>Doryab</surname>
            <given-names>Afsaneh</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">https://orcid.org/0000-0003-1575-385X</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
        <label>1</label>
        <institution>Department of Systems and Information Engineering</institution>
        <institution>University of Virginia</institution>
        <addr-line>Charlottesville, VA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff2">
        <label>2</label>
        <institution>NuRelm</institution>
        <addr-line>Pittsburgh, PA</addr-line>
        <country>United States</country>
      </aff>
      <aff id="aff3">
        <label>3</label>
        <institution>Department of Pediatrics</institution>
        <institution>University of Pittsburgh</institution>
        <addr-line>Pittsburgh, PA</addr-line>
        <country>United States</country>
      </aff>
      <author-notes>
        <corresp>Corresponding Author: Tahsin Mullick <email>tum7q@virginia.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection">
        <year>2024</year>
      </pub-date>
      <pub-date pub-type="epub">
        <day>20</day>
        <month>5</month>
        <year>2024</year>
      </pub-date>
      <volume>3</volume>
      <elocation-id>e47805</elocation-id>
      <history>
        <date date-type="received">
          <day>2</day>
          <month>4</month>
          <year>2023</year>
        </date>
        <date date-type="rev-request">
          <day>2</day>
          <month>7</month>
          <year>2023</year>
        </date>
        <date date-type="rev-recd">
          <day>16</day>
          <month>9</month>
          <year>2023</year>
        </date>
        <date date-type="accepted">
          <day>9</day>
          <month>4</month>
          <year>2024</year>
        </date>
      </history>
      <copyright-statement>©Tahsin Mullick, Sam Shaaban, Ana Radovic, Afsaneh Doryab. Originally published in JMIR AI (https://ai.jmir.org), 20.05.2024.</copyright-statement>
      <copyright-year>2024</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR AI, is properly cited. The complete bibliographic information, a link to the original publication on https://www.ai.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="https://ai.jmir.org/2024/1/e47805" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>Passive mobile sensing provides opportunities for measuring and monitoring health status in the wild and outside of clinics. However, longitudinal, multimodal mobile sensor data can be small, noisy, and incomplete. This makes processing, modeling, and prediction of these data challenging. The small size of the data set restricts it from being modeled using complex deep learning networks. The current state of the art (SOTA) tackles small sensor data sets following a singular modeling paradigm based on traditional machine learning (ML) algorithms. These opt for either a user-agnostic modeling approach, making the model susceptible to a larger degree of noise, or a personalized approach, where training on individual data alludes to a more limited data set, giving rise to overfitting, therefore, ultimately, having to seek a trade-off by choosing 1 of the 2 modeling approaches to reach predictions.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>The objective of this study was to filter, rank, and output the best predictions for small, multimodal, longitudinal sensor data using a framework that is designed to tackle data sets that are limited in size (particularly targeting health studies that use passive multimodal sensors) and that combines both user agnostic and personalized approaches, along with a combination of ranking strategies to filter predictions.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>In this paper, we introduced a novel ranking framework for longitudinal multimodal sensors (FLMS) to address challenges encountered in health studies involving passive multimodal sensors. Using the FLMS, we (1) built a tensor-based aggregation and ranking strategy for final interpretation, (2) processed various combinations of sensor fusions, and (3) balanced user-agnostic and personalized modeling approaches with appropriate cross-validation strategies. The performance of the FLMS was validated with the help of a real data set of adolescents diagnosed with major depressive disorder for the prediction of change in depression in the adolescent participants.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>Predictions output by the proposed FLMS achieved a 7% increase in accuracy and a 13% increase in recall for the real data set. Experiments with existing SOTA ML algorithms showed an 11% increase in accuracy for the depression data set and how overfitting and sparsity were handled.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>The FLMS aims to fill the gap that currently exists when modeling passive sensor data with a small number of data points. It achieves this through leveraging both user-agnostic and personalized modeling techniques in tandem with an effective ranking strategy to filter predictions.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>machine learning</kwd>
        <kwd>AI</kwd>
        <kwd>artificial intelligence</kwd>
        <kwd>passive sensing</kwd>
        <kwd>ranking framework</kwd>
        <kwd>small health data set</kwd>
        <kwd>ranking</kwd>
        <kwd>algorithm</kwd>
        <kwd>algorithms</kwd>
        <kwd>sensor</kwd>
        <kwd>multimodal</kwd>
        <kwd>predict</kwd>
        <kwd>prediction</kwd>
        <kwd>agnostic</kwd>
        <kwd>framework</kwd>
        <kwd>validation</kwd>
        <kwd>data set</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>Mobile and wearable sensing has garnered increasing interest in areas of physical health [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>], mental health [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref5">5</xref>], and activity recognition [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref7">7</xref>]. Multimodal passive sensing accommodates data collection without disrupting the human routine, allowing it to be an important tool to understand human behavior. However, passive sensing, unlike other forms of data, encounters common fundamental challenges in mobile health studies pertaining to physical and mental health. These challenges include small data sets, noisy or sparse data, and sensor selection criteria. Next, we explain these challenges and discuss how our framework can help in alleviating them.</p>
        <p>One of the primary challenges in passive sensing studies is small data sets. These arise due to limitations in the sample size of participants, the study duration, and ground truth restrictions. In this study, we explored this challenge from the viewpoint of studies conducted on passive sensing. Studies related to physical health (eg, [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>]) have investigated dietary behavior with the help of passive sensing. Participant sample sizes in Rabbi et al [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>] were 17 and 16, respectively, which is a limited participant count. This type of data limitation is even more prominent in mental health research that relies on passive sensing. Studies on depression [<xref ref-type="bibr" rid="ref3">3</xref>] and schizophrenia [<xref ref-type="bibr" rid="ref4">4</xref>], for example, had participant sample sizes of 28 and 5, respectively. The limited data sets in passive sensing research are also a factor of the study duration. To understand this, we can observe the duration of study. For example, the study duration in Rabbi et al [<xref ref-type="bibr" rid="ref1">1</xref>,<xref ref-type="bibr" rid="ref2">2</xref>] was 21 and 98 days, respectively, while the study by Canzian and Musolesi [<xref ref-type="bibr" rid="ref3">3</xref>] lasted for 70 days and that by Difrancesco et al [<xref ref-type="bibr" rid="ref4">4</xref>] was limited to only 5 days. The limitation in data led researchers away from using complex deep learning (DL) models, as demonstrated in previous studies [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref4">4</xref>]. This is because DL models have more hyperparameters and succumb to overfitting due to memorization of the data the models are trained on [<xref ref-type="bibr" rid="ref8">8</xref>]. In this study, we took inspiration from the existing work and selected specific traditional machine learning (ML) algorithms that are less susceptible to overfitting in small-data scenarios. However, unlike previous studies [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref17">17</xref>], we also ensured that our predictions were ranked based on 2 different modeling paradigms that further helped circumvent overfitting and also assisted in noise removal, as explained later.</p>
        <p>The second challenge commonly faced when tackling passive sensor data is that of sparsity or noise. This challenge arises due to signal inconsistencies and noise in sensor data collection because of software issues, data sync, or hardware problems. Discussions of sparsity and the negative effect it has on modeling have been previously documented [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref18">18</xref>-<xref ref-type="bibr" rid="ref20">20</xref>]. These studies have presented an overview of the passive sensing landscape and highlighted the role signal inconsistencies can play in predictive modeling of passively sensed data. The fact that data are noisy, especially in the case of wearable sensors, was mentioned by Plötz [<xref ref-type="bibr" rid="ref18">18</xref>]. Cornet and Holden [<xref ref-type="bibr" rid="ref19">19</xref>] reported that a lack of sensor precision leads to sparsity, and Xu et al [<xref ref-type="bibr" rid="ref20">20</xref>] documented the level of noise in data that prevents user-agnostic models from generalizing well. Our proposed framework attempts to reduce the effect of noise by forming a balance between predictions from user-agnostic modeling paradigms and personalized modeling paradigms. In addition, choosing specific ML algorithms, such as Extreme Gradient Boosting (XGBoost), Adaptive Boosting (AdaBoost), elastic-net, and extra-tree, and ranking predictions from them help lessen the impact of sparsity [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>].</p>
        <p>Sensor selection is the third type of challenge that has not received significant attention in passive or mobile sensing literature. Studies have tested various feature combinations mainly in the light of performing feature selection or feature reduction [<xref ref-type="bibr" rid="ref25">25</xref>]. Joshi and Boyd [<xref ref-type="bibr" rid="ref26">26</xref>] and Altenbach et al [<xref ref-type="bibr" rid="ref27">27</xref>], for example, used heuristic-based convex optimization to select sensors from an array of sensors. However, both these studies were purely from the perspective of sensor placement. They did not investigate which combination of sensors provided the best outcome for prediction-based modeling and were more in favor of wireless sensor network establishment. Mobile or wearable devices are laced with multiple sensors, and building and knowing which sensors create optimum models are vital particularly to mental and physical health–related studies. Through our framework, we present a way to test combinations of sensor data and derive and rank predictions from among those combinations, allowing investigators to understand which combinations of sensor data yield the best predictions for their passive sensing experimental setup.</p>
        <p>All the aforementioned challenges are common to passive sensing data sets. However, they exhibit significant presence in mental and physical health–related studies [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref4">4</xref>]. Xu et al [<xref ref-type="bibr" rid="ref20">20</xref>] talked of the general sequence of steps researchers take to build models and the struggles of working with passively sensed data. A strong framework to yield the best predictions can prove to be beneficial to the community at large and bring about greater insight from studies conducted with small data sets.</p>
        <p>In this paper, we present our ML modeling and ranking framework to address these challenges. The framework is designed to induce improved predictions for multimodal sensing. It balances both user-agnostic and personalized modeling of small data sets encountered often in mental and physical health–based studies. Our framework makes the following contributions: (1) prediction filtering and ranking through tensor-based aggregation of small, multimodal sensing data sets, (2) sensor combination selection to derive the best predictions, and (3) a reduction in overfitting predictions due to limited data and noise through ensembling of user-agnostic and personalized modeling strategies.</p>
        <p>Importantly, it should be noted that by the size of the data set, we refer to the final data sets where raw sensor readings are aggregated into intervals to align with the sampling frequency of ground truth data. In this work, we defined small data sets as those comprising fewer than 1000 data points for training ML models. Sparse or noisy data sets were those that either consisted of many zero entries or data sets for which highly varying sensor values were observed among different participants in the study.</p>
        <p>We evaluated the framework through its performance in the context of predicting changes in depression severity in a group of adolescent patients. The results showed the framework’s ability to use multiple modeling approaches for providing robust predictions in critical cases, such as mental health.</p>
        <p>Passive sensing data for human behavior modeling are different from other data formats, such as images, audio, or normal tabular data. Researchers in the field of passive sensing agree that passive sensing data have some common properties, such as they are time series data, multimodal, longitudinal, nonlinear, and noisy, as previously discussed [<xref ref-type="bibr" rid="ref20">20</xref>]. Xu et al [<xref ref-type="bibr" rid="ref20">20</xref>] also emphasized the researcher’s need for tools that can help ease the time lost in traversing the common pitfalls of passively sensed data. Our work endeavors to resolve such pitfalls for cases where passive sensing data are limited. Next, we discuss the related work highlighting the state of the art (SOTA) in passively sensed small, multimodal data sets.</p>
      </sec>
      <sec>
        <title>Related Work</title>
        <p>Despite the growing body of work using multimodal passive sensing in physical and mental health applications [<xref ref-type="bibr" rid="ref28">28</xref>-<xref ref-type="bibr" rid="ref32">32</xref>], there exists scope for improvement in small-data scenarios.</p>
        <p>In this section, we underline what exists in the current SOTA and why we need a ranking-based framework to address scenarios with small data sets. Keeping in line with our contribution, it will prove beneficial to present the current SOTA through understanding:</p>
        <list list-type="bullet">
          <list-item>
            <p>How traditional ML algorithms are applied in the context of passive sensing</p>
          </list-item>
          <list-item>
            <p>Why complex DL models do not work well in limited data scenarios</p>
          </list-item>
          <list-item>
            <p>How ensemble modeling has been adapted in passive sensing studies</p>
          </list-item>
          <list-item>
            <p>What the role of data fusion is in modeling passive sensing data</p>
          </list-item>
        </list>
        <sec>
          <title>Traditional Machine Learning Algorithms Applied in Passive Sensing</title>
          <p>Traditional ML algorithms have been applied to passive sensing in the space of human activity recognition (HAR) [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>], general health [<xref ref-type="bibr" rid="ref12">12</xref>-<xref ref-type="bibr" rid="ref15">15</xref>], and mental health [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. A deeper dive into the studies reveals some common takeaways that include the following:</p>
          <list list-type="bullet">
            <list-item>
              <p>All of them test multiple ML algorithms, followed by selecting predictions based on the overall chosen validation metric.</p>
            </list-item>
            <list-item>
              <p>They all follow a singular modeling strategy, resorting to either user-agnostic or personalized modeling.</p>
            </list-item>
            <list-item>
              <p>Cross-validation (CV) is either K-fold or leave-one-out CV.</p>
            </list-item>
          </list>
          <p>This is a repetition of steps that authors in the field make independently and is discussed extensively in the highlighted literature presented in <xref ref-type="table" rid="table1">Table 1</xref>. Following a single modeling strategy is restricting as choosing to follow a user-agnostic approach exposes the model to a greater degree of noise due to the heterogeneity in sensor values among participants, while solely following a personalized approach reduces data availability further as the model learns from individuals’ data rather than the general population data. Our endeavor through this ranking framework is to combine both the approaches, while using traditional ML algorithms.</p>
          <table-wrap position="float" id="table1">
            <label>Table 1</label>
            <caption>
              <p>Summary of SOTA<sup>a</sup> literature using traditional ML<sup>b</sup> for passive sensing, with special focus on CV<sup>c</sup>, the overall modeling strategy, and ML algorithms.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="200"/>
              <col width="200"/>
              <col width="200"/>
              <col width="200"/>
              <col width="200"/>
              <thead>
                <tr valign="top">
                  <td>Study</td>
                  <td>Application</td>
                  <td>CV</td>
                  <td>Modeling strategy</td>
                  <td>ML algorithm</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Kwapisz et al [<xref ref-type="bibr" rid="ref9">9</xref>]</td>
                  <td>HAR<sup>d</sup></td>
                  <td>10-fold</td>
                  <td>User agnostic</td>
                  <td>DT<sup>e</sup>, LR<sup>f</sup>, MLP<sup>g</sup></td>
                </tr>
                <tr valign="top">
                  <td>Shukla et al [<xref ref-type="bibr" rid="ref10">10</xref>]</td>
                  <td>HAR</td>
                  <td>5-fold</td>
                  <td>User agnostic</td>
                  <td>KNN<sup>h</sup>, SVM<sup>i</sup></td>
                </tr>
                <tr valign="top">
                  <td>Chen and Chen [<xref ref-type="bibr" rid="ref11">11</xref>]</td>
                  <td>HAR</td>
                  <td>10-fold</td>
                  <td>User agnostic</td>
                  <td>RF<sup>j</sup>, SVM, KNN</td>
                </tr>
                <tr valign="top">
                  <td>Huang et al [<xref ref-type="bibr" rid="ref12">12</xref>]</td>
                  <td>Sleep</td>
                  <td>10-fold</td>
                  <td>User agnostic</td>
                  <td>SVM</td>
                </tr>
                <tr valign="top">
                  <td>Montanini et al [<xref ref-type="bibr" rid="ref13">13</xref>]</td>
                  <td>Sleep</td>
                  <td>K-fold/leave 1 out</td>
                  <td>User agnostic/personalized</td>
                  <td>KNN, DT, RF, SVM</td>
                </tr>
                <tr valign="top">
                  <td>Teng et al [<xref ref-type="bibr" rid="ref14">14</xref>]</td>
                  <td>Parkinson’s tremors</td>
                  <td>5-fold</td>
                  <td>User agnostic</td>
                  <td>XGBoost<sup>k</sup>, DT, RF</td>
                </tr>
                <tr valign="top">
                  <td>Azam et al [<xref ref-type="bibr" rid="ref15">15</xref>]</td>
                  <td>Breath</td>
                  <td>K-fold</td>
                  <td>User agnostic</td>
                  <td>SVM</td>
                </tr>
                <tr valign="top">
                  <td>Canzian and Musolesi [<xref ref-type="bibr" rid="ref3">3</xref>]</td>
                  <td>Depression</td>
                  <td>Leave 1 out</td>
                  <td>User agnostic</td>
                  <td>SVM</td>
                </tr>
                <tr valign="top">
                  <td>Grunerbl et al [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                  <td>Bipolar disorder</td>
                  <td>K-fold</td>
                  <td>User agnostic/personalized</td>
                  <td>NB<sup>l</sup>, KNN, DT</td>
                </tr>
                <tr valign="top">
                  <td>Saeb et al [<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                  <td>Depression/anxiety</td>
                  <td>10-fold</td>
                  <td>User agnostic</td>
                  <td>XGBoost, DT</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table1fn1">
                <p><sup>a</sup>SOTA: state of the art.</p>
              </fn>
              <fn id="table1fn2">
                <p><sup>b</sup>ML: machine learning.</p>
              </fn>
              <fn id="table1fn3">
                <p><sup>c</sup>CV: cross-validation.</p>
              </fn>
              <fn id="table1fn4">
                <p><sup>d</sup>HAR: human activity recognition.</p>
              </fn>
              <fn id="table1fn5">
                <p><sup>e</sup>DT: decision tree.</p>
              </fn>
              <fn id="table1fn6">
                <p><sup>f</sup>LR: linear regression<sup>.</sup></p>
              </fn>
              <fn id="table1fn7">
                <p><sup>g</sup>MLP: multilayer perceptron<sup>.</sup></p>
              </fn>
              <fn id="table1fn8">
                <p><sup>h</sup>KNN: K-nearest neighbor<sup>.</sup></p>
              </fn>
              <fn id="table1fn9">
                <p><sup>i</sup>SVM: support vector machine.</p>
              </fn>
              <fn id="table1fn10">
                <p><sup>j</sup>RF: random forest<sup>.</sup></p>
              </fn>
              <fn id="table1fn11">
                <p><sup>k</sup>XGBoost: Extreme Gradient Boosting<sup>.</sup></p>
              </fn>
              <fn id="table1fn12">
                <p><sup>l</sup>NB: naive Bayes<sup>.</sup></p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
        </sec>
        <sec>
          <title>Limitation of Deep Learning in Small-Data Scenarios</title>
          <p>A common replacement for traditional ML algorithms is DL. Here, we explain why DL models are not ideal solutions for the problem addressed in this study. DL models have gained immense popularity in the literature [<xref ref-type="bibr" rid="ref33">33</xref>]. Their power lies in modeling the nonlinearity and noisy nature of passively sensed data. DL has a toolkit of strategies to handle small data that includes data augmentation [<xref ref-type="bibr" rid="ref1">1</xref>], transfer learning [<xref ref-type="bibr" rid="ref19">19</xref>], and ensembling [<xref ref-type="bibr" rid="ref29">29</xref>]. However, the size of a small data set in DL studies ranges from 1000 to 10,000 training points [<xref ref-type="bibr" rid="ref18">18</xref>]. This is unlike the ranking framework presented in this paper, which has been designed for data sets with fewer than 1000 data points. Therefore, despite their superiority in modeling larger passive sensing data sets, the performance of DL models suffers in cases where study data are limited and in the hundreds. The complexity of DL models results in overfitting to small data sets [<xref ref-type="bibr" rid="ref14">14</xref>]. In this paper, we worked to solve the problem of limiting data by providing researchers with a reproducible way to run multiple models and select the best predictions from among them. By using traditional ML in conjunction with ranked predictions from user-agnostic and personalized models, the issue of overfitting due to model complexity is dealt with in the proposed work.</p>
        </sec>
        <sec>
          <title>Ensemble Learning to Build Robust Models for Passive Sensing Data</title>
          <p>Among the different ways of dealing with overfitting, ensemble learning has been instrumental. Ensemble ML is a widely used approach in passive sensing studies [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref35">35</xref>]. It mainly exists in the form of boosting [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref34">34</xref>], bagging [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref16">16</xref>], weighted ensembles [<xref ref-type="bibr" rid="ref35">35</xref>], and max voting [<xref ref-type="bibr" rid="ref36">36</xref>] ML algorithms. Ensemble learning presents better results in terms of evaluation metrics. Ensemble learners are trained using a single modeling strategy. Therefore, they are either personalized ensembles [<xref ref-type="bibr" rid="ref35">35</xref>], which allows learners to derive interesting artifacts at personal levels, or user-agnostic ensembles [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>,<xref ref-type="bibr" rid="ref34">34</xref>,<xref ref-type="bibr" rid="ref36">36</xref>-<xref ref-type="bibr" rid="ref38">38</xref>], which only generate macrolevel information. Our contribution through the ranking framework is to provide a balance of both macrolevel patterns and user-specific patterns through a weighted ensemble of both approaches. Ensembling in this manner will allow us to reduce the noise that is picked up due to varying sensor values among users and account for user-specific patterns through the predictions on personalized data.</p>
        </sec>
        <sec>
          <title>Role of Data Fusion in Passive Sensing Studies</title>
          <p>The use of data fusion in passive sensing has seen a steady growth due to the use of multimodal sensors in passive sensing studies. Earlier studies were often restricted to single sensors, which were then manipulated to obtain a handful of features. For example, Canzian and Musolesi [<xref ref-type="bibr" rid="ref3">3</xref>] primarily used GPS sensor data, while Kwapisz et al [<xref ref-type="bibr" rid="ref9">9</xref>] only opted for an accelerometer to base their predictive modeling. The way data fusion is approached has a common link among the surveyed studies in the current literature. The studies have applied feature-level fusion [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref39">39</xref>-<xref ref-type="bibr" rid="ref43">43</xref>], where fusion takes place after feature extraction from raw signals. A single feature set is generated and then passed on to dimensionality reduction, such as linear discriminant analysis (LDA) [<xref ref-type="bibr" rid="ref10">10</xref>] or principal component analysis (PCA) [<xref ref-type="bibr" rid="ref40">40</xref>-<xref ref-type="bibr" rid="ref42">42</xref>]. The focus in these papers tends to be a reduction in dimension, without trying to study the impact of multiple distinct feature combinations. In comparison, our contribution of feature selection focuses on studying the relationship between each group of sensors by creating multiple feature sets based on sensor availability. This will allow us to select the best set of features to work with for a specific type of study. An illustration of the difference in the existing literature and our feature fusion approach is shown in <xref rid="figure1" ref-type="fig">Figure 1</xref> [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref39">39</xref>-<xref ref-type="bibr" rid="ref43">43</xref>].</p>
          <p>Overall, our ranking framework is motivated to aid researchers in situations in which data sets are small, sparse, or noisy and multimodal by taking advantage of its multiple model generation and the balanced outcome of the best predictions.</p>
          <fig id="figure1" position="float">
            <label>Figure 1</label>
            <caption>
              <p>(A) Data fusion approach in the current literature and (B) proposed FLMS data fusion approach, where s1-s6 represent distinct sensors and f1-f3 represent feature set combinations, which were then fused prior to ML modeling. FLMS: framework for longitudinal multimodal sensors; LDA: linear discriminant analysis; ML: machine learning; PCA: principal component analysis.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Ethical Considerations</title>
        <p>The data collection was approved by the Institutional Review Board of the University of Pittsburgh Human Research Protections Office (STUDY18120176).</p>
      </sec>
      <sec>
        <title>Data Description</title>
        <p>The study used passive sensing data and is presented through the lens of depression change prediction among adolescents. The data set comprised 55 adolescents from 12 to 17 years old, with an average age of 15.5 (SD 1.5) years. The AWARE app was used to collect the participants’ smartphone and Fitbit data. The data completeness rate for AWARE and Fitbit was, on average, 65.11% and 30.36%, respectively. The levels of completeness echoed the difficulty in collecting passive sensing data. Smartphone and Fitbit data were collected from each participant over 24 weeks.</p>
        <p>The 9-item Patient Health Questionnaire (PHQ-9) [<xref ref-type="bibr" rid="ref44">44</xref>] was used to collect weekly self-reports of depression severity from the participants. The questionnaire consists of a set of 9 questions, which can be scored from 0 to 3, giving a score range of 0-27. We used PHQ-9 scores as the ground truth to compare the prediction accuracy of our models.</p>
        <sec>
          <title>Relation of Sensor Data to Mental Health</title>
          <p>Raw sensor data, including calls, location, conversation, screen usage, Wi-Fi, steps, sleep, and heart rate, were processed, and relevant features were extracted at daily intervals. We used RAPIDS [<xref ref-type="bibr" rid="ref45">45</xref>] to extract 72 features from the sensors. The existing literature [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref46">46</xref>-<xref ref-type="bibr" rid="ref51">51</xref>] shows how location [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref50">50</xref>,<xref ref-type="bibr" rid="ref52">52</xref>], calls [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref53">53</xref>], screen usage [<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref55">55</xref>], conversations [<xref ref-type="bibr" rid="ref55">55</xref>-<xref ref-type="bibr" rid="ref58">58</xref>], Wi-Fi [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref59">59</xref>], steps [<xref ref-type="bibr" rid="ref60">60</xref>], and heart rate [<xref ref-type="bibr" rid="ref61">61</xref>] can be effective in predicting mental health behavior. Studies [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref50">50</xref>] have used location sensors, such as the GPS, and shown a strong relation to depressive symptom severity. Clinical measures, such as the PHQ-9 [<xref ref-type="bibr" rid="ref44">44</xref>], the PHQ-8 [<xref ref-type="bibr" rid="ref62">62</xref>], the Hamilton Rating Scale for Depression (HAM-D) [<xref ref-type="bibr" rid="ref63">63</xref>], and the Hamilton Rating Scale for Anxiety (HAM-A) [<xref ref-type="bibr" rid="ref64">64</xref>], have been used as target labels for prediction using sensor-based features, establishing a proof of association between sensor features and mental health predictions. Studies [<xref ref-type="bibr" rid="ref47">47</xref>,<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref51">51</xref>,<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref60">60</xref>] have used multimodal sensors of smartphones that included the sensors we chose for this study: calls, location, conversation, screen usage, Wi-Fi, Fitbit steps, and Fitbit heart rate. In the <italic>Results</italic> section, we further elaborate on the feature engineering from each of the sensors. The validity of using the sensors to predict mental health, in particular the choice of sensors, was motivated by the aforementioned studies, which showed strong predictive capability of sensors in the area of mental health prediction.</p>
        </sec>
      </sec>
      <sec>
        <title>Framework Design and Modeling</title>
        <p>We proposed a framework for longitudinal multimodal sensors (FLMS) as a ranking framework to rigorously handle longitudinal, multimodal sensor data and incorporate different analysis and modeling strategies suited for small and sparse time series data sets to produce better results. The FLMS incorporates 4 stages to improve, rank, and filter data set predictions (see <xref rid="figure1" ref-type="fig">Figure 1</xref>):</p>
        <list list-type="bullet">
          <list-item>
            <p>Stage 1: multimodal sensor fusion to explore the data set from multiple views and to identify the minimum number of sensors necessary to yield a good prediction. It also addresses sparsity.</p>
          </list-item>
          <list-item>
            <p>Stage 2: ML modeling with combined user-agnostic and personalized approach. This stage is designed to leverage user-agnostic and personalized predictions. The ML algorithms used in this stage were chosen due to their superior prediction capability in small-data scenarios and their ability to tackle sparse data sets.</p>
          </list-item>
          <list-item>
            <p>Stage 3: tensor-based aggregation and ranking leverage predictions from all fused combinations and modeling strategies to calculate more robust predictions.</p>
          </list-item>
          <list-item>
            <p>Stage 4: final prediction informed by the ensemble weighted average of both user-agnostic and personalized predictions to reduce the effect of overfitting in small data sets. This stage uses weights calculated via hamming distances to prevent any modeling approach from dominating the predictions.</p>
          </list-item>
        </list>
        <p>A high-level view in <xref rid="figure2" ref-type="fig">Figure 2</xref> illustrates how the FLMS is different from conventional ML approaches. Observing <xref rid="figure2" ref-type="fig">Figure 2</xref>A, we understand that the conventional modeling strategy uses a single algorithm with either a user-agnostic CV, where all users are included in the training and test sets, or a personalized CV strategy, where a single user’s data are used to derive predictions. However, <xref rid="figure2" ref-type="fig">Figure 2</xref>B displays how the FLMS uses different combinations of sensors as input data, followed by multiple algorithms and a combination of user-agnostic and personalized modeling. The modeling stage is followed by a ranking of predictions and finally an ensemble of the predictions to yield the final output.</p>
        <p>A detailed explanation of the stages of the FLMS and their utility is provided next.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>(A) Conventional modeling approach and (B) proposed FLMS approach. FLMS: framework for longitudinal multimodal sensors.</p>
          </caption>
          <graphic xlink:href="ai_v3i1e47805_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <sec>
          <title>Stage 1: Multimodal Sensor Fusion</title>
          <p>Stage 1 was designed for the early fusion of sensors at a feature level. Sensor fusions followed a combinatorial approach using <inline-graphic xlink:href="ai_v3i1e47805_fig12.png" xlink:type="simple" mimetype="image"/>, where Z is the total number of modalities available and x is the number of sensors to fuse. Our case study had 6-sensor modalities that generated a set of 63 separate data sets calculated as <inline-graphic xlink:href="ai_v3i1e47805_fig13.png" xlink:type="simple" mimetype="image"/>.</p>
          <p>Data set preprocessing steps involved normalization and log transforms. Imputations to fill missing feature observations were also conducted. The framework allowed for implementation of the K-nearest neighbor (KNN) algorithm for imputation, which is also the first level of defense against sparsity. The generated data sets were in 2D tabular data format. The sensor data were aggregated according to the granularity of the ground truth. Our case study collected PHQ-9 scores as an accepted depression measure. The total score range of the 9 questions was 0-27. This was collected on a weekly basis, and thus, our daily data were aggregated in weekly intervals.</p>
        </sec>
        <sec>
          <title>Stage 2: ML Modeling With a Combined User-Agnostic and Personalized Approach</title>
          <p>Stage 2 focused on modeling and predictions based on the data sets generated in stage 1. All stage 1 data sets were run through the modeling suite, which encompasses a series of ML algorithms and CV strategies to help build user-agnostic and personalized models.</p>
          <p>The ML suite includes case-specific linear and nonlinear algorithms. For our case study on adolescent depression, we followed a regression-based approach, and therefore, we selected algorithms such as linear regression (LR), elastic-net, random forest (RF), AdaBoost, extra-tree, gradient boosting, and XGBoost. The algorithms were chosen based on (1) their performance in the existing literature when working with small data and robustness to sparsity, and (2) tree-based models, which were specifically chosen to provide added tractability for researchers to inspect which features mainly contributed to the models’ predictive capability. The algorithms were used in each modeling strategy. The predictions of the ML algorithms for each time unit were stored in arrays for each participant and later used to select the best model for each participant. The best model selection strategy chose the model with the minimum error (in the case of regression) or the maximum accuracy (in the case of classification) among all algorithms. For example, among l number of regression algorithms, the best model was chosen as follows:</p>
          <disp-formula>
            <graphic xlink:href="ai_v3i1e47805_fig14.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>
            <bold>(1)</bold>
          </p>
          <p>,where alg refers to the algorithm with the lowest absolute sum error and pred<sub>m</sub>(alg<sub>t</sub>) is the prediction made by an algorithm l at unit time t. The array of prediction by the best model was retained for each respective participant.</p>
        </sec>
        <sec>
          <title>User-Agnostic Model Building</title>
          <p>To leverage as much data as possible, we implemented the leave-one-participant-out (LOPO) and leave-time-unit-X-out (LTXO) strategies. This is illustrated in <xref rid="figure3" ref-type="fig">Figure 3</xref>A,B.</p>
          <p>In LOPO, we held out all data from a single participant for validation and trained the model on other participants. This strategy reflected the cold start case where a new user started using the health app.</p>
          <p>The LTXO is based on the unit of time for ground truth data (eg, a week). For training, we held out a given time unit of all participants and trained the model on the rest of the time units. This strategy evaluated the impact of time-specific segments of data on prediction. The training phase captures the similarity and variation of data during different time units to build user-agnostic models.</p>
          <fig id="figure3" position="float">
            <label>Figure 3</label>
            <caption>
              <p>User-agnostic model building: (A) LOPO and (B) LTXO strategies. Algo: algorithm; LOPO: leave one participant out; LTXO: leave time unit X out.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig3.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Personalized Model Building</title>
          <p>The personalized modeling strategy leverages each user’s historical and cross-time data samples in a sliding window and the leave-one-time-unit-out approach.</p>
          <p>For each participant, the accumulated-time-unit (ATU) strategy built a model from X<sub>t</sub> time units of data to predict X<sub>t+1</sub>. For example, the model built from weeks 1 and 2 predicted depression in week 3. In the next iteration, the sliding window was increased by T time units (eg, 2 weeks) to repeat the model-building process. This process continued until the maximum number of time units was reached. This method examined the forecasting capability of the framework.</p>
          <p>The leave-one-time-unit-one-participant-out (LOTPO) strategy trained the models on all time units of a participant across time to predict the target label for the current time unit. For example, for a participant with 10 weeks of data, we built a model from data in weeks 1-5 and weeks 7-10 to predict depression in week 6. This method evaluated the feasibility of past and future data for each participant to predict an outcome (<xref rid="figure4" ref-type="fig">Figure 4</xref>A,B).</p>
          <fig id="figure4" position="float">
            <label>Figure 4</label>
            <caption>
              <p>Personalized model building: (A) LOTPO and (B) ATU strategies. Algo: algorithm; ATU: accumulated time unit; LOTPO: leave one time unit of participant out.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig4.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Stage 3: Tensor-Based Aggregation and Ranking</title>
          <p>The output of stage 2 was a set of best prediction matrices for sensor fusion combinations, where each slot in the matrix represented prediction results for a participant in a particular time unit. We represented these predictions in the form of Z-dimensional tensors (<xref rid="figure5" ref-type="fig">Figure 5</xref>), where Z is the number of modalities being used. For example, a study with 6 modalities and 45 users over 24 weeks was represented in tensor form as (6, 45, 24). The tensor representation helped represent the high dimensionality of sensor combinations.</p>
          <p>The predicted values for each slot across tensors were then aggregated using an aggregation function (eg, mean). This process took advantage of the stage 2 combinations to help reduce the error in prediction. For example, we aggregated predictions of 6 tensors (generated from 5-sensor fusion) into 1 tensor by calculating the mean of the predictions from the 6 combinations (see <xref rid="figure3" ref-type="fig">Figure 3</xref>). This was done for both user-agnostic and personalized models. The aggregated mean was calculated using the following equation:</p>
          <disp-formula>
            <graphic xlink:href="ai_v3i1e47805_fig15.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>
            <bold>(2)</bold>
          </p>
          <p>,where M<sub>agg</sub> is the aggregated mean, k is the total number of sensor combinations aggregated, i is the combination number, j is the corresponding time unit, and <inline-graphic xlink:href="ai_v3i1e47805_fig16.png" xlink:type="simple" mimetype="image"/> is the prediction across each set of combinations. The data were now in a format where each 2D tensor represented a particular sensor fusion prediction set (<xref rid="figure6" ref-type="fig">Figure 6</xref>).</p>
          <p>The predictions were next encoded into 0s and 1s to counter the large variance in the regression values from the original values. This logic can be set based on the type of ML problem the framework is being used to address. For example, in our case study, if the regressed change in depression score values was 0 or negative value, we classified it as 0, and if it was positive, we represented it as 1 (<xref rid="figure7" ref-type="fig">Figure 7</xref>).</p>
          <p>The next step in this stage measured the hamming distance between the 0-1‑encoded tensor and the true labels tensor, as shown in <xref rid="figure8" ref-type="fig">Figure 8</xref>. These hamming distances were then aggregated (D<sub>u</sub>) for the respective 2D tensor as follows:</p>
          <disp-formula>
            <graphic xlink:href="ai_v3i1e47805_fig17.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>
            <bold>(3)</bold>
          </p>
          <p>,where d(p<sub>i</sub>, a<sub>i</sub>) is the hamming distance between unit time predictions p<sub>i</sub> and the true value a<sub>i</sub>. Based on the measured distance, we ranked and chose the best set of predictions. This metric helped inform the choice of weightage to associate with a particular modeling strategy. The hamming distance helped further reduce errors after encoding and filtered down to the best set of predictions from each strategy.</p>
          <fig id="figure5" position="float">
            <label>Figure 5</label>
            <caption>
              <p>An example of tensor representation of 6-sensor fusion predictions.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig5.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure6" position="float">
            <label>Figure 6</label>
            <caption>
              <p>Instance of ATU where it shows how the mean aggregated prediction set is generated according to Equation (2). ATU: accumulated time unit; avg: average.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig6.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure7" position="float">
            <label>Figure 7</label>
            <caption>
              <p>The 0-1 encoding process resolves dealing with large variances in regression values. ATU: accumulated time unit; LOPO: leave one participant out; LOTPO: leave one time unit of participant out; LTXO: leave time unit X out.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig7.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure8" position="float">
            <label>Figure 8</label>
            <caption>
              <p>Hamming distance calculations reduce error and also determine the weight each of the 4 modeling approaches will contribute to stage 4’s ensembled weighted average. ATU: accumulated time unit.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig8.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
        <sec>
          <title>Stage 4: Weighted Ensemble</title>
          <p>The final stage formed the most robust set of predictions via an ensemble weighted average approach, where weights were calculated based on the minimum hamming distances derived from each modeling strategy in stage 3 (<xref rid="figure9" ref-type="fig">Figure 9</xref>):</p>
          <disp-formula>
            <graphic xlink:href="ai_v3i1e47805_fig18.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </disp-formula>
          <p>
            <bold>(4)</bold>
          </p>
          <p>,where P<sub>ij</sub> is the prediction tensor, w<sub>k</sub> is the weight based on the minimum hamming distance, and i and j are the number of users and time units, respectively. The data were then encoded back to 0s and 1s. A complete version of the FLMS with all its stages is presented in <xref rid="figure10" ref-type="fig">Figure 10</xref> (see <xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref> for a higher quality image). </p>
          <fig id="figure9" position="float">
            <label>Figure 9</label>
            <caption>
              <p>Ensemble average based on weights derived from the hamming distance to arrive at best-ranked predictions. ATU: accumulated time unit; LOPO: leave one participant out; LOTPO: leave one time unit of participant out; LTXO: leave time unit X out.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig9.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
          <fig id="figure10" position="float">
            <label>Figure 10</label>
            <caption>
              <p>FLMS ranking overview. Algo: algorithm; ATU: accumulated time unit; avg: average; CV: cross-validation; FLMS: framework for longitudinal multimodal sensors; LOPO: leave one participant out; LOTPO: leave one time unit of participant out; LTXO: leave time unit X out; ML: machine learning.</p>
            </caption>
            <graphic xlink:href="ai_v3i1e47805_fig10.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
          </fig>
        </sec>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Stagewise Description of Framework Processing on an Adolescent Data Set</title>
        <p>To evaluate the performance of the proposed FLMS, we used a depression data set of adolescents. This was a small data set, comprising noisy, multimodal sensor values from multiple participants—a suitable case study for our purpose of evaluating the performance of our proposed framework. Before presenting the experimental results, we first provide an understanding of how the adolescent data set was processed at each stage of the FLMS.</p>
        <p>The passively sensed depression data set was longitudinal, with a varying number of observations per participant. The goal was to predict changes in the depression score. This was achieved by passing the small set of observations through our ranking framework, which processed, modeled, ranked, and output the best set of overall predictions based on multiple modeling approaches. A prediction of change in depression is difficult and becomes even more challenging when the amount of data provided to the ML algorithms is limited.</p>
        <sec>
          <title>Stage 1 Outcome</title>
          <p>As part of stage 1, daily data were aggregated in weekly intervals to align with weekly ground truth values. Based on our extensive exploratory data analysis (EDA), we set thresholds for sparsity and adopted KNN as the imputation strategy.</p>
          <p>Our final data set consisted of 507 data points with 72 features, with an average of 13 weekly data points per participant. A series of data sets were then produced from an early fusion of 6-sensor features. Each data set retained 45 (81.8%) of the 55 participants. We had to drop 11 (20%) participants as they were missing more than 60% of their sensor data. The true depression state of the participants was given by the PHQ-9 weekly survey. The change in participant depression scores was calculated as W<sub>m</sub> – W<sub>m–1</sub>, where W<sub>m</sub> is the score for the m-th week; this served as the ground truth for our analysis.</p>
        </sec>
        <sec>
          <title>Stage 2: ML Modeling Outcome</title>
          <p>The ML algorithms in stage 2 regressed on the change in the depression score, with positive changes exhibiting a rise in the depression score in that week, negative changes representing a decrease, and 0 marking no change. The best predictive models of depression for each participant were built and selected following the steps in stage 2.</p>
        </sec>
        <sec>
          <title>Stage 3: Encoding and Prediction Filtering Outcome</title>
          <p>This led to stage 3, where after the mean aggregation, we encoded the regressed values as our goal was to predict whether the change in the depression score was positive, negative, or constant, rather than determining the exact value of the change. This step was followed by hamming distance calculations to further rank and filter the best set of predictions.</p>
        </sec>
        <sec>
          <title>Stage 4: Final Prediction Ensembling of Adolescent Data</title>
          <p>The predictions evaluated by the minimum hamming distances entered stage 4, where we calculated the final ensemble predictions. The predictions used weights determined by hamming distance calculations, which enabled us to balance between personalized and user-agnostic models. This step completed the offline training and prediction of change in depression in the adolescent data set.</p>
        </sec>
      </sec>
      <sec>
        <title>Experiment Design and Results</title>
        <p>In this section, we present the depression change prediction results of the FLMS. The experiments were designed to test the framework’s claims of reducing overfitting on a small data set, reducing the impact of noise or sparsity, and identifying the best combination for sensor fusion.</p>
        <p>We conducted 3 main experiments in support of our claims:</p>
        <list list-type="bullet">
          <list-item>
            <p>Experiment 1 tested FLMS predictions against singular modeling strategies used in SOTA. This experiment evaluated our claim regarding the advantage of the overall framework that took steps to reduce noise and identify the best sensor combinations versus a singular modeling strategy.</p>
          </list-item>
          <list-item>
            <p>Experiment 2 was a SOTA comparison test conducted to evaluate how our prediction-ranking framework performed in comparison to existing ML and DL approaches used in the current literature. This comparison also substantiated the FLMS performance to overfitting versus the existing strategies in the literature from prediction in small-data scenarios.</p>
          </list-item>
          <list-item>
            <p>Experiment 3 was designed to compare the FLMS performance with that of commonly used ML algorithms that have been shown to perform well with sparse data. It is important to note that there is an overlap of ML algorithms used to tackle sparsity and those used in passive sensing studies for mental health, particularly for small data sets.</p>
          </list-item>
        </list>
        <sec>
          <title>Evaluation Metrics</title>
          <p>The task of the FLMS is to model, rank, and output the best set of predictions from multiple modeling approaches. The output of the FLMS are predictions encoded as 0s or 1s (ie, binary values). Therefore, our choice of evaluation metrics for the framework predictions was the average accuracy, average recall, and average <italic>F</italic><sub>1</sub>-scores amongst users.</p>
        </sec>
        <sec>
          <title>Experiment Metadata</title>
          <p>The metadata pertaining to each experiment is provided at the end of the experiments. The information included as metadata is based on the best practices used [<xref ref-type="bibr" rid="ref65">65</xref>] to help with reproducibility of results. They include (1) feature preprocessing steps, (2) modeling CV strategy, (3) ML algorithms used, (4) random state, and (5) evaluation metrics specific to the experiments. They are presented in the form of tables following the corresponding results for each experiment.</p>
        </sec>
        <sec>
          <title>Data Set Used in the Experiments</title>
          <p>To standardize our experiments, we maintained a consistent data set, a combination of 6-sensor feature sets that included calls, location, screen usage, conversation, Fitbit, and Wi-Fi. After the stages of preprocessing, missing data imputation using the KNN strategy, and the removal of highly corelated features, the final data set comprised 61 features and 507 data points belonging to a total of 45 (81.8%) participants.</p>
        </sec>
        <sec>
          <title>Feature Engineering in Experiments</title>
          <p>Since we maintained a consistent data set for all our experiments, feature engineering for all the experiments was achieved through data collected from 6 sensors. As discussed earlier, the data were collected from participants’ smartphones using the AWARE app [<xref ref-type="bibr" rid="ref66">66</xref>] and then passed through the RAPIDS application programming interface (API). The features extracted using the API are discussed in detail next.</p>
          <sec>
            <title>Call Sensor Features</title>
            <p>The calls sensor features provide a context of how frequently the user has been in contact with someone else. Studies have revealed that higher degrees of depression are linked to reduced contact with social circles [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref53">53</xref>]. As part of call sensor features, we extracted the total number of missed calls; the counts of missed calls from distinct contacts, calls from the most frequent contacts for a time segment, incoming calls, and outgoing calls; the mean (SD), maximum, and minimum duration of both incoming and outgoing calls; and the entropy duration of outgoing and incoming calls, which provided an estimate of the Shannon entropy for the duration of all calls of a particular call type (ie, incoming, outgoing, or missed). All the extracted features were mean-aggregated over the period of 1 week to match the ground truth.</p>
          </sec>
          <sec>
            <title>Location Sensor Features</title>
            <p>Location sensor features provide a contextual idea of the amount of movement users of the sensors go through and show the correlation to mental health [<xref ref-type="bibr" rid="ref3">3</xref>,<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref49">49</xref>,<xref ref-type="bibr" rid="ref50">50</xref>]. The location data are collected through the phones’ GPS or the cellular towers around the phones. Location has been proven to be able to predict depressive states [<xref ref-type="bibr" rid="ref3">3</xref>]. The features extracted from the location sensors included the location variance calculated through the sum of variance in longitude and latitude coordinates, the log of the location variance, the total distance covered, and the circadian movement [<xref ref-type="bibr" rid="ref17">17</xref>] calculated using the Lomb-Scargle method that maps a person’s location patterns following the 24-hour circadian cycle. The speed was also captured as a feature, and static labeled samples were clustered and K-means clustering was used to locate significant places visited by the participants. In addition, location entropy was also engineered to provide the proportion of time spent at each significant location visited during a day.</p>
          </sec>
          <sec>
            <title>Screen Sensor Features</title>
            <p>Screen sensor features are a strong indicator of how engaged users are with their phones. To capture this information, we extracted features that includes the minimum, maximum, sum, and mean (SD) of unlock episodes, along with the number of all unlock episodes and minutes until the first unlock episode. These features have been used in prior studies that proved their correlation to depressive symptom severity [<xref ref-type="bibr" rid="ref46">46</xref>,<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref55">55</xref>].</p>
          </sec>
          <sec>
            <title>Conversation Sensor Features</title>
            <p>Conversation is yet another interesting set of features that provide information pertaining to social interactions and has been used in a number of studies relating to mental health [<xref ref-type="bibr" rid="ref55">55</xref>-<xref ref-type="bibr" rid="ref58">58</xref>]. The computed features included the minimum, maximum, sum, and mean (SD) of the duration of all conversations. We also recorded the minutes of voice, silence, and noise. The energy associated with noise, which is the L2-norm and the sum of all energy values when noise or voice, was inferred.</p>
          </sec>
          <sec>
            <title>Fitbit</title>
            <p>Fitbit offers 2 features, which we extracted based on their application in previous studies relating to mental health [<xref ref-type="bibr" rid="ref54">54</xref>,<xref ref-type="bibr" rid="ref60">60</xref>,<xref ref-type="bibr" rid="ref61">61</xref>], and included the maximum resting heart rate (average maximum heart rate over 1 week) and the maximum number of steps (average step count over 1 week). These features provided an idea of the physical movement and stress experienced by participants.</p>
          </sec>
          <sec>
            <title>Wi-Fi</title>
            <p>Wi-Fi can be a good indicator of social context. We extracted the Wi-Fi count scans that told us the number of scanned Wi-Fi access points connected to by the phone during a time segment and the number of unique connected devices during a time segment based on the hardware address. In addition, we extracted the most scanned connected device. The use of Wi-Fi-based features in mental health prediction have been previously covered [<xref ref-type="bibr" rid="ref48">48</xref>,<xref ref-type="bibr" rid="ref59">59</xref>].</p>
            <p>The data set used in our experiments had all the features discussed, which were part of the 61 features. Feature engineering helped provide a context to the data gathered from all the smartphones and Fitbit sensors and form predictions for ML models.</p>
          </sec>
        </sec>
      </sec>
      <sec>
        <title>Results of Experiment 1</title>
        <p>Experiment 1 showcased the overall performance of the FLMS in comparison with traditional user-agnostic and personalized models. The FLMS achieved a mean accuracy of 0.66 (SD 0.53) and a mean recall of 0.59 (SD 0.50), which are 7% and 13% higher than the best baseline performance achieved by ATU modeling. Among the singular modeling approaches, the ATU, a personalized strategy, performed best overall, with a mean accuracy of 0.59 (SD 0.50) and a mean recall of 0.46 (SD 0.66). The worst performances were shown by user-agnostic LOPO and LTXO approaches, both of which had a mean accuracy of 0.45 (SD 0.80) and 0.47 (SD 0.83), respectively. These results are presented in <xref ref-type="table" rid="table2">Table 2</xref> and show that singular modeling approaches used in different studies [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref17">17</xref>] underperform when modeling involves small, noisy, multimodal sensor data in comparison to our FLMS. The FLMS uses a balance of these strategies to improve predictions.</p>
        <p>Experiment 1 was also designed to show how the FLMS suggests the best feature combinations for the various modeling strategies it uses through the utility of hamming distance from stage 3. The lowest hamming distance in stage 3 for the various modeling approaches used is presented in <xref ref-type="table" rid="table3">Table 3</xref>. We observed that the ATU approach led to the lowest hamming distance of 226, followed by LOTPO, with a minimum hamming distance of 267. The highest hamming distances were those of LOPO at 350 and LTXO at 378. The lower the hamming distance, the closer the predictions to ground truth. Based on this, we saw that overall, 6-sensor fusion works best for this data set. The metadata of experiment 1 are shown in <xref ref-type="table" rid="table4">Table 4</xref>.</p>
        <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Experiment 1 performance of the FLMS<sup>a</sup> in comparison to singular modeling strategies.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <thead>
              <tr valign="top">
                <td>Modeling strategy</td>
                <td>Type of modeling strategy</td>
                <td>Test accuracy, mean (SD)</td>
                <td>Test recall, mean (SD)</td>
                <td>Test <italic>F</italic><sub>1</sub>-score, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>FLMS</td>
                <td>User agnostic + personalized</td>
                <td>0.66 (0.53)</td>
                <td>0.59 (0.50)</td>
                <td>0.56 (0.55)</td>
              </tr>
              <tr valign="top">
                <td>ATU<sup>b</sup></td>
                <td>Personalized</td>
                <td>0.59 (0.60)</td>
                <td>0.46 (0.66)</td>
                <td>0.50 (0.57)</td>
              </tr>
              <tr valign="top">
                <td>LOTPO<sup>c</sup></td>
                <td>Personalized</td>
                <td>0.53 (0.65)</td>
                <td>0.45 (0.70)</td>
                <td>0.32 (0.73)</td>
              </tr>
              <tr valign="top">
                <td>LOPO<sup>d</sup></td>
                <td>User agnostic</td>
                <td>0.45 (0.80)</td>
                <td>0.43 (0.72)</td>
                <td>0.40 (0.87)</td>
              </tr>
              <tr valign="top">
                <td>LTXO<sup>e</sup></td>
                <td>User agnostic</td>
                <td>0.47 (0.83)</td>
                <td>0.35 (0.81)</td>
                <td>0.33 (0.86)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table2fn2">
              <p><sup>b</sup>ATU: accumulated time unit.</p>
            </fn>
            <fn id="table2fn3">
              <p><sup>c</sup>LOTPO: leave one time unit one participant out.</p>
            </fn>
            <fn id="table2fn4">
              <p><sup>d</sup>LOPO: leave one participant out.</p>
            </fn>
            <fn id="table2fn5">
              <p><sup>e</sup>LTXO: leave time unit X out.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table3">
          <label>Table 3</label>
          <caption>
            <p>Experiment 1 minimum hamming distance for choosing the best sensor combination for the experiment.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="520"/>
            <col width="320"/>
            <col width="160"/>
            <thead>
              <tr valign="top">
                <td>Best sensor fusion</td>
                <td>Modeling approach in the FLMS<sup>a</sup></td>
                <td>Hamming distance</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>6-sensor fusion (calls + location + screen usage + conversation + Fitbit + Wi-Fi)</td>
                <td>ATU<sup>b</sup></td>
                <td>226</td>
              </tr>
              <tr valign="top">
                <td>6-sensor fusion (calls + location + screen usage + conversation + Fitbit + Wi-Fi)</td>
                <td>LOTPO<sup>c</sup></td>
                <td>267</td>
              </tr>
              <tr valign="top">
                <td>1-sensor fusion (location)</td>
                <td>LOPO<sup>d</sup></td>
                <td>350</td>
              </tr>
              <tr valign="top">
                <td>2-sensor fusion (calls + location)</td>
                <td>LTXO<sup>e</sup></td>
                <td>378</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table3fn1">
              <p><sup>a</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table3fn2">
              <p><sup>b</sup>ATU: accumulated time unit.</p>
            </fn>
            <fn id="table3fn3">
              <p><sup>c</sup>LOTPO: leave one time unit one participant out.</p>
            </fn>
            <fn id="table3fn4">
              <p><sup>d</sup>LOPO: leave one participant out.</p>
            </fn>
            <fn id="table3fn5">
              <p><sup>e</sup>LTXO: leave time unit X out.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table4">
          <label>Table 4</label>
          <caption>
            <p>Experiment 1 metadata.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="290"/>
            <col width="710"/>
            <thead>
              <tr valign="top">
                <td>Metadata</td>
                <td>Experiment 1</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Feature preprocessing</td>
                <td>KNN<sup>a</sup> imputation, dropping highly co-related columns, sklearn StandardScaler</td>
              </tr>
              <tr valign="top">
                <td>Modeling CV<sup>b</sup> strategy</td>
                <td>FLMS<sup>c</sup>, ATU<sup>d</sup>, LOTPO<sup>e</sup>, LTXO<sup>f</sup>, LOPO<sup>g</sup></td>
              </tr>
              <tr valign="top">
                <td>ML<sup>h</sup> algorithms used</td>
                <td>import XGBoost<sup>i</sup> as xgb<break/>sklearn.linear_model import LinearRegression<break/>sklearn.ensemble import RandomForestRegressor<break/>sklearn.linear_model import ElasticNet<break/>sklearn.ensemble import GradientBoostingRegressor<break/>sklearn.ensemble import ExtraTreesRegressor<break/>sklearn.ensemble import AdaBoostRegressor</td>
              </tr>
              <tr valign="top">
                <td>Random state</td>
                <td>42</td>
              </tr>
              <tr valign="top">
                <td>Evaluation metrics</td>
                <td>Accuracy, recall, <italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table4fn1">
              <p><sup>a</sup>KNN: K-nearest neighbor.</p>
            </fn>
            <fn id="table4fn2">
              <p><sup>b</sup>CV: cross-validation.</p>
            </fn>
            <fn id="table4fn3">
              <p><sup>c</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table4fn4">
              <p><sup>d</sup>ATU: accumulated time unit.</p>
            </fn>
            <fn id="table4fn5">
              <p><sup>e</sup>LOTPO: leave one time unit one participant out.</p>
            </fn>
            <fn id="table4fn6">
              <p><sup>f</sup>LTXO: leave time unit X out.</p>
            </fn>
            <fn id="table4fn7">
              <p><sup>g</sup>LOPO: leave one participant out.</p>
            </fn>
            <fn id="table4fn8">
              <p><sup>h</sup>ML: machine learning.</p>
            </fn>
            <fn id="table4fn9">
              <p><sup>i</sup>XGBoost: Extreme Gradient Boosting.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Results of Experiment 2</title>
        <p>In experiment 2, we compared FLMS ranking results with ML algorithms that have been used in multiple studies on sensor-based assessment of mental health, as listed in <xref ref-type="table" rid="table1">Table 1</xref>. The ML algorithms XGBoost and KNN were chosen based on the popularity of their usage in the community, while the DL algorithm was chosen to be a basic multilayer perceptron (MLP) network and a long short-term memory (LSTM) network. These were also the best-performing algorithms compared to other ML algorithms in the literature on our data set. We initially tried using K-fold validation for the SOTA algorithms, but due to poor results, we switched to the leave-one-out strategy, which performed relatively better. This experiment first compared the overall performance of the FLMS with other SOTA algorithms based on the average test accuracy, recall, and <italic>F</italic><sub>1</sub>-score. Second, the experiment substantiated the claim that the FLMS is better in tackling overfitting, as shown by the mean training accuracy versus the mean test accuracy compared to the ML algorithms in <xref rid="figure11" ref-type="fig">Figure 11</xref>. The models with only the single ML algorithm performed no better than the majority baseline approach, with XGBoost showing a mean test accuracy 0.50 (SD 0.55) and the KNN showing around the same mean accuracy of 0.52 (SD 0.54), as shown in <xref ref-type="table" rid="table5">Table 5</xref>. The MLP achieved higher accuracy but a low test <italic>F</italic><sub>1</sub>-score, indicating the model’s performance has high false-positive and false-negative rates. The LSTM was no different and showed a similar recall and <italic>F</italic><sub>1</sub>-score outcomes. The overfitting of the SOTA models is illustrated in <xref rid="figure11" ref-type="fig">Figure 11</xref>, where we compared the FLMS and the rest of the algorithms based on their respective performances using training and test accuracies. <xref rid="figure11" ref-type="fig">Figure 11</xref> shows that the FLMS had a relatively consistent performance between a training accuracy of 68% and a test accuracy of 66%, while XGBoost, KNN, MLP, and LSTM models had high training accuracies but low test accuracies. The metadata of experiment 2 are shown in <xref ref-type="table" rid="table6">Table 6</xref>.</p>
        <p>The experiments demonstrated support for the points highlighted in the contribution of this paper—that our ranking framework works well with small data sets in comparison to existing approaches and can reduce overfitting by using a balance-weighted ensembling of user-agnostic and personalized models.</p>
        <fig id="figure11" position="float">
          <label>Figure 11</label>
          <caption>
            <p>Experiment 2 shows FLMS training and test accuracies in comparison to SOTA models. The FLMS is better at adapting to overfitting compared to the other algorithms. FLMS: framework for longitudinal multimodal sensors; KNN: K-nearest neighbor; LSTM: long short-term memory; ML: machine learning; MLP: multilayer perceptron; SOTA: state of the art; XGBoost: Extreme Gradient Boosting.</p>
          </caption>
          <graphic xlink:href="ai_v3i1e47805_fig11.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <table-wrap position="float" id="table5">
          <label>Table 5</label>
          <caption>
            <p>Experiment 2 performance of the FLMS<sup>a</sup> compared to ML<sup>b</sup> and DL<sup>c</sup> algorithms used in the current literature on adolescent data.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <col width="200"/>
            <thead>
              <tr valign="top">
                <td>Predictive learning approach</td>
                <td>Modeling strategy</td>
                <td>Test accuracy, mean (SD)</td>
                <td>Test recall, mean (SD)</td>
                <td>Test <italic>F</italic><sub>1</sub>-score, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>FLMS</td>
                <td>ATU<sup>d</sup> + LOTPO<sup>e</sup> + LOPO<sup>f</sup> + LTXO<sup>g</sup></td>
                <td>0.66 (0.53)</td>
                <td>0.59 (0.50)</td>
                <td>0.56 (0.55)</td>
              </tr>
              <tr valign="top">
                <td>XGBoost<sup>h</sup> [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.50 (0.55)</td>
                <td>0.33 (0.52)</td>
                <td>0.28 (0.57)</td>
              </tr>
              <tr valign="top">
                <td>KNN<sup>i</sup> [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.52 (0.54)</td>
                <td>0.40 (0.61)</td>
                <td>0.30 (0.73)</td>
              </tr>
              <tr valign="top">
                <td>MLP<sup>j</sup> [<xref ref-type="bibr" rid="ref9">9</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.55 (0.70)</td>
                <td>0.50 (0.71)</td>
                <td>0.33 (0.70)</td>
              </tr>
              <tr valign="top">
                <td>LSTM<sup>k</sup> [<xref ref-type="bibr" rid="ref67">67</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.41 (0.66)</td>
                <td>0.25 (0.70)</td>
                <td>0.35 (0.70)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table5fn1">
              <p><sup>a</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table5fn2">
              <p><sup>b</sup>ML: machine learning.</p>
            </fn>
            <fn id="table5fn3">
              <p><sup>c</sup>DL: deep learning.</p>
            </fn>
            <fn id="table5fn4">
              <p><sup>d</sup>ATU: accumulated time unit.</p>
            </fn>
            <fn id="table5fn5">
              <p><sup>e</sup>LOTPO: leave one time unit one participant out.</p>
            </fn>
            <fn id="table5fn6">
              <p><sup>f</sup>LOPO: leave one participant out.</p>
            </fn>
            <fn id="table5fn7">
              <p><sup>g</sup>LTXO: leave time unit X out.</p>
            </fn>
            <fn id="table5fn8">
              <p><sup>h</sup>XGBoost: Extreme Gradient Boosting.</p>
            </fn>
            <fn id="table5fn9">
              <p><sup>i</sup>KNN: K-nearest neighbor.</p>
            </fn>
            <fn id="table5fn10">
              <p><sup>j</sup>MLP: multilayer perceptron.</p>
            </fn>
            <fn id="table5fn11">
              <p><sup>k</sup>LSTM: long short-term memory.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table6">
          <label>Table 6</label>
          <caption>
            <p>Experiment 2 metadata.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="340"/>
            <col width="660"/>
            <thead>
              <tr valign="top">
                <td>Metadata</td>
                <td>Experiment 2</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Feature preprocessing</td>
                <td>KNN<sup>a</sup> imputation, dropping highly co-related columns, sklearn StandardScaler</td>
              </tr>
              <tr valign="top">
                <td>Modeling CV<sup>b</sup> strategy</td>
                <td>FLMS<sup>c</sup>, leave 1 out</td>
              </tr>
              <tr valign="top">
                <td>ML<sup>d</sup> algorithms used</td>
                <td>import XGBoost<sup>e</sup> as xgb<break/>sklearn.neural_network import MLPClassifier<break/>sklearn.neighbors import KNeighborsClassifier<break/>keras.layers import LSTM<sup>f</sup></td>
              </tr>
              <tr valign="top">
                <td>Random state</td>
                <td>42</td>
              </tr>
              <tr valign="top">
                <td>Evaluation metrics</td>
                <td>Accuracy, recall, <italic>F</italic><sub>1</sub>-Score</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table6fn1">
              <p><sup>a</sup>KNN: K-nearest neighbor.</p>
            </fn>
            <fn id="table6fn2">
              <p><sup>b</sup>CV: cross-validation.</p>
            </fn>
            <fn id="table6fn3">
              <p><sup>c</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table6fn4">
              <p><sup>d</sup>ML: machine learning.</p>
            </fn>
            <fn id="table6fn5">
              <p><sup>e</sup>XGBoost: Extreme Gradient Boosting.</p>
            </fn>
            <fn id="table6fn6">
              <p><sup>f</sup>LSTM: long short-term memory.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Results of Experiment 3</title>
        <p>Sparsity is a challenge in dealing with small data sets. The large number of 0s or missing values can misdirect models and lead to overfitting [<xref ref-type="bibr" rid="ref68">68</xref>]. Therefore, it is important to handle the problem of sparsity. Our experiment was designed specifically for small data sets, where sparsity proves to be a challenge. To tackle sparsity in small-data scenarios, the commonly used ML algorithms are KNN, MLP, support vector machine (SVM), decision tree (DT), random forest (RF), XGBoost, and AdaBoost [<xref ref-type="bibr" rid="ref21">21</xref>-<xref ref-type="bibr" rid="ref24">24</xref>,<xref ref-type="bibr" rid="ref69">69</xref>-<xref ref-type="bibr" rid="ref71">71</xref>].</p>
        <p>In our experiment, we showcased a comparison of the FLMS with all the mentioned ML algorithms. We first calculated the sparsity of the adolescent data set that comprised all 6-sensor feature sets. The reason for continuing to use the 6-sensor feature sets as in the prior experiment was to test the algorithms with a data set that had a higher degree of sparsity compared to other feature combinations with lower number of sensors. The sparsity for this data set was calculated as the ratio of 0s to the total number of elements in the data set and is given as follows:</p>
        <disp-formula>
          <graphic xlink:href="ai_v3i1e47805_fig19.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </disp-formula>
        <p>
          <bold>(5)</bold>
        </p>
        <p>The sparsity of the data set used for this experiment was 35%. In a small data set, this is a significant amount of sparsity to negatively impact ML algorithms.</p>
        <p>We performed the modeling and evaluated the performance based on <italic>F</italic><sub>1</sub>-scores as in the case of the prediction of mental health, the <italic>F</italic><sub>1</sub>-score is a good reflection of how sparsity affects the models’ judgment in detecting positive and false cases. The models already shown in <xref ref-type="table" rid="table4">Table 4</xref> remained, in addition to other models that have been mentioned in the literature to perform well on sparse data sets. Among the ML algorithms used in the literature, the best performance was shown by the RF, with an <italic>F</italic><sub>1</sub>-score of 0.35, while the FLMS showed an <italic>F</italic><sub>1</sub>-score 0.21 higher than that of the RF. Both MLP and AdaBoost performed close to the RF, with an <italic>F</italic><sub>1</sub>-score of 0.33. The algorithm that performed the worst in handling sparsity was the SVM, with an <italic>F</italic><sub>1</sub>-score of only 0.15. This experiment highlights the fact that due to the combination of modeling, the FLMS performs better when dealing with highly sparse small data sets (<xref ref-type="table" rid="table7">Table 7</xref>). The metadata of experiment 3 are shown in <xref ref-type="table" rid="table8">Table 8</xref>.</p>
        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Experiment 3 performance of the FLMS<sup>a</sup> compared to common ML<sup>b</sup> algorithms for tackling sparsity on the adolescent data set.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="370"/>
            <col width="360"/>
            <col width="270"/>
            <thead>
              <tr valign="top">
                <td>Predictive learning approach</td>
                <td>Modeling strategy</td>
                <td>Test <italic>F</italic><sub>1</sub>-score, mean (SD)</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>FLMS</td>
                <td>ATU<sup>c</sup> + LOTPO<sup>d</sup> + LOPO<sup>e</sup> + LTXO<sup>f</sup></td>
                <td>0.56 (0.55)</td>
              </tr>
              <tr valign="top">
                <td>XGBoost<sup>g</sup> [<xref ref-type="bibr" rid="ref14">14</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.28 (0.57)</td>
              </tr>
              <tr valign="top">
                <td>KNN<sup>h</sup> [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref13">13</xref>,<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.30 (0.73)</td>
              </tr>
              <tr valign="top">
                <td>MLP<sup>i</sup> [<xref ref-type="bibr" rid="ref9">9</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.33 (0.70)</td>
              </tr>
              <tr valign="top">
                <td>SVM<sup>j</sup> [<xref ref-type="bibr" rid="ref12">12</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.15 (0.62)</td>
              </tr>
              <tr valign="top">
                <td>DT<sup>k</sup> [<xref ref-type="bibr" rid="ref13">13</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.24 (0.70)</td>
              </tr>
              <tr valign="top">
                <td>RF<sup>l</sup> [<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref13">13</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.35 (0.65)</td>
              </tr>
              <tr valign="top">
                <td>AdaBoost<sup>m</sup> [<xref ref-type="bibr" rid="ref14">14</xref>]</td>
                <td>Leave 1 out</td>
                <td>0.33 (0.60)</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table7fn1">
              <p><sup>a</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table7fn2">
              <p><sup>b</sup>ML: machine learning.</p>
            </fn>
            <fn id="table7fn3">
              <p><sup>c</sup>ATU: accumulated time unit.</p>
            </fn>
            <fn id="table7fn4">
              <p><sup>d</sup>LOTPO: leave one time unit one participant out.</p>
            </fn>
            <fn id="table7fn5">
              <p><sup>e</sup>LOPO: leave one participant out.</p>
            </fn>
            <fn id="table7fn6">
              <p><sup>f</sup>LTXO: leave time unit X out.</p>
            </fn>
            <fn id="table7fn7">
              <p><sup>g</sup>XGBoost: Extreme Gradient Boosting.</p>
            </fn>
            <fn id="table7fn8">
              <p><sup>h</sup>KNN: K-nearest neighbor.</p>
            </fn>
            <fn id="table7fn9">
              <p><sup>i</sup>MLP: multilayer perceptron.</p>
            </fn>
            <fn id="table7fn10">
              <p><sup>j</sup>SVM: support vector machine.</p>
            </fn>
            <fn id="table7fn11">
              <p><sup>k</sup>DT: decision tree.</p>
            </fn>
            <fn id="table7fn12">
              <p><sup>l</sup>RF: random forest.</p>
            </fn>
            <fn id="table7fn13">
              <p><sup>m</sup>AdaBoost: Adaptive Boosting.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Experiment 3 metadata.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="320"/>
            <col width="680"/>
            <thead>
              <tr valign="top">
                <td>Metadata</td>
                <td>Experiment 3</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Feature preprocessing</td>
                <td>KNN<sup>a</sup> imputation, dropping highly corelated columns, sklearn StandardScaler</td>
              </tr>
              <tr valign="top">
                <td>Modeling CV<sup>b</sup> strategy</td>
                <td>FLMS<sup>c</sup>, leave 1 out</td>
              </tr>
              <tr valign="top">
                <td>ML<sup>d</sup> algorithms used</td>
                <td>import XGBoost<sup>e</sup> as xgb<break/>from sklearn.svm import SVM<sup>f</sup><break/>sklearn.neural_network import MLPClassifier<break/>sklearn.neighbors import KNeighborsClassifier<break/>sklearn.tree import DecisionTreeClassifier<break/>sklearn.ensemble import RandomForestClassifier<break/>sklearn.ensemble import AdaBoostClassifier</td>
              </tr>
              <tr valign="top">
                <td>Random state</td>
                <td>42</td>
              </tr>
              <tr valign="top">
                <td>Evaluation metrics</td>
                <td><italic>F</italic><sub>1</sub>-score</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table8fn1">
              <p><sup>a</sup>KNN: K-nearest neighbor.</p>
            </fn>
            <fn id="table8fn2">
              <p><sup>b</sup>CV: cross-validation.</p>
            </fn>
            <fn id="table8fn3">
              <p><sup>c</sup>FLMS: framework for longitudinal multimodal sensors.</p>
            </fn>
            <fn id="table8fn4">
              <p><sup>d</sup>ML: machine learning.</p>
            </fn>
            <fn id="table8fn5">
              <p><sup>e</sup>XGBoost: Extreme Gradient Boosting.</p>
            </fn>
            <fn id="table8fn6">
              <p><sup>f</sup>SVM: support vector machine.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <sec>
        <title>Principal Findings</title>
        <p>Solving the problem of limited and sparse data sets is not a singular modeling-based endeavor. It requires flexibility and a combination of strategies to achieve predictions that can be trusted. In this section, we discuss our ranking framework’s overarching aims, performance, and limitations based on our assessments.</p>
        <p>In experiment 1, we tested the FLMS in comparison to baseline user-agnostic and personalized models. Our framework achieved a higher accuracy, recall, and <italic>F</italic><sub>1</sub>-score for the predictions when compared to singular modeling approaches, as seen in <xref ref-type="table" rid="table2">Table 2</xref>. We also demonstrated how we arrived at the sensor combination for the best set of predictions using hamming distances in stage 3 of the FLMS, as reflected in <xref ref-type="table" rid="table3">Table 3</xref>. In experiment 2, we compared the FLMS with SOTA algorithms used in the literature for predicting mental health states using sensors. The results from this experiment showed the FLMS to perform better than the existing algorithms in terms of accuracy, recall, and <italic>F</italic><sub>1</sub>-scores (<xref ref-type="table" rid="table4">Table 4</xref>). Experiment 2 also highlighted the FLMS’s ability to reduce overfitting in comparison to the SOTA algorithms. The FLMS showed that the training accuracy and test accuracy did not diverge by large margins, indicating it had not been overfitting the models. Lastly, we compared the FLMS ranking with that of existing ML algorithms that perform well with sparse data in experiment 3. We saw that the data set we used in our experiments exhibited 35% sparsity, which is a significant amount in an already small data set. The FLMS had a higher <italic>F</italic><sub>1</sub>-score compared to the rest of the ML algorithms.</p>
      </sec>
      <sec>
        <title>Comparison With Previous Research</title>
        <p>The results of baseline modeling are consistent with previous studies [<xref ref-type="bibr" rid="ref10">10</xref>,<xref ref-type="bibr" rid="ref29">29</xref>] that showed superior performance when models were personalized. The increase in accuracy shows that our framework was able to narrow down the best set of predictions overall.</p>
        <p>Hamming distance results showed that in LOPO and LTXO approaches, single-sensor deployment and a dual-sensor combination perform equally well as 6-sensor combinations and achieve a minimum hamming distance. This brings forth the advantage of our framework to prioritize sensor selection for yielding best predictions overall and for only the necessary number of feature sets.</p>
        <p>The results of experiment 2 provide us with further evidence of the ranking frameworks’ efficacy in balancing reliance between both user-agnostic and personalized approaches. Despite a higher accuracy, the recall of the FLMS does not overfit like that of other SOTA ML algorithms. The FLMS uses weights to balance out such effects, thus reducing the impact of overfitting in prediction performance. The test with popular existing ML algorithms showed that, despite the success of the models in previous studies [<xref ref-type="bibr" rid="ref9">9</xref>-<xref ref-type="bibr" rid="ref11">11</xref>,<xref ref-type="bibr" rid="ref13">13</xref>-<xref ref-type="bibr" rid="ref17">17</xref>], they struggle when the data set is small and noisy, as is the case of the depression data set presented in this work. This performance result is similar when we look at the capability of ML algorithms that are better at handling sparsity. We found the FLMS to perform better than those algorithms.</p>
        <p>Overall, seeking a single user-agnostic model that fits all is an elusive problem as most existing works suggest better performance for specialized approaches. However, specialized modeling does not perform well on heterogeneous data sets. Therefore, neither user-agnostic nor personalized modeling alone can be applicable to a specific problem area. Our framework provides a practical way to balance the 2 approaches, particularly for dealing with limited data sets.</p>
      </sec>
      <sec>
        <title>Limitations and Future Directions</title>
        <p>We encountered a few limitations with this study that can be addressed in future work. The FLMS was tested on the case of depression in adolescents. As such, we have not been able to establish a lower bound on the data set size that our framework is capable of handling.</p>
        <p>Another area that we could not elaborate on is the computing speed of such a framework that might be impacted if sensor numbers rise to higher levels. Lastly, the framework was equipped with lightweight and widely used ML algorithms. Methods such as the generalized linear mixed model (GLMM) for handling longitudinal data could not be tested.</p>
        <p>Future work can address these limitations with exposure of the framework to more multimodal, longitudinal data sets and adapting and testing other ML algorithms. Interesting future directions for the framework include its online adaptation and a similarity-based cold-start solution.</p>
      </sec>
      <sec>
        <title>Conclusion</title>
        <p>In this study, we presented a novel prediction-ranking framework for modeling limited noisy or sparse, multimodal, longitudinal passive sensor data. We tested our framework on an adolescent depression data set consisting of 45 participants over a period of 24 weeks. The results showed that despite the complexity and limitations of the data set, our framework is able to provide better predictions compared to singular modeling approaches. In experiment 1, our model achieved a 7% increase in accuracy and a 13% increase in recall. In experiment 2 with synthetic data, our model achieved a 5% increase in accuracy and avoided overestimating the recall value through ensembling predictions. The framework also showed its ability to explore sensor combinations through feature fusion. Our tests with existing popular SOTA algorithms showed that the models struggle when data tend to be limited and noisy. We also tested the FLMS with algorithms that perform well with sparsity and found the FLMS to exhibit a better performance. In conclusion, the FLMS can be an effective tool for passive sensing studies.</p>
      </sec>
    </sec>
  </body>
  <back>
    <app-group>
      <supplementary-material id="app1">
        <label>Multimedia Appendix 1</label>
        <p>FLMS ranking overview. Algo: algorithm; ATU: accumulated time unit; avg: average; CV: cross-validation; FLMS: framework for
longitudinal multimodal sensors; LOPO: leave one participant out; LOTPO: leave one time unit of participant out; LTXO: leave time unit X out; ML:
machine learning.</p>
        <media xlink:href="ai_v3i1e47805_app1.png" xlink:title="PNG File , 2313 KB"/>
      </supplementary-material>
    </app-group>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">AdaBoost</term>
          <def>
            <p>Adaptive Boosting</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">API</term>
          <def>
            <p>application programming interface</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">ATU</term>
          <def>
            <p>accumulated time unit</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">CV</term>
          <def>
            <p>cross-validation</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">DL</term>
          <def>
            <p>deep learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">DT</term>
          <def>
            <p>decision tree</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb7">FLMS</term>
          <def>
            <p>framework for longitudinal multimodal sensors</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb8">HAR</term>
          <def>
            <p>human activity recognition</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb9">KNN</term>
          <def>
            <p>K-nearest neighbor</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb10">LDA</term>
          <def>
            <p>linear discriminant analysis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb11">LOPO</term>
          <def>
            <p>leave one participant out</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb12">LOTPO</term>
          <def>
            <p>leave one time unit one participant out</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb13">LR</term>
          <def>
            <p>linear regression</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb14">LSTM</term>
          <def>
            <p>long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb15">LTXO</term>
          <def>
            <p>leave time unit X out</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb16">ML</term>
          <def>
            <p>machine learning</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb17">MLP</term>
          <def>
            <p>multilayer perceptron</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb18">PCA</term>
          <def>
            <p>principal component analysis</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb19">PHQ-9</term>
          <def>
            <p>9-item Patient Health Questionnaire</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb20">RF</term>
          <def>
            <p>random forest</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb21">SOTA</term>
          <def>
            <p>state of the art</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb22">SVM</term>
          <def>
            <p>support vector machine</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb23">XGBoost</term>
          <def>
            <p>Extreme Gradient Boosting</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This study was supported by a grant from the National Institute of Mental Health (NIMH)(1R44MH122067); the NIMH-funded “The Center for Enhancing Triage and Utilization for Depression and Emergent Suicidality (ETUDES) in Pediatric Primary Care” (P50MH115838); the Center for Behavioral Health, Media, and Technology; and a career development award (NIMH 1K23MH111922-01A1). Research recruitment was supported by the Clinical and Translational Science Institute at the University of Pittsburgh by the National Institutes of Health Clinical and Translational Science Award (CTSA) program (grant UL1 TR001857).</p>
    </ack>
    <fn-group>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rabbi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Pfammatter</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Spring</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhury</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Automated personalized feedback for physical activity and dietary behavior change with mobile phones: a randomized controlled trial on adults</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2015</year>
          <month>05</month>
          <day>14</day>
          <volume>3</volume>
          <issue>2</issue>
          <fpage>e42</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2015/2/e42/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/mhealth.4160</pub-id>
          <pub-id pub-id-type="medline">25977197</pub-id>
          <pub-id pub-id-type="pii">v3i2e42</pub-id>
          <pub-id pub-id-type="pmcid">PMC4812832</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rabbi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Aung</surname>
              <given-names>MH</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhury</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>MyBehavior: automatic personalized health feedback from user behaviors and preferences using smartphones</article-title>
          <year>2015</year>
          <conf-name>UbiComp '15: 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
          <conf-date>September 9-11, 2015</conf-date>
          <conf-loc>Osaka, Japan</conf-loc>
          <fpage>707</fpage>
          <lpage>718</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/2750858.2805840"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/2750858.2805840</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Canzian</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Musolesi</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Trajectories of depression: unobtrusive monitoring of depressive states by means of smartphone mobility traces analysis</article-title>
          <year>2015</year>
          <conf-name>UbiComp '15: 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
          <conf-date>September 9-11, 2015</conf-date>
          <conf-loc>Osaka, Japan</conf-loc>
          <fpage>1293</fpage>
          <lpage>1304</lpage>
          <pub-id pub-id-type="doi">10.1145/2750858.2805845</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Difrancesco</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Fraccaro</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>van der Veer</surname>
              <given-names>SN</given-names>
            </name>
            <name name-style="western">
              <surname>Alshoumr</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Ainsworth</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bellazzi</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Out-of-home activity recognition from GPS data in schizophrenic patients</article-title>
          <year>2016</year>
          <conf-name>CBMS 2016: IEEE 29th International Symposium on Computer-Based Medical Systems</conf-name>
          <conf-date>June 20-24, 2016</conf-date>
          <conf-loc>Belfast and Dublin, Ireland</conf-loc>
          <fpage>324</fpage>
          <lpage>328</lpage>
          <pub-id pub-id-type="doi">10.1109/cbms.2016.54</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sano</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Phillips</surname>
              <given-names>AJ</given-names>
            </name>
            <name name-style="western">
              <surname>Amy</surname>
              <given-names>ZY</given-names>
            </name>
            <name name-style="western">
              <surname>McHill</surname>
              <given-names>AW</given-names>
            </name>
            <name name-style="western">
              <surname>Taylor</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Jaques</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Czeisler</surname>
              <given-names>CA</given-names>
            </name>
            <name name-style="western">
              <surname>Klerman</surname>
              <given-names>EB</given-names>
            </name>
            <name name-style="western">
              <surname>Picard</surname>
              <given-names>RW</given-names>
            </name>
          </person-group>
          <article-title>Recognizing academic performance, sleep quality, stress level, and mental health using personality traits, wearable sensors and mobile phones</article-title>
          <year>2015</year>
          <conf-name>BSN 2015: 12th IEEE International Conference on Wearable and Implantable Body Sensor Networks</conf-name>
          <conf-date>June 9-12, 2015</conf-date>
          <conf-loc>Cambridge, MA</conf-loc>
          <fpage>1</fpage>
          <lpage>6</lpage>
          <pub-id pub-id-type="doi">10.1109/bsn.2015.7299420</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Murahari</surname>
              <given-names>VS</given-names>
            </name>
            <name name-style="western">
              <surname>Plötz</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>On attention models for human activity recognition</article-title>
          <year>2018</year>
          <conf-name>ISWC '18: 2018 ACM International Symposium on Wearable Computers</conf-name>
          <conf-date>October 8-12, 2018</conf-date>
          <conf-loc>Singapore</conf-loc>
          <fpage>100</fpage>
          <lpage>103</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3267242.3267287"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3267242.3267287</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Allan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Henrik</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Sourav</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Thor</surname>
              <given-names>SP</given-names>
            </name>
            <name name-style="western">
              <surname>Mikkel</surname>
              <given-names>BK</given-names>
            </name>
            <name name-style="western">
              <surname>Anind</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Tobias</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mads</surname>
              <given-names>MJ</given-names>
            </name>
          </person-group>
          <article-title>Smart devices are different: assessing and mitigating mobile sensing heterogeneities for activity recognition</article-title>
          <year>2015</year>
          <conf-name>SenSys '15: 13th ACM Conference on Embedded Networked Sensor Systems</conf-name>
          <conf-date>November 1-4, 2015</conf-date>
          <conf-loc>Seoul, South Korea</conf-loc>
          <fpage>127</fpage>
          <lpage>140</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/2809695.2809718"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/2809695.2809718</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dietterich</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Overfitting and undercomputing in machine learning</article-title>
          <source>ACM Comput Surv</source>
          <year>1995</year>
          <month>09</month>
          <volume>27</volume>
          <issue>3</issue>
          <fpage>326</fpage>
          <lpage>327</lpage>
          <pub-id pub-id-type="doi">10.1145/212094.212114</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kwapisz</surname>
              <given-names>JR</given-names>
            </name>
            <name name-style="western">
              <surname>Weiss</surname>
              <given-names>GM</given-names>
            </name>
            <name name-style="western">
              <surname>Moore</surname>
              <given-names>SA</given-names>
            </name>
          </person-group>
          <article-title>Activity recognition using cell phone accelerometers</article-title>
          <source>SIGKDD Explor Newsl</source>
          <year>2011</year>
          <month>03</month>
          <day>31</day>
          <volume>12</volume>
          <issue>2</issue>
          <fpage>74</fpage>
          <lpage>82</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/1964897.1964918"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/1964897.1964918</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shukla</surname>
              <given-names>PK</given-names>
            </name>
            <name name-style="western">
              <surname>Vijayvargiya</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kumar</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Human activity recognition using accelerometer and gyroscope data from smartphones</article-title>
          <year>2020</year>
          <conf-name>ICONC3: 2020 IEEE International Conference on Emerging Trends in Communication, Control and Computing</conf-name>
          <conf-date>February 21-22, 2020</conf-date>
          <conf-loc>Lakshmangarh, Sikar, India</conf-loc>
          <pub-id pub-id-type="doi">10.1109/iconc345789.2020.9117456</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Shen</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Performance analysis of smartphone-sensor behavior for human activity recognition</article-title>
          <source>IEEE Access</source>
          <year>2017</year>
          <volume>5</volume>
          <fpage>3095</fpage>
          <lpage>3110</lpage>
          <pub-id pub-id-type="doi">10.1109/access.2017.2676168</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Ding</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Guanling</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ding</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>Monitoring sleep and detecting irregular nights through unconstrained smartphone sensing</article-title>
          <year>2015</year>
          <conf-name>2015 IEEE UIC-ATC-ScalCom</conf-name>
          <conf-date>August 10-14, 2015</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>10</fpage>
          <lpage>14</lpage>
          <pub-id pub-id-type="doi">10.1109/uic-atc-scalcom-cbdcom-iop.2015.30</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Montanini</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Sabino</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Spinsante</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Gambio</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Smartphone as unobtrusive sensor for real-time sleep recognition</article-title>
          <year>2018</year>
          <conf-name>2018 IEEE International Conference on Consumer Electronics (ICCE)</conf-name>
          <conf-date>January 12-14, 2018</conf-date>
          <conf-loc>Las Vegas</conf-loc>
          <fpage>12</fpage>
          <lpage>14</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/ICCE.2018.8326220"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/icce.2018.8326220</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Teng</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cheng</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Ji</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Zhou</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>W</given-names>
            </name>
          </person-group>
          <article-title>PDGes: an interpretable detection model for Parkinson’s disease using smartphones</article-title>
          <source>ACM Trans Sen Netw</source>
          <year>2023</year>
          <month>04</month>
          <day>20</day>
          <volume>19</volume>
          <issue>4</issue>
          <fpage>1</fpage>
          <lpage>21</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3585314"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3585314</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Azam</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Shahzadi</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Khalid</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Anwar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Naeem</surname>
              <given-names>U</given-names>
            </name>
          </person-group>
          <article-title>Smartphone based human breath analysis from respiratory sounds</article-title>
          <year>2018</year>
          <conf-name>2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</conf-name>
          <conf-date>July 17-21, 2018</conf-date>
          <conf-loc>Honolulu, HI</conf-loc>
          <fpage>445</fpage>
          <lpage>448</lpage>
          <pub-id pub-id-type="doi">10.1109/embc.2018.8512452</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Grunerbl</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Osmani</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Bahle</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Carrasco</surname>
              <given-names>JC</given-names>
            </name>
            <name name-style="western">
              <surname>Oehler</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mayora</surname>
              <given-names>O</given-names>
            </name>
            <name name-style="western">
              <surname>Haring</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Lukowicz</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Using smart phone mobility traces for the diagnosis of depressive and manic episodes in bipolar patients</article-title>
          <year>2014</year>
          <conf-name>AH '14: 5th Augmented Human International Conference</conf-name>
          <conf-date>March 7-9, 2014</conf-date>
          <conf-loc>Kobe, Japan</conf-loc>
          <fpage>1</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1145/2582051.2582089</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Saeb</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lattie</surname>
              <given-names>EG</given-names>
            </name>
            <name name-style="western">
              <surname>Kording</surname>
              <given-names>KP</given-names>
            </name>
            <name name-style="western">
              <surname>Mohr</surname>
              <given-names>DC</given-names>
            </name>
          </person-group>
          <article-title>Mobile phone detection of semantic location and its relationship to depression and anxiety</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2017</year>
          <month>08</month>
          <day>10</day>
          <volume>5</volume>
          <issue>8</issue>
          <fpage>e112</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2017/8/e112/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/mhealth.7297</pub-id>
          <pub-id pub-id-type="medline">28798010</pub-id>
          <pub-id pub-id-type="pii">v5i8e112</pub-id>
          <pub-id pub-id-type="pmcid">PMC5571235</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Plötz</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>If only we had more data!: sensor-based human activity recognition in challenging scenarios</article-title>
          <year>2023</year>
          <conf-name>2023 IEEE International Conference on Pervasive Computing and Communications Workshops and other Affiliated Events (PerCom Workshops)</conf-name>
          <conf-date>March 13-17, 2023</conf-date>
          <conf-loc>Atlanta, GA</conf-loc>
          <fpage>565</fpage>
          <lpage>570</lpage>
          <pub-id pub-id-type="doi">10.1109/percomworkshops56833.2023.10150267</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Cornet</surname>
              <given-names>VP</given-names>
            </name>
            <name name-style="western">
              <surname>Holden</surname>
              <given-names>RJ</given-names>
            </name>
          </person-group>
          <article-title>Systematic review of smartphone-based passive sensing for health and wellbeing</article-title>
          <source>J Biomed Inform</source>
          <year>2018</year>
          <month>01</month>
          <volume>77</volume>
          <fpage>120</fpage>
          <lpage>132</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(17)30278-2"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2017.12.008</pub-id>
          <pub-id pub-id-type="medline">29248628</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(17)30278-2</pub-id>
          <pub-id pub-id-type="pmcid">PMC5793918</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Mankoff</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>Ak</given-names>
            </name>
          </person-group>
          <article-title>Understanding practices and needs of researchers in human state modeling by passive mobile sensing</article-title>
          <source>CCF Trans Pervasive Comp Interact</source>
          <year>2021</year>
          <month>07</month>
          <day>06</day>
          <volume>3</volume>
          <issue>4</issue>
          <fpage>344</fpage>
          <lpage>366</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1007/s42486-021-00072-4"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s42486-021-00072-4</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Guestrin</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>XGBoost: a scalable tree boosting system</article-title>
          <year>2016</year>
          <conf-name>KDD '16: 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>
          <conf-date>August 13-17, 2016</conf-date>
          <conf-loc>San Francisco, CA</conf-loc>
          <fpage>785</fpage>
          <lpage>794</lpage>
          <pub-id pub-id-type="doi">10.1145/2939672.2939785</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xi</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Xiang</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Ramadge</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Schapire</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Speed and sparsity of regularized boosting</article-title>
          <source>PMLR</source>
          <year>2009</year>
          <volume>5</volume>
          <fpage>615</fpage>
          <lpage>622</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zou</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Hastie</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>Regularization and variable selection via the elastic net</article-title>
          <source>J R Stat Soc B: Stat Methodol</source>
          <year>2005</year>
          <month>4</month>
          <volume>67</volume>
          <issue>2</issue>
          <fpage>301</fpage>
          <lpage>320</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1111/j.14679868.2005.00503.x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1111/j.1467-9868.2005.00503.x</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Geurts</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Ernst</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Wehenkel</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>Extremely randomized trees</article-title>
          <source>Mach Learn</source>
          <year>2006</year>
          <month>3</month>
          <day>2</day>
          <volume>63</volume>
          <issue>1</issue>
          <fpage>3</fpage>
          <lpage>42</lpage>
          <pub-id pub-id-type="doi">10.1007/s10994-006-6226-1</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Muhammad</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Alshehri</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Karray</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Saddik</surname>
              <given-names>AE</given-names>
            </name>
            <name name-style="western">
              <surname>Alsulaiman</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Falk</surname>
              <given-names>TH</given-names>
            </name>
          </person-group>
          <article-title>A comprehensive survey on multimodal medical signals fusion for smart healthcare systems</article-title>
          <source>Inf Fusion</source>
          <year>2021</year>
          <month>12</month>
          <volume>76</volume>
          <fpage>355</fpage>
          <lpage>375</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.inffus.2021.06.007"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.inffus.2021.06.007</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Joshi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Boyd</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Sensor selection via convex optimization</article-title>
          <source>IEEE Trans Signal Process</source>
          <year>2009</year>
          <month>02</month>
          <volume>57</volume>
          <issue>2</issue>
          <fpage>451</fpage>
          <lpage>462</lpage>
          <pub-id pub-id-type="doi">10.1109/TSP.2008.2007095</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Altenbach</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Corroy</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Böcherer</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Mathar</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>Strategies for distributed sensor selection using convex optimization</article-title>
          <year>2012</year>
          <conf-name>2012 IEEE Global Communications Conference (GLOBECOM)</conf-name>
          <conf-date>December 3-7, 2012</conf-date>
          <conf-loc>Anaheim, CA</conf-loc>
          <fpage>2367</fpage>
          <lpage>2372</lpage>
          <pub-id pub-id-type="doi">10.1109/glocom.2012.6503470</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Shorten</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Khoshgoftaar</surname>
              <given-names>TM</given-names>
            </name>
          </person-group>
          <article-title>A survey on image data augmentation for deep learning</article-title>
          <source>J Big Data</source>
          <year>2019</year>
          <month>7</month>
          <day>6</day>
          <volume>6</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>48</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1186/s40537-019-0197-0"/>
          </comment>
          <pub-id pub-id-type="doi">10.1186/s40537-019-0197-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Feng</surname>
              <given-names>SY</given-names>
            </name>
            <name name-style="western">
              <surname>Gangal</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Wei</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Chandar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Vosoughi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Mitamura</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Hovy</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>A survey of data augmentation approaches for NLP</article-title>
          <source>arXiv. Preprint posted online 2021. [doi: 10.48550/arXiv.2105.03075]</source>
          <year>2021</year>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.48550/arXiv.2105.03075"/>
          </comment>
          <pub-id pub-id-type="doi">10.48550/arXiv.2105.03075</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Florez</surname>
              <given-names>AYC</given-names>
            </name>
            <name name-style="western">
              <surname>Scabora</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Amer-Yahia</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Júnior</surname>
              <given-names>JFR</given-names>
            </name>
          </person-group>
          <article-title>Augmentation techniques for sequential clinical data to improve deep learning prediction technique</article-title>
          <year>2020</year>
          <conf-name>2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS)</conf-name>
          <conf-date>July 28-30, 2020</conf-date>
          <conf-loc>Rochester, MN</conf-loc>
          <fpage>597</fpage>
          <lpage>602</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1109/CBMS49503.2020.00118"/>
          </comment>
          <pub-id pub-id-type="doi">10.1109/cbms49503.2020.00118</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Müller</surname>
              <given-names>SR</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>XL</given-names>
            </name>
            <name name-style="western">
              <surname>Peters</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Chaintreau</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Matz</surname>
              <given-names>SC</given-names>
            </name>
          </person-group>
          <article-title>Depression predictions from GPS-based mobility do not generalize well to large demographically heterogeneous samples</article-title>
          <source>Sci Rep</source>
          <year>2021</year>
          <month>07</month>
          <day>07</day>
          <volume>11</volume>
          <issue>1</issue>
          <fpage>14007</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1038/s41598-021-93087-x"/>
          </comment>
          <pub-id pub-id-type="doi">10.1038/s41598-021-93087-x</pub-id>
          <pub-id pub-id-type="medline">34234186</pub-id>
          <pub-id pub-id-type="pii">10.1038/s41598-021-93087-x</pub-id>
          <pub-id pub-id-type="pmcid">PMC8263566</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Xu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Chikersal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Dutcher</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Sefidgar</surname>
              <given-names>YS</given-names>
            </name>
            <name name-style="western">
              <surname>Seo</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>Tumminia</surname>
              <given-names>MJ</given-names>
            </name>
            <name name-style="western">
              <surname>Villalba</surname>
              <given-names>DK</given-names>
            </name>
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>KG</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Doryab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Nurius</surname>
              <given-names>PS</given-names>
            </name>
            <name name-style="western">
              <surname>Riskin</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>AK</given-names>
            </name>
            <name name-style="western">
              <surname>Mankoff</surname>
              <given-names>J</given-names>
            </name>
          </person-group>
          <article-title>Leveraging collaborative-filtering for personalized behavior modeling: a case study of depression detection among college students</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2021</year>
          <month>03</month>
          <day>30</day>
          <volume>5</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>27</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3448107"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3448107</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Maxhuni</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Hernandez-Leal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sucar</surname>
              <given-names>LE</given-names>
            </name>
            <name name-style="western">
              <surname>Osmani</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Morales</surname>
              <given-names>EF</given-names>
            </name>
            <name name-style="western">
              <surname>Mayora</surname>
              <given-names>O</given-names>
            </name>
          </person-group>
          <article-title>Stress modelling and prediction in presence of scarce data</article-title>
          <source>J Biomed Inform</source>
          <year>2016</year>
          <month>10</month>
          <volume>63</volume>
          <fpage>344</fpage>
          <lpage>356</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://linkinghub.elsevier.com/retrieve/pii/S1532-0464(16)30109-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jbi.2016.08.023</pub-id>
          <pub-id pub-id-type="medline">27592309</pub-id>
          <pub-id pub-id-type="pii">S1532-0464(16)30109-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Jacobson</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Lekkas</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Huang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Thomas</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Deep learning paired with wearable passive sensing data predicts deterioration in anxiety disorder symptoms across 17-18 years</article-title>
          <source>J Affect Disord</source>
          <year>2021</year>
          <month>03</month>
          <day>01</day>
          <volume>282</volume>
          <fpage>104</fpage>
          <lpage>111</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/33401123"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.jad.2020.12.086</pub-id>
          <pub-id pub-id-type="medline">33401123</pub-id>
          <pub-id pub-id-type="pii">S0165-0327(20)33176-1</pub-id>
          <pub-id pub-id-type="pmcid">PMC7889722</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ren</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Balkind</surname>
              <given-names>EG</given-names>
            </name>
            <name name-style="western">
              <surname>Pastro</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Israel</surname>
              <given-names>ES</given-names>
            </name>
            <name name-style="western">
              <surname>Pizzagalli</surname>
              <given-names>DA</given-names>
            </name>
            <name name-style="western">
              <surname>Rahimi-Eichi</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Baker</surname>
              <given-names>JT</given-names>
            </name>
            <name name-style="western">
              <surname>Webb</surname>
              <given-names>CA</given-names>
            </name>
          </person-group>
          <article-title>Predicting states of elevated negative affect in adolescents from smartphone sensors: a novel personalized machine learning approach</article-title>
          <source>Psychol Med</source>
          <year>2022</year>
          <month>07</month>
          <day>27</day>
          <volume>53</volume>
          <issue>11</issue>
          <fpage>5146</fpage>
          <lpage>5154</lpage>
          <pub-id pub-id-type="doi">10.1017/s0033291722002161</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Adhikary</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Majumder</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Chatterjee</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Shaw</surname>
              <given-names>RN</given-names>
            </name>
            <name name-style="western">
              <surname>Ghosh</surname>
              <given-names>A</given-names>
            </name>
          </person-group>
          <person-group person-group-type="editor">
            <name name-style="western">
              <surname>Shaw</surname>
              <given-names>RN</given-names>
            </name>
            <name name-style="western">
              <surname>Das</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Piuri</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Bianchini</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Human activity recognition for disease detection using machine learning techniques—a comparative study</article-title>
          <source>Advanced Computing and Intelligent Technologies. Lecture Notes in Electrical Engineering, Vol 914</source>
          <year>2022</year>
          <publisher-loc>Singapore</publisher-loc>
          <publisher-name>Springer</publisher-name>
        </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Messalas</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Kanellopoulos</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Makris</surname>
              <given-names>C</given-names>
            </name>
          </person-group>
          <article-title>Model-agnostic interpretability with Shapley values</article-title>
          <year>2019</year>
          <conf-name>IISA 2019: 10th IEEE International Conference on Information, Intelligence, Systems and Applications</conf-name>
          <conf-date>July 15-17, 2019</conf-date>
          <conf-loc>Patras, Greece</conf-loc>
          <fpage>1</fpage>
          <lpage>7</lpage>
          <pub-id pub-id-type="doi">10.1109/iisa.2019.8900669</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref38">
        <label>38</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Li</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Qiao</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Yu</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>H-Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liao</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Zhu</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>Interpretable tree-based ensemble model for predicting beach water quality</article-title>
          <source>Water Res</source>
          <year>2022</year>
          <month>03</month>
          <day>01</day>
          <volume>211</volume>
          <fpage>118078</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1016/j.watres.2022.118078"/>
          </comment>
          <pub-id pub-id-type="doi">10.1016/j.watres.2022.118078</pub-id>
          <pub-id pub-id-type="medline">35066260</pub-id>
          <pub-id pub-id-type="pii">S0043-1354(22)00041-0</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref39">
        <label>39</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Debie</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Fernandez Rojas</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Fidock</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Barlow</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Kasmarik</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Anavatti</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Garratt</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Abbass</surname>
              <given-names>HA</given-names>
            </name>
          </person-group>
          <article-title>Multimodal fusion for objective assessment of cognitive workload: a review</article-title>
          <source>IEEE Trans Cybern</source>
          <year>2021</year>
          <month>3</month>
          <volume>51</volume>
          <issue>3</issue>
          <fpage>1542</fpage>
          <lpage>1555</lpage>
          <pub-id pub-id-type="doi">10.1109/tcyb.2019.2939399</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref40">
        <label>40</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Pedrelli</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Fedor</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Ghandeharioun</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Howe</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Ionescu</surname>
              <given-names>DF</given-names>
            </name>
            <name name-style="western">
              <surname>Bhathena</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Fisher</surname>
              <given-names>LB</given-names>
            </name>
            <name name-style="western">
              <surname>Cusin</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Nyer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yeung</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Sangermano</surname>
              <given-names>L</given-names>
            </name>
            <name name-style="western">
              <surname>Mischoulon</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Alpert</surname>
              <given-names>JE</given-names>
            </name>
            <name name-style="western">
              <surname>Picard</surname>
              <given-names>RW</given-names>
            </name>
          </person-group>
          <article-title>Monitoring changes in depression severity using wearable and mobile sensors</article-title>
          <source>Front Psychiatry</source>
          <year>2020</year>
          <month>12</month>
          <day>18</day>
          <volume>11</volume>
          <fpage>584711</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/33391050"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fpsyt.2020.584711</pub-id>
          <pub-id pub-id-type="medline">33391050</pub-id>
          <pub-id pub-id-type="pmcid">PMC7775362</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref41">
        <label>41</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>R</given-names>
            </name>
          </person-group>
          <article-title>On predicting relapse in schizophrenia using mobile sensing in a randomized control trial</article-title>
          <year>2020</year>
          <conf-name>2020 IEEE International Conference on Pervasive Computing and Communications (PerCom)</conf-name>
          <conf-date>March 23-27, 2020</conf-date>
          <conf-loc>Austin, TX</conf-loc>
          <fpage>1</fpage>
          <lpage>8</lpage>
          <pub-id pub-id-type="doi">10.1109/percom45495.2020.9127365</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref42">
        <label>42</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Folarin</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Cummins</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Garcia-Dias</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Stewart</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ranjan</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Rashid</surname>
              <given-names>Z</given-names>
            </name>
            <name name-style="western">
              <surname>Conde</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Laiou</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Sankesara</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Matcham</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Leightley</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>White</surname>
              <given-names>KM</given-names>
            </name>
            <name name-style="western">
              <surname>Oetzmann</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Ivan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Lamers</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Siddi</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Simblett</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Nica</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Rintala</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Mohr</surname>
              <given-names>DC</given-names>
            </name>
            <name name-style="western">
              <surname>Myin-Germeys</surname>
              <given-names>I</given-names>
            </name>
            <name name-style="western">
              <surname>Wykes</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Haro</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Penninx</surname>
              <given-names>BWJH</given-names>
            </name>
            <name name-style="western">
              <surname>Vairavan</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Narayan</surname>
              <given-names>VA</given-names>
            </name>
            <name name-style="western">
              <surname>Annas</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Hotopf</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dobson</surname>
              <given-names>RJB</given-names>
            </name>
            <collab>RADAR-CNS Consortium</collab>
          </person-group>
          <article-title>Challenges in using mHealth data from smartphones and wearable devices to predict depression symptom severity: retrospective analysis</article-title>
          <source>J Med Internet Res</source>
          <year>2023</year>
          <month>08</month>
          <day>14</day>
          <volume>25</volume>
          <fpage>e45233</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2023//e45233/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/45233</pub-id>
          <pub-id pub-id-type="medline">37578823</pub-id>
          <pub-id pub-id-type="pii">v25i1e45233</pub-id>
          <pub-id pub-id-type="pmcid">PMC10463088</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref43">
        <label>43</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Tlachac</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Toto</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Lovering</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kayastha</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Taurich</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Rundensteiner</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>EMU: early mental health uncovering framework and dataset</article-title>
          <year>2021</year>
          <conf-name>ICMLA 2021: 20th IEEE International Conference on Machine Learning and Applications</conf-name>
          <conf-date>December 13-16, 2021</conf-date>
          <conf-loc>Pasadena, CA</conf-loc>
          <fpage>1311</fpage>
          <lpage>1318</lpage>
          <pub-id pub-id-type="doi">10.1109/icmla52953.2021.00213</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref44">
        <label>44</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Negeri</surname>
              <given-names>ZF</given-names>
            </name>
            <name name-style="western">
              <surname>Levis</surname>
              <given-names>B</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>He</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Krishnan</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Bhandari</surname>
              <given-names>PM</given-names>
            </name>
            <name name-style="western">
              <surname>Neupane</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Brehaut</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Benedetti</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Thombs</surname>
              <given-names>BD</given-names>
            </name>
            <collab>Depression Screening Data (DEPRESSD) PHQ Group</collab>
          </person-group>
          <article-title>Accuracy of the Patient Health Questionnaire-9 for screening to detect major depression: updated systematic review and individual participant data meta-analysis</article-title>
          <source>BMJ</source>
          <year>2021</year>
          <month>10</month>
          <day>05</day>
          <volume>375</volume>
          <fpage>n2183</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="http://www.bmj.com/lookup/pmidlookup?view=long&#38;pmid=34610915"/>
          </comment>
          <pub-id pub-id-type="doi">10.1136/bmj.n2183</pub-id>
          <pub-id pub-id-type="medline">34610915</pub-id>
          <pub-id pub-id-type="pmcid">PMC8491108</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref45">
        <label>45</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Vega</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Aguillera</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Goel</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Joshi</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Khandekar</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Durica</surname>
              <given-names>KC</given-names>
            </name>
            <name name-style="western">
              <surname>Kunta</surname>
              <given-names>AR</given-names>
            </name>
            <name name-style="western">
              <surname>Low</surname>
              <given-names>CA</given-names>
            </name>
          </person-group>
          <article-title>Reproducible analysis pipeline for data streams: open-source software to process data collected with mobile devices</article-title>
          <source>Front Digit Health</source>
          <year>2021</year>
          <month>11</month>
          <day>18</day>
          <volume>3</volume>
          <fpage>769823</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://europepmc.org/abstract/MED/34870271"/>
          </comment>
          <pub-id pub-id-type="doi">10.3389/fdgth.2021.769823</pub-id>
          <pub-id pub-id-type="medline">34870271</pub-id>
          <pub-id pub-id-type="pmcid">PMC8636712</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref46">
        <label>46</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Saeb</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Karr</surname>
              <given-names>CJ</given-names>
            </name>
            <name name-style="western">
              <surname>Schueller</surname>
              <given-names>SM</given-names>
            </name>
            <name name-style="western">
              <surname>Corden</surname>
              <given-names>ME</given-names>
            </name>
            <name name-style="western">
              <surname>Kording</surname>
              <given-names>KP</given-names>
            </name>
            <name name-style="western">
              <surname>Mohr</surname>
              <given-names>DC</given-names>
            </name>
          </person-group>
          <article-title>Mobile phone sensor correlates of depressive symptom severity in daily-life behavior: an exploratory study</article-title>
          <source>J Med Internet Res</source>
          <year>2015</year>
          <month>07</month>
          <day>15</day>
          <volume>17</volume>
          <issue>7</issue>
          <fpage>e175</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.jmir.org/2015/7/e175/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/jmir.4273</pub-id>
          <pub-id pub-id-type="medline">26180009</pub-id>
          <pub-id pub-id-type="pii">v17i7e175</pub-id>
          <pub-id pub-id-type="pmcid">PMC4526997</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref47">
        <label>47</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>W</given-names>
            </name>
            <name name-style="western">
              <surname>daSilva</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Huckins</surname>
              <given-names>JF</given-names>
            </name>
            <name name-style="western">
              <surname>Kelley</surname>
              <given-names>WM</given-names>
            </name>
            <name name-style="western">
              <surname>Heatherton</surname>
              <given-names>TF</given-names>
            </name>
            <name name-style="western">
              <surname>Campbell</surname>
              <given-names>AT</given-names>
            </name>
          </person-group>
          <article-title>Tracking depression dynamics in college students using mobile phone and wearable sensing</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2018</year>
          <month>03</month>
          <day>26</day>
          <volume>2</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>26</lpage>
          <pub-id pub-id-type="doi">10.1145/3191775</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref48">
        <label>48</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wahle</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Kowatsch</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Fleisch</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Rufer</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Weidt</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Mobile sensing and support for people with depression: a pilot trial in the wild</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2016</year>
          <month>09</month>
          <day>21</day>
          <volume>4</volume>
          <issue>3</issue>
          <fpage>e111,e5960</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2016/3/e111/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/mhealth.5960</pub-id>
          <pub-id pub-id-type="medline">27655245</pub-id>
          <pub-id pub-id-type="pii">v4i3e111</pub-id>
          <pub-id pub-id-type="pmcid">PMC5052463</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref49">
        <label>49</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mehrotra</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Musolesi</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>Using autoencoders to automatically extract mobility features for predicting depressive states</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2018</year>
          <month>09</month>
          <day>18</day>
          <volume>2</volume>
          <issue>3</issue>
          <fpage>1</fpage>
          <lpage>20</lpage>
          <pub-id pub-id-type="doi">10.1145/3264937</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref50">
        <label>50</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Farhan</surname>
              <given-names>AA</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Morillo</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Ware</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Kamath</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Russell</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bamis</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Behavior vs. introspection: refining prediction of clinical depression via smartphone sensing data</article-title>
          <year>2016</year>
          <conf-name>2016 IEEE Wireless Health</conf-name>
          <conf-date>October 25-27, 2016</conf-date>
          <conf-loc>Bethesda, MD</conf-loc>
          <fpage>30</fpage>
          <lpage>37</lpage>
          <pub-id pub-id-type="doi">10.1109/wh.2016.7764553</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref51">
        <label>51</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Chikersal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Doryab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Tumminia</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Villalba</surname>
              <given-names>DK</given-names>
            </name>
            <name name-style="western">
              <surname>Dutcher</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>KG</given-names>
            </name>
            <name name-style="western">
              <surname>Mankoff</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Goel</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>AK</given-names>
            </name>
          </person-group>
          <article-title>Detecting depression and predicting its onset using longitudinal symptoms captured by passive sensing: a machine learning approach with robust feature selection</article-title>
          <source>ACM Trans Comput-Hum Interact</source>
          <year>2021</year>
          <month>01</month>
          <day>20</day>
          <volume>28</volume>
          <issue>1</issue>
          <fpage>1</fpage>
          <lpage>41</lpage>
          <pub-id pub-id-type="doi">10.1145/3422821</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref52">
        <label>52</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lane</surname>
              <given-names>ND</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Mohammod</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>H</given-names>
            </name>
            <name name-style="western">
              <surname>Cardone</surname>
              <given-names>G</given-names>
            </name>
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Doryab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Berke</surname>
              <given-names>E</given-names>
            </name>
            <name name-style="western">
              <surname>Campbell</surname>
              <given-names>AT</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhury</surname>
              <given-names>T</given-names>
            </name>
          </person-group>
          <article-title>BeWell: sensing sleep, physical activities and social interactions to promote wellbeing</article-title>
          <source>Mobile Netw Appl</source>
          <year>2014</year>
          <month>1</month>
          <day>9</day>
          <volume>19</volume>
          <issue>3</issue>
          <fpage>345</fpage>
          <lpage>359</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1007/s11036-013-0484-5"/>
          </comment>
          <pub-id pub-id-type="doi">10.1007/s11036-013-0484-5</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref53">
        <label>53</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>LiKamWa</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Lane</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Zhong</surname>
              <given-names>L</given-names>
            </name>
          </person-group>
          <article-title>MoodScope: building a mood sensor from smartphone usage patterns</article-title>
          <year>2013</year>
          <conf-name>MobiSys'13: 11th Annual International Conference on Mobile Systems, Applications, and Services</conf-name>
          <conf-date>June 25-28, 2013</conf-date>
          <conf-loc>Taipei, Taiwan</conf-loc>
          <fpage>25</fpage>
          <lpage>28</lpage>
          <pub-id pub-id-type="doi">10.1145/2462456.2464449</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref54">
        <label>54</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Doryab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Villalba</surname>
              <given-names>DK</given-names>
            </name>
            <name name-style="western">
              <surname>Chikersal</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Dutcher</surname>
              <given-names>JM</given-names>
            </name>
            <name name-style="western">
              <surname>Tumminia</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Cohen</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Mankoff</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Creswell</surname>
              <given-names>JD</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>AK</given-names>
            </name>
          </person-group>
          <article-title>Identifying behavioral phenotypes of loneliness and social isolation with passive sensing: statistical analysis, data mining and machine learning of smartphone and Fitbit data</article-title>
          <source>JMIR Mhealth Uhealth</source>
          <year>2019</year>
          <month>07</month>
          <day>24</day>
          <volume>7</volume>
          <issue>7</issue>
          <fpage>e13209</fpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://mhealth.jmir.org/2019/7/e13209/"/>
          </comment>
          <pub-id pub-id-type="doi">10.2196/13209</pub-id>
          <pub-id pub-id-type="medline">31342903</pub-id>
          <pub-id pub-id-type="pii">v7i7e13209</pub-id>
          <pub-id pub-id-type="pmcid">PMC6685126</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref55">
        <label>55</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Aung</surname>
              <given-names>MSH</given-names>
            </name>
            <name name-style="western">
              <surname>Abdullah</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>CrossCheck: toward passive sensing and detection of mental health changes in people with schizophrenia</article-title>
          <year>2016</year>
          <conf-name>UbiComp '16: 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
          <conf-date>September 12-16, 2016</conf-date>
          <conf-loc>Heidelberg, Germany</conf-loc>
          <pub-id pub-id-type="doi">10.1145/2971648.2971740</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref56">
        <label>56</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Lane</surname>
              <given-names>N</given-names>
            </name>
            <name name-style="western">
              <surname>Rabbi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Lin</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Yang</surname>
              <given-names>x</given-names>
            </name>
          </person-group>
          <article-title>Bewell: a smartphone application to monitor, model and promote wellbeing</article-title>
          <year>2012</year>
          <conf-name>5th International ICST Conference on Pervasive Computing Technologies for Healthcare</conf-name>
          <conf-date>May 23-26, 2011</conf-date>
          <conf-loc>Dublin, Ireland</conf-loc>
          <pub-id pub-id-type="doi">10.4108/icst.pervasivehealth.2011.246161</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref57">
        <label>57</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Mashfiqui</surname>
              <given-names>r</given-names>
            </name>
            <name name-style="western">
              <surname>Ali</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Choudhury</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Berke</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Passive and in-situ assessment of mental and physical well-being using mobile sensors</article-title>
          <year>2011</year>
          <conf-name>UbiComp '11: 13th International Conference on Ubiquitous Computing</conf-name>
          <conf-date>September 17-21, 2011</conf-date>
          <conf-loc>Beijing, China</conf-loc>
          <fpage>385</fpage>
          <lpage>394</lpage>
          <pub-id pub-id-type="doi">10.1145/2030112.2030164</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref58">
        <label>58</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>F</given-names>
            </name>
            <name name-style="western">
              <surname>Chen</surname>
              <given-names>Z</given-names>
            </name>
          </person-group>
          <article-title>StudentLife: assessing mental health, academic performance and behavioral trends of college students using smartphones</article-title>
          <year>2014</year>
          <conf-name>UbiComp '14: 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing</conf-name>
          <conf-date>September 13-17, 2014</conf-date>
          <conf-loc>Seattle, WA</conf-loc>
          <fpage>3</fpage>
          <lpage>14</lpage>
          <pub-id pub-id-type="doi">10.1145/2632048.2632054</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref59">
        <label>59</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ware</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Yue</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Morillo</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Lu</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Shang</surname>
              <given-names>C</given-names>
            </name>
            <name name-style="western">
              <surname>Kamath</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Bamis</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Bi</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Russell</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Wang</surname>
              <given-names>B</given-names>
            </name>
          </person-group>
          <article-title>Large-scale automatic depression screening using meta-data from WiFi infrastructure</article-title>
          <source>Proc ACM Interact Mob Wearable Ubiquitous Technol</source>
          <year>2018</year>
          <month>12</month>
          <day>27</day>
          <volume>2</volume>
          <issue>4</issue>
          <fpage>1</fpage>
          <lpage>27</lpage>
          <pub-id pub-id-type="doi">10.1145/3287073</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref60">
        <label>60</label>
        <nlm-citation citation-type="confproc">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Dai</surname>
              <given-names>R</given-names>
            </name>
            <name name-style="western">
              <surname>Kannampallil</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Kim</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Detecting mental disorders with wearables: a large cohort study</article-title>
          <year>2023</year>
          <conf-name>IoTDI '23: 8th ACM/IEEE Conference on Internet of Things Design and Implementation</conf-name>
          <conf-date>May 9-12, 2023</conf-date>
          <conf-loc>San Antonio, TX</conf-loc>
          <fpage>39</fpage>
          <lpage>51</lpage>
          <pub-id pub-id-type="doi">10.1145/3576842.3582389</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref61">
        <label>61</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Doryab</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Chikarsel</surname>
              <given-names>P</given-names>
            </name>
            <name name-style="western">
              <surname>Liu</surname>
              <given-names>X</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>AK</given-names>
            </name>
          </person-group>
          <article-title>Extraction of behavioral features from smartphone and wearable data</article-title>
          <source>arXiv. Preprint posted online 2018. [doi: 10.48550/arXiv.1812.10394]</source>
          <year>2021</year>
          <pub-id pub-id-type="doi">10.48550/arXiv.1812.10394</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref62">
        <label>62</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kroenke</surname>
              <given-names>K</given-names>
            </name>
            <name name-style="western">
              <surname>Strine</surname>
              <given-names>TW</given-names>
            </name>
            <name name-style="western">
              <surname>Spitzer</surname>
              <given-names>RL</given-names>
            </name>
            <name name-style="western">
              <surname>Williams</surname>
              <given-names>JB</given-names>
            </name>
            <name name-style="western">
              <surname>Berry</surname>
              <given-names>JT</given-names>
            </name>
            <name name-style="western">
              <surname>Mokdad</surname>
              <given-names>AH</given-names>
            </name>
          </person-group>
          <article-title>The PHQ-8 as a measure of current depression in the general population</article-title>
          <source>J Affect Disord</source>
          <year>2009</year>
          <month>04</month>
          <volume>114</volume>
          <issue>1-3</issue>
          <fpage>163</fpage>
          <lpage>173</lpage>
          <pub-id pub-id-type="doi">10.1016/j.jad.2008.06.026</pub-id>
          <pub-id pub-id-type="medline">18752852</pub-id>
          <pub-id pub-id-type="pii">S0165-0327(08)00282-6</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref63">
        <label>63</label>
        <nlm-citation citation-type="book">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Hamilton</surname>
              <given-names>M</given-names>
            </name>
          </person-group>
          <article-title>The Hamilton Rating Scale for depression</article-title>
          <source>Assessment of Depression</source>
          <year>1986</year>
          <publisher-loc>Berlin, Heidelberg</publisher-loc>
          <publisher-name>Springer</publisher-name>
          <fpage>143</fpage>
          <lpage>152</lpage>
        </nlm-citation>
      </ref>
      <ref id="ref64">
        <label>64</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Thompson</surname>
              <given-names>E</given-names>
            </name>
          </person-group>
          <article-title>Hamilton Rating Scale for anxiety (HAM-A)</article-title>
          <source>Occup Med (Lond)</source>
          <year>2015</year>
          <month>10</month>
          <day>13</day>
          <volume>65</volume>
          <issue>7</issue>
          <fpage>601</fpage>
          <pub-id pub-id-type="doi">10.1093/occmed/kqv054</pub-id>
          <pub-id pub-id-type="medline">26370845</pub-id>
          <pub-id pub-id-type="pii">kqv054</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref65">
        <label>65</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Schelter</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Böse</surname>
              <given-names>J-H</given-names>
            </name>
            <name name-style="western">
              <surname>Kirschnick</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Klein</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Seufert</surname>
              <given-names>S</given-names>
            </name>
          </person-group>
          <article-title>Automatically tracking metadata and provenance of machine learning experiments</article-title>
          <source>Amazon Science</source>
          <year>2017</year>
          <access-date>2024-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://assets.amazon.science/2f/39/4b32cf354e4c993b439d88258597/automatically-tracking-metadata-and-provenance-of-machine-learning-experiments.pdf">https://assets.amazon.science/2f/39/4b32cf354e4c993b439d88258597/automatical ly-tracking-metadata-and-provenance-of-machine-learning-experiments.pdf</ext-link>
          </comment>
        </nlm-citation>
      </ref>
      <ref id="ref66">
        <label>66</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Ferreira</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Kostakos</surname>
              <given-names>V</given-names>
            </name>
            <name name-style="western">
              <surname>Dey</surname>
              <given-names>AK</given-names>
            </name>
          </person-group>
          <article-title>AWARE: mobile context instrumentation framework</article-title>
          <source>Front ICT</source>
          <year>2015</year>
          <month>04</month>
          <day>20</day>
          <volume>2</volume>
          <fpage>6</fpage>
          <pub-id pub-id-type="doi">10.3389/fict.2015.00006</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref67">
        <label>67</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Acikmese</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Alptekin</surname>
              <given-names>SE</given-names>
            </name>
          </person-group>
          <article-title>Prediction of stress levels with LSTM and passive mobile sensors</article-title>
          <source>Procedia Comput Sci</source>
          <year>2019</year>
          <volume>159</volume>
          <fpage>658</fpage>
          <lpage>667</lpage>
          <pub-id pub-id-type="doi">10.1016/j.procs.2019.09.221</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref68">
        <label>68</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Kucukozer-Cavdar</surname>
              <given-names>S</given-names>
            </name>
            <name name-style="western">
              <surname>Taskaya-Temizel</surname>
              <given-names>T</given-names>
            </name>
            <name name-style="western">
              <surname>Mehrotra</surname>
              <given-names>A</given-names>
            </name>
            <name name-style="western">
              <surname>Musolesi</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Tino</surname>
              <given-names>P</given-names>
            </name>
          </person-group>
          <article-title>Designing robust models for behaviour prediction using sparse data from mobile sensing: a case study of office workers’ availability for well-being interventions</article-title>
          <source>ACM Trans Comput Healthc</source>
          <year>2021</year>
          <month>07</month>
          <day>18</day>
          <volume>2</volume>
          <issue>4</issue>
          <fpage>1</fpage>
          <lpage>33</lpage>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://doi.org/10.1145/3458753"/>
          </comment>
          <pub-id pub-id-type="doi">10.1145/3458753</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref69">
        <label>69</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Yin</surname>
              <given-names>D</given-names>
            </name>
            <name name-style="western">
              <surname>Li</surname>
              <given-names>J</given-names>
            </name>
            <name name-style="western">
              <surname>Wu</surname>
              <given-names>G</given-names>
            </name>
          </person-group>
          <article-title>Solving the data sparsity problem in predicting the success of the startups with machine learning methods</article-title>
          <source>arXiv. Preprint posted online 2021. [doi: 10.48550/arXiv.2112.07985]</source>
          <year>2021</year>
          <volume>07985</volume>
          <issue>2021</issue>
          <pub-id pub-id-type="doi">10.48550/arXiv.2112.07985</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref70">
        <label>70</label>
        <nlm-citation citation-type="journal">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Zhang</surname>
              <given-names>M</given-names>
            </name>
            <name name-style="western">
              <surname>Sun</surname>
              <given-names>Y</given-names>
            </name>
            <name name-style="western">
              <surname>Liang</surname>
              <given-names>F</given-names>
            </name>
          </person-group>
          <article-title>Sparse deep learning for time series data: theory and applications</article-title>
          <source>arXiv. Preprint posted online 2023. [doi: 10.48550/arXiv.2310.03243]</source>
          <year>2021</year>
          <pub-id pub-id-type="doi">10.48550/arXiv.2310.03243</pub-id>
        </nlm-citation>
      </ref>
      <ref id="ref71">
        <label>71</label>
        <nlm-citation citation-type="web">
          <person-group person-group-type="author">
            <name name-style="western">
              <surname>Rosidi</surname>
              <given-names>N</given-names>
            </name>
          </person-group>
          <article-title>Best machine learning model for sparse data</article-title>
          <source>KD nuggets</source>
          <year>2023</year>
          <month>4</month>
          <day>7</day>
          <access-date>2024-05-01</access-date>
          <comment>
            <ext-link ext-link-type="uri" xlink:type="simple" xlink:href="https://www.kdnuggets.com/2023/04/best-machine-learning-model-sparse-data.html">https://www.kdnuggets.com/2023/04/best-machine-learning-model-sparse-data.html</ext-link>
          </comment>
        </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
