<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Public Health</journal-id>
<journal-title-group>
<journal-title>Frontiers in Public Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Public Health</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-2565</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpubh.2025.1620031</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Digital health interventions from the humanistic perspective of sports: strategies to promote health for all</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Yang</surname> <given-names>Yan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<uri xlink:href="https://loop.frontiersin.org/people/3050783"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Huang</surname> <given-names>Xianzhong</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Shi</surname> <given-names>Bing</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Physical Education and Research, Shangluo University, Shangluo</institution>, <city>Shaanxi</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Shangluo Smart Sports Service and Research Center, Shangluo</institution>, <city>Shaanxi</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>Physical Education School of Shaanxi Normal University, Xi&#x00027;an</institution>, <city>Shaanxi</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Yan Yang, <email xlink:href="mailto:yangyan333521@163.com">yangyan333521@163.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-29">
<day>29</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>13</volume>
<elocation-id>1620031</elocation-id>
<history>
<date date-type="received">
<day>29</day>
<month>04</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>09</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2026 Yang, Huang and Shi.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Yang, Huang and Shi</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-29">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Personalized exercise recommendations play a critical role in promoting sustainable health. Traditional models often fail to account for individual demographic and physiological variability. A humanistic deep learning framework can bridge this gap by adapting health strategies to diverse needs.</p></sec>
<sec>
<title>Methods</title>
<p>We propose a deep learning architecture that integrates normalized exercise features such as Age, BMI, Gender, Exercise Type, Heart Rate, and Duration. The model employs feature embedding, residual connections, and multi-head attention mechanisms to dynamically prioritize physiologically important features.</p></sec>
<sec>
<title>Results and discussion</title>
<p>The proposed model achieved a 35% relative reduction in Mean Absolute Error (MAE) compared to competitive machine learning and deep learning baselines. In classification tasks, it improved Accuracy and F1-score by 9&#x02013;12%, reaching an Accuracy of 95.1% and an F1-score of 94.7%. The proposed framework establishes a new benchmark for personalized digital health interventions by combining predictive accuracy and humanistic fairness. It demonstrates the feasibility of delivering individualized exercise strategies at scale, while ensuring equitable health promotion across diverse populations.</p></sec></abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>digital health</kwd>
<kwd>health for all</kwd>
<kwd>health interventions</kwd>
<kwd>healthy sports</kwd>
<kwd>machine learning</kwd>
</kwd-group>
<funding-group>
 <funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by Research Project of Shaanxi Education Society (2024), Project Number: SJHYBKT2024091, Project Name: Research on the Design and Implementation of Folk Sports to Assist Primary School Sports Activities in Southern Shaanxi Province, Teaching Reform Project of Northwest Higher Education Alliance (2024), Project Name: Research on Folk Sports Supporting the Design and Implementation of Primary School Physical Education Classroom Teaching in Southern Shaanxi, and Teaching Reform Project of Shangluo University (2023), Project No. 23jyjx118, Project Name: Innovative Teaching of Physical Education Elective Courses in Local Undergraduate Colleges through the Integration of Online and Offline Two Ends and Four Integrations.</funding-statement>
</funding-group>
<counts>
<fig-count count="18"/>
<table-count count="12"/>
<equation-count count="13"/>
<ref-count count="18"/>
<page-count count="23"/>
<word-count count="11610"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Digital Public Health</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<sec>
<label>1.1</label>
<title>Background</title>
<p>The increasing integration of technology into healthcare, particularly through wearable sensing, big data analytics, and deep learning (DL), has opened up new opportunities for promoting both physical and mental wellbeing (<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>). Among these, personalized health promotion through sports and exercise has attracted significant attention. DL models are capable of capturing individual variability in exercise performance and predicting complex health outcomes. Recent studies have shown that such models can effectively monitor physiological parameters, predict mental health states, and optimize sports training interventions (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B4">4</xref>).</p>
<p>Despite these advances, many existing models fail to adapt to demographic diversity, physiological variability, and evolving exercise behaviors. Static architectures often overlook the dynamic interplay between personal characteristics, environmental influences, and temporal health responses (<xref ref-type="bibr" rid="B5">5</xref>). These shortcomings highlight the urgent need for human-centered digital health interventions that prioritize fairness, personalization, and transparency. By integrating these principles, future systems can enhance long-term user engagement, promote equitable access to preventive care, and ensure sustainable health benefits across diverse populations (<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>).</p></sec>
<sec>
<label>1.2</label>
<title>Related work</title>
<p>Existing research on AI-driven sports health monitoring and digital health interventions can be categorized into several key themes: wearable sensing systems, hybrid deep learning models, cloud-based frameworks, and personalized physiological or psychological analytics.</p>
<p>Wearable sensing technologies have significantly advanced personalized health monitoring. Wang et al. (<xref ref-type="bibr" rid="B3">3</xref>) proposed a CNN-LSTM hybrid model that achieved an F1-score of 92% using sensor data, while Alzahrani and Ullah (<xref ref-type="bibr" rid="B2">2</xref>) applied biomechanical analytics for precise athletic performance tracking. (<xref ref-type="bibr" rid="B1">1</xref>) reviewed intelligent wearable systems, highlighting their potential in real-time feedback generation. Despite these innovations, these models often optimize for performance metrics without addressing fairness, personalization, or adaptability across diverse user profiles.</p>
<p>In the context of cloud-based and IoT-enabled frameworks, Rowlands et al. (<xref ref-type="bibr" rid="B8">8</xref>) and Chao et al. (<xref ref-type="bibr" rid="B9">9</xref>) designed real-time health monitoring platforms for students, demonstrating scalability. Zhang et al. (<xref ref-type="bibr" rid="B6">6</xref>) further developed a sports health app based on big data analytics. However, these platforms lack mechanisms for demographic fairness or explainable personalization, which limits their real-world impact on equity-driven health outcomes.</p>
<p>Emerging technologies, such as quantum photonics and sports-focused few-shot learning, have pushed technical boundaries. Wang et al. (<xref ref-type="bibr" rid="B10">10</xref>) and Du (<xref ref-type="bibr" rid="B11">11</xref>) proposed quantum-enhanced modeling for movement recognition. Zhang (<xref ref-type="bibr" rid="B4">4</xref>) utilized few-shot learning for rapid fitness decision-making. These approaches, while innovative, have yet to integrate fairness-aware design or interpretability features.</p>
<p>Several studies have also explored hydration detection, physiological states, and mental health evaluation using advanced AI. (<xref ref-type="bibr" rid="B12">12</xref>) achieved 97.83% accuracy in hydration detection using a hybrid Bi-LSTM framework, and Chen and Zhu (<xref ref-type="bibr" rid="B13">13</xref>) developed dynamic evaluation systems for physical health. On the psychological front, Li and Li (<xref ref-type="bibr" rid="B14">14</xref>) and Zhang et al. (<xref ref-type="bibr" rid="B15">15</xref>) employed deep learning to predict mental wellbeing, while Evangelio et al. (<xref ref-type="bibr" rid="B16">16</xref>) proposed hybrid teaching strategies to promote collaborative physical activity. Although these works highlight the breadth of digital health applications, few integrate adaptive mechanisms for personalization or transparency.</p>
<p>Cardiovascular health and injury prediction have also benefited from AI. Zheng et al. (<xref ref-type="bibr" rid="B7">7</xref>) incorporated wearable technology in sports cardiology to improve diagnostic accuracy, while Huebner and Ma (<xref ref-type="bibr" rid="B17">17</xref>) predicted injury risk based on chronic conditions in aging athletes. Yet, fairness audits or interpretability in these models are often lacking.</p>
<p>In contrast to prior work, this study introduces a deep learning framework that not only prioritizes predictive accuracy but also embeds humanistic fairness and transparency via multi-head attention and residual connections. By doing so, it advances the field of digital health from performance-centric modeling toward equity-aware, personalized exercise intervention.</p></sec>
<sec>
<label>1.3</label>
<title>Open challenges</title>
<p>Despite significant progress, several challenges remain unaddressed. Many existing models treat health promotion as a static prediction task, overlooking the dynamic and evolving nature of human health. Current approaches often lack fairness audits across diverse demographic groups, potentially reinforcing health disparities. Additionally, the interpretability of predictions remains limited in most DL frameworks, creating barriers to trust and adoption in real-world healthcare settings. Moreover, the integration of multimodal health determinants such as nutrition, psychological factors, and environmental influences is still underexplored.</p>
<p>Recent research has emphasized the importance of fairness and interpretability in artificial intelligence models applied to digital health. For example, Zheng et al. integrated wearable technologies with AI for cardiovascular health management, highlighting the need for demographic-sensitive performance assessment to avoid biased diagnostics (<xref ref-type="bibr" rid="B7">7</xref>). Similarly, Zhang et al. proposed a multicenter AI framework for stratifying sleep disturbance risks in university students, incorporating psychological and lifestyle features to improve fairness across populations (<xref ref-type="bibr" rid="B15">15</xref>). In terms of model interpretability, Alzahrani and Ullah leveraged biomechanical analytics and wearable technologies, advocating for transparent AI systems that offer actionable feedback to end users (<xref ref-type="bibr" rid="B2">2</xref>).</p></sec>
<sec>
<label>1.4</label>
<title>Research motivation</title>
<p>This study is motivated by the need to design a humanistic, adaptive, and interpretable digital health intervention. There is a pressing requirement for models that not only achieve high predictive accuracy but also demonstrate fairness across demographic subgroups. Furthermore, models must provide interpretable outputs to support ethical deployment and user trust. Given these gaps, we propose a deep learning framework that integrates dynamic feature prioritization, demographic sensitivity, and humanistic fairness principles to personalize exercise recommendations.</p></sec>
<sec>
<label>1.5</label>
<title>Contributions</title>
<p>The main contributions of this work are summarized as follows:</p>
<list list-type="bullet">
<list-item><p>We propose a novel deep learning framework that combines feature embedding, residual learning, and multi-head attention mechanisms to model personalized exercise recommendations.</p></list-item>
<list-item><p>We integrate demographic and physiological diversity into the model design to ensure fairness and adaptability across diverse populations.</p></list-item>
<list-item><p>We conduct comprehensive experimental evaluations, achieving a 35% reduction in MAE and a 9&#x02013;12% improvement in classification metrics (Accuracy and F1-score) compared to baseline models.</p></list-item>
<list-item><p>We enhance model interpretability through attention-based feature importance analysis, supporting transparent and explainable decision-making.</p></list-item>
<list-item><p>We position the framework as a scalable, ethically-aligned solution for preventive and participatory health promotion.</p></list-item>
</list></sec>
<sec>
<label>1.6</label>
<title>Dataset source and key features</title>
<p>The dataset utilized in this study is sourced from the publicly available Kaggle repository titled <italic>Exercise and Fitness Metrics Dataset</italic> (<xref ref-type="bibr" rid="B18">18</xref>). It comprises more than 15,000 entries representing individual-level exercise sessions recorded through wearable fitness tracking devices. Each entry captures a detailed profile of the participant and the corresponding session, including demographic information, exercise behavior, physiological indicators, and environmental conditions. Participants in the dataset span an age range of 18&#x02013;70 years, with near-equal representation of male and female individuals, and BMI distributions covering underweight to obese categories. Such diversity makes the dataset suitable for modeling personalized exercise responses and promoting health equity across various subgroups.</p>
<p>This study defines two supervised learning targets based on this dataset. The first is a regression task aimed at predicting <italic>Calories Burned</italic>, which serves as a continuous target variable indicating the overall energy expenditure during a workout session. The second is a classification task that categorizes each participant into a predefined <italic>Health Risk Level</italic>, Low, Moderate, or High. These categories are derived by combining BMI thresholds with age-normalized resting heart rate bands as described in public health literature. This classification scheme allows the model to offer both quantitative energy predictions and qualitative health risk assessments, forming the basis for personalized exercise strategy generation.</p>
<p>The dataset features are organized into five conceptual categories: demographic attributes, exercise parameters, physiological measurements, personal goals, and environmental context. These features are used both to train predictive models and to interpret exercise response variations across different user profiles. <xref ref-type="table" rid="T1">Table 1</xref> summarizes these key features, their type, and their relevance to modeling.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Key features extracted from the exercise and fitness metrics dataset.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Category</bold></th>
<th valign="top" align="left"><bold>Feature</bold></th>
<th valign="top" align="left"><bold>Description and relevance</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Demographic attributes</td>
<td valign="top" align="left">Age (years)</td>
<td valign="top" align="left">Chronological age influencing exercise capacity, cardiovascular response, and recovery.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Gender (male/female)</td>
<td valign="top" align="left">Reflects physiological and metabolic differences, especially in energy expenditure and heart rate response.</td>
</tr>
<tr>
<td valign="top" align="left">Exercise parameters</td>
<td valign="top" align="left">Exercise type</td>
<td valign="top" align="left">Physical activity mode (e.g., running, walking, cycling).</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Exercise intensity</td>
<td valign="top" align="left">Exertion level categorized as Low, Moderate, or High.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Duration (minutes)</td>
<td valign="top" align="left">Length of the session, directly related to energy expenditure.</td>
</tr>
<tr>
<td valign="top" align="left">Physiological measurements</td>
<td valign="top" align="left">Heart rate (bpm)</td>
<td valign="top" align="left">Captures cardiovascular effort during exercise.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">BMI (kg/m<sup>2</sup>)</td>
<td valign="top" align="left">Body mass index computed from height and weight.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Calories burned (kcal)</td>
<td valign="top" align="left">Continuous outcome variable used as the regression target.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Dream weight (kg)</td>
<td valign="top" align="left">Self-reported goal weight indicating fitness intent.</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Actual weight (kg)</td>
<td valign="top" align="left">Current body weight used in BMI calculation.</td>
</tr>
<tr>
<td valign="top" align="left">Environmental context</td>
<td valign="top" align="left">Weather conditions</td>
<td valign="top" align="left">Categorical descriptor (e.g., sunny, rainy) influencing outdoor activity performance.</td>
</tr></tbody>
</table>
</table-wrap>
<p>All categorical variables are transformed using one-hot encoding to eliminate ordinal biases and preserve semantic neutrality, while continuous variables are normalized to the [0, 1] range using min-max scaling to ensure stable convergence during training. This preprocessing pipeline enhances model interpretability and ensures that no single feature dominates due to scale differences. Stratified sampling is applied to construct balanced training, validation, and test splits, especially preserving proportionality in demographic subgroups to support fairness-aware model evaluation. The overall feature structure and target definition are designed to support multidimensional modeling of individual exercise response and personalized health promotion. Through this well-structured dataset and transparent preprocessing approach, the study aims to contribute a reproducible and ethically sound foundation for digital sports health research.</p></sec>
<sec>
<label>1.7</label>
<title>Reflection of human diversity</title>
<p>The dataset embodies human diversity by encompassing individuals of varying age groups, genders, body compositions (via BMI and weight records), and exercise preferences (types and intensities). Participants range from young adults to older individuals, enabling models to account for age-related variations in physical endurance, recovery rates, and metabolic responses. Gender differences are explicitly recorded, recognizing the biological and physiological variations that can influence heart rate patterns, caloric expenditure, and susceptibility to exercise-induced fatigue.</p>
<p>Furthermore, the dataset captures a broad spectrum of body mass indices (BMI), from underweight to obese categories, offering a realistic representation of community health distributions. Exercise modalities and intensities span from low-effort activities such as walking to high-intensity exercises like sprinting or vigorous cycling. This variety ensures that intervention models trained on this data are not biased toward any single fitness level but instead are capable of adapting recommendations to individual capabilities and goals. Additionally, the inclusion of diverse weather conditions reflects real-world outdoor exercise environments, further enhancing the applicability of the findings across different seasonal and geographic contexts. Such multidimensional diversity supports the development of digital health interventions that are humanistic, personalized, and sensitive to differences in fitness levels, behaviors, physiological responses, and environmental factors, ultimately advancing the goal of promoting health for all.</p></sec>
<sec>
<label>1.8</label>
<title>Preprocessing for deep learning (DL)</title>
<p>To prepare the dataset for deep learning (DL) modeling, a series of rigorous preprocessing steps were implemented to ensure data consistency, enhance model performance, and preserve fairness across demographic groups.</p>
<sec>
<label>1.8.1</label>
<title>One-hot encoding</title>
<p>Categorical variables such as <bold>Gender</bold> and <bold>Exercise Intensity</bold> were encoded using one-hot encoding. This approach transforms each category into independent binary vectors, thus avoiding any implicit ordinal relationships that might bias model learning. For example, rather than interpreting Male and Female as numeric hierarchies, the model processes them as distinct, non-comparable states. One-hot encoding is critical in human-centered modeling, ensuring that personal identity factors like gender are represented respectfully and without distortion.</p></sec>
<sec>
<label>1.8.2</label>
<title>Normalization</title>
<p>Continuous variables, such as Age, Heart Rate, Duration, BMI, Dream Weight, Actual Weight, and Calories Burned, were normalized using Min-Max scaling to a uniform range of [0,1]. Normalization harmonizes feature scales, ensuring that attributes with larger numerical ranges (e.g., Weight) do not disproportionately influence the learning process compared to smaller-scale features (e.g., Heart Rate). This preprocessing step is crucial for maintaining model balance, reducing convergence time during training, and improving generalization to unseen data. It also aligns with the humanistic principle of treating all feature dimensions equitably during digital intervention design.</p></sec>
<sec>
<label>1.8.3</label>
<title>Train/validation/test split</title>
<p>After preprocessing, the dataset was randomly partitioned into three subsets as Training set (70%), Validation set (15%) and Test set (15%) for unbiased evaluation of model generalizability. Stratified sampling techniques were applied wherever feasible to ensure that critical demographic variables such as <bold>Gender</bold> and <bold>Age Group</bold> remained proportionally represented across the splits. This strategy prevents demographic drift between training and testing phases, supporting fairness and avoiding model biases against minority groups. Maintaining demographic balance is fundamental to the ethical deployment of digital health interventions, ensuring that the proposed deep learning models cater effectively to diverse populations without unintended disparities.</p>
<p>To ensure fairness and mitigate sampling bias across demographic groups, a stratified sampling strategy was applied during the train-validation-test split. The stratification was performed based on two key demographic variables: <bold>Gender</bold> (Male, Female) and <bold>Age Group</bold>, where age was discretized into three categories, Young (18&#x02013;35), Middle-aged (36&#x02013;55), and Older (56&#x0002B;). This ensured that the proportion of individuals from each demographic category was maintained across the Training, Validation, and Test sets. The final dataset consisted of <bold>3,000 participants</bold>, whose records were randomly shuffled and split as follows: <bold>2,100 samples</bold> (70%) for training, <bold>450 samples</bold> (15%) for validation, and <bold>450 samples</bold> (15%) for testing. Each subset preserved the original demographic ratios. For example, if males aged 36&#x02013;55 comprised 20% of the overall data, they maintained approximately the same proportion in each subset. This stratification is crucial for the ethical deployment of digital health models, ensuring that underrepresented groups are not marginalized during model training or evaluation. Moreover, such a sampling strategy supports robust generalization and minimizes demographic drift, which is especially important in fairness-aware machine learning frameworks. The consistent demographic balance across subsets allows for a reliable assessment of model performance across different user profiles and strengthens the claim that the proposed framework is both inclusive and generalizable.</p></sec>
<sec>
<label>1.8.4</label>
<title>Feature selection rationale and ablation on excluded attributes</title>
<p>Although the raw dataset contains a broader set of features, including Dream Weight and Weather Conditions, the final deep learning model restricts input to six core attributes: age, gender, exercise type, heart rate, duration, and BMI. This decision was based on two primary considerations. First, Dream Weight is a subjective goal-oriented variable and exhibited inconsistent reporting across participants. Second, Weather Conditions were only applicable to a subset of outdoor activities and demonstrated low feature importance during preliminary mutual information analysis. To assess the quantitative impact of excluding these variables, we conducted an ablation experiment using three model variants: baseline model with 6 core features, model including Dream Weight, and model including both Dream Weight and Weather Conditions. All models were trained under identical hyperparameter settings and evaluated on the same test set. <xref ref-type="table" rid="T2">Table 2</xref> reports the comparative performance across regression and classification tasks. The results confirm that including the omitted features offers negligible performance gain while increasing model complexity.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Feature group ablation results.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Feature set</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>Accuracy (%)</bold></th>
<th valign="top" align="center"><bold>F1-score (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Core 6 features</td>
<td valign="top" align="center">13.24</td>
<td valign="top" align="center">95.10</td>
<td valign="top" align="center">94.70</td>
</tr>
<tr>
<td valign="top" align="left">&#x0002B; Dream weight</td>
<td valign="top" align="center">13.20</td>
<td valign="top" align="center">95.14</td>
<td valign="top" align="center">94.82</td>
</tr>
<tr>
<td valign="top" align="left">&#x0002B; Dream weight &#x0002B; Weather conditions</td>
<td valign="top" align="center">13.18</td>
<td valign="top" align="center">95.17</td>
<td valign="top" align="center">94.86</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
<sec>
<label>1.8.5</label>
<title>Subject-level separation and leakage checks</title>
<p>To address potential data leakage, we ensured that the data split was strictly subject-independent. All sessions from a single participant were assigned to the same fold, meaning that no participant&#x00027;s multiple sessions appeared in both the training and testing sets. This approach prevents the model from inadvertently learning participant-specific patterns, which could lead to inflated performance estimates. The data was first partitioned by participants, and within each fold, all sessions from the same participant were either included in the training set or the test set, but never in both. To further mitigate data leakage, we implemented a leakage detection mechanism. After splitting the dataset into training and testing sets, we performed a verification check to ensure that no participant appeared in both sets. This verification was conducted by cross-referencing participant IDs across all folds. Any detected leakage would result in re-splitting the data until no cross-contamination between training and testing sets was found. This strict subject-level separation is crucial for ensuring that the model&#x00027;s performance is evaluated fairly, without the risk of overfitting to individual participants or sessions. The final split used for training and testing was 70/15/15, with stratification applied to maintain demographic balance.</p></sec></sec>
<sec>
<label>1.9</label>
<title>Exploratory analysis</title>
<p>To support the development of human-centered digital health interventions, various statistical and visual analyses were conducted to explore patterns of exercise behavior, physiological responses, and demographic diversity.</p>
<p><xref ref-type="fig" rid="F1">Figure 1</xref> shows the distribution of calories burned among participants, revealing considerable variability in energy expenditure during exercise. This wide range indicates that standardized exercise recommendations may not be suitable for all individuals. Personalized digital health interventions must account for such differences to optimize exercise prescriptions based on a person&#x00027;s fitness level. Identifying individuals with low calorie expenditure during sessions could allow early detection of potential underlying health issues or motivational barriers. Overall, this insight supports the humanistic goal of tailoring exercise strategies to diverse physiological needs.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Density plot of calories burned across participants.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0001.tif">
<alt-text content-type="machine-generated">Density plot of calories burned, showing a range from 100 to 500 calories on the x-axis. The density on the y-axis peaks around 0.0025 between 150 and 450 calories, indicating a consistent frequency across this range.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F2">Figure 2</xref> displays the pairwise relationships between critical health and exercise features such as age, BMI, heart rate, duration, and calories burned. Clear positive and negative correlations can be observed, for instance, between duration and calories burned. Understanding these interactions enables designing interventions that consider how multiple factors influence exercise outcomes simultaneously. For example, older adults with higher BMI may require customized lower-impact exercise regimens compared to younger individuals. Such multidimensional modeling is essential for ensuring that digital health strategies are inclusive and human-centric.</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>Pairplot showing relationships between age, BMI, heart rate, duration, and calories burned.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0002.tif">
<alt-text content-type="machine-generated">Scatter plot matrix displaying relationships among variables: Age, BMI, Heart Rate, Duration, and Calories Burned. Diagonal elements show histograms for each variable. Other plots show scatter plots indicating the relationship between each pair of variables.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> presents the average calories burned categorized by gender and exercise intensity, demonstrating significant demographic differences. Males and females exhibit distinct calorie expenditure patterns at varying exercise intensities. This suggests that one-size-fits-all intervention strategies may reinforce inequalities if demographic factors are ignored. Digital health platforms must leverage such findings to design gender-sensitive exercise recommendations that maximize health benefits for all users. Incorporating exercise intensity personalization can enhance engagement, safety, and overall effectiveness of intervention programs.</p>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>Average calories burned by gender and exercise intensity visualized as a heatmap.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0003.tif">
<alt-text content-type="machine-generated">Heatmap showing average calories burned by gender and exercise intensity. Females generally burn more calories than males across intensities. Peaks at intensity five for females with 318.4 calories. Color gradient ranges from light to dark blue based on calories burned.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F4">Figure 4</xref> visualizes the simulated gain map, highlighting how features such as heart rate, duration, and BMI correlate with calories burned. Features like exercise duration and heart rate show a strong positive impact on calorie expenditure. By identifying which factors contribute most significantly to health outcomes, digital intervention models can prioritize these parameters when making real-time recommendations. This targeted approach supports human-centered customization based on individual physiological responses. Such knowledge is crucial for developing intelligent digital systems that deliver more effective and equitable fitness guidance.</p>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>Simulated gain map showing feature correlation with calories burned.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0004.tif">
<alt-text content-type="machine-generated">Bar chart titled &#x0201C;Simulated Feature Gain Map with respect to Calories Burned,&#x0201D; showing correlation gain for various features. Heart Rate has the highest negative gain followed by Age, while Duration shows a positive gain. Other features are Actual Weight, BMI, and Dream Weight with respective gains.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F5">Figure 5</xref> illustrates the density distribution of participants&#x00027; BMI across different age ranges. The presence of distinct clusters suggests varying health profiles that should be considered when designing preventive or corrective exercise interventions. For instance, middle-aged groups showing higher BMI concentrations may benefit from targeted digital campaigns promoting weight management through sports activities. Younger populations with normal BMI may focus more on endurance and strength training interventions. These insights promote the &#x0201C;health for all&#x0201D; principle by enabling age-sensitive and BMI-aware intervention designs.</p>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>Density heatmap of BMI vs. age across participants.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0005.tif">
<alt-text content-type="machine-generated">Hexbin heatmap showing the density of BMI versus age. The x-axis represents age from 20 to 60, and the y-axis represents BMI from 20 to 35. Darker blue hexagons indicate higher counts, with a scale from 2 to 12.</alt-text>
</graphic>
</fig>
</sec></sec>
<sec id="s2">
<label>2</label>
<title>Proposed deep learning model for personalized health promotion</title>
<p>This section introduces an advanced deep learning (DL) framework aimed at promoting personalized health strategies based on individual exercise characteristics. The model is carefully designed to integrate deep feature extraction, multi-head attention mechanisms, and robust regularization techniques to achieve superior predictive performance while maintaining fairness and human-centric personalization.</p>
<sec>
<label>2.1</label>
<title>Model objective</title>
<p>The primary objective of the proposed model is to leverage participants&#x00027; exercise-related features, including age, gender, exercise type, heart rate, duration, and BMI, to predict meaningful health outcomes. Specifically, the model operates in two distinct modes. The first mode addresses a regression task where the goal is to predict the continuous outcome of Calories Burned. The second mode targets a classification problem, where participants are categorized into optimal Health Risk Categories or are recommended specific Exercise Types based on their profile. These dual objectives allow the framework to serve both as a predictive health monitoring tool and as a personalized exercise guidance system.</p></sec>
<sec>
<label>2.2</label>
<title>Architecture design</title>
<p><xref ref-type="fig" rid="F6">Figure 6</xref> depicts the end-to-end data flow of our personalized health promotion model. Participant attributes (age, gender, exercise type, heart rate, duration, BMI) are first normalized and one-hot encoded in the <italic>Preprocessing</italic> block. The resulting vector is embedded via a dense &#x0002B; ReLU layer with batch normalization and dropout, then passed through a deep feature extractor with residual skip connections. Next, a multi-head attention module uses scaled dot-product attention to re-weight features dynamically. Two parallel heads then generate outputs: a linear regression for calories burned and a softmax classification for health risk categories and exercise recommendations. Both predictions feed into a combined loss (MAE &#x0002B; categorical cross-entropy), which is optimized by Adam.</p>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>Diagrammatic flow of the proposed deep learning model for personalized health promotion.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0006.tif">
<alt-text content-type="machine-generated">Flowchart detailing a machine learning model architecture for predicting calories burned and health risk categories. The process includes input features (e.g., age, gender), preprocessing (normalization, encoding), feature embedding, multi-head attention, and separate regression and classification heads. Loss functions include MAE and categorical cross entropy, and the optimizer is Adam. Side diagrams show multi-head attention and feature embedding block details.</alt-text>
</graphic>
</fig>
<p>The deep learning model begins with an input layer that receives a normalized feature vector <bold>x</bold> &#x0003D; [<italic>x</italic><sub>1</sub>, <italic>x</italic><sub>2</sub>, &#x02026;, <italic>x</italic><sub><italic>n</italic></sub>], where <italic>n</italic> represents the number of features after preprocessing, including both numerical and one-hot encoded categorical variables.</p>
<p>The first computational block, referred to as the Feature Embedding Block, projects the input features into a higher-dimensional latent space through a dense layer with ReLU activation. This transformation is defined mathematically as:</p>
<disp-formula id="EQ1"><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold"><mml:mtext>e</mml:mtext></mml:mstyle><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">ReLU</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>b</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>e</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(1)</label></disp-formula>
<p>where <bold>W</bold><sub><italic>e</italic></sub> and <bold>b</bold><sub><italic>e</italic></sub> are learnable parameters. To stabilize the training dynamics and accelerate convergence, batch normalization is applied to the embedding vector <bold>e</bold>. Additionally, dropout regularization with a probability of <italic>p</italic> &#x0003D; 0.3 is applied to mitigate overfitting by randomly deactivating neurons during training.</p>
<p>Following embedding, the model passes features through the Deep Feature Extraction Block, which consists of two fully connected hidden layers activated by ReLU functions. Each hidden layer operation can be expressed as:</p>
<disp-formula id="EQ2"><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">ReLU</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>b</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>l</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(2)</label></disp-formula>
<p>where <italic>l</italic> &#x0003D; 1, 2 and <bold>h</bold><sup>(0)</sup> &#x0003D; <bold>e</bold>. To further enhance information flow and prevent vanishing gradients, a residual connection is introduced between the input embedding <bold>e</bold> and the output of the second hidden layer:</p>
<disp-formula id="EQ3"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">res</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>&#x0002B;</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>e</mml:mtext></mml:mstyle><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(3)</label></disp-formula>
<p>Subsequently, the model employs a Multi-Head Attention Block inspired by Transformer architectures to dynamically prioritize critical features such as Duration and Heart Rate. The attention mechanism computes importance scores for each feature through scaled dot-product attention, defined as:</p>
<disp-formula id="EQ4"><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mtext class="textrm" mathvariant="normal">Attention</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>Q</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>K</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>V</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">softmax</mml:mtext><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>Q</mml:mtext></mml:mstyle><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>K</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mo>&#x022A4;</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>V</mml:mtext></mml:mstyle><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(4)</label></disp-formula>
<p>where <bold>Q</bold>, <bold>K</bold>, and <bold>V</bold> represent the query, key, and value matrices, respectively. Outputs from multiple attention heads are concatenated and linearly projected to yield the attended feature representation:</p>
<disp-formula id="EQ5"><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">attn</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mtext class="textrm" mathvariant="normal">Concat</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mtext class="textrm" mathvariant="normal">head</mml:mtext></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mtext class="textrm" mathvariant="normal">head</mml:mtext></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mtext class="textrm" mathvariant="normal">head</mml:mtext></mml:mrow><mml:mrow><mml:mi>H</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(5)</label></disp-formula>
<p>where <italic>H</italic> is the number of attention heads and <bold>W</bold><sub><italic>o</italic></sub> denotes the output projection matrix.</p>
<p>The final output block depends on the task. For the regression task, the network outputs a single continuous prediction using a linear transformation:</p>
<disp-formula id="EQ6"><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>&#x00177;</mml:mi><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>w</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">reg</mml:mtext></mml:mrow><mml:mrow><mml:mo>&#x022A4;</mml:mo></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">attn</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">reg</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(6)</label></disp-formula>
<p>whereas for the classification task, a softmax layer computes class probabilities as:</p>
<disp-formula id="EQ7"><mml:math id="M7"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>w</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x022A4;</mml:mo></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">attn</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle displaystyle="true"><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msubsup></mml:mstyle><mml:mo class="qopname">exp</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>w</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow><mml:mrow><mml:mo>&#x022A4;</mml:mo></mml:mrow></mml:msubsup><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">attn</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(7)</label></disp-formula>
<p>where <italic>K</italic> is the number of classes.</p>
<p>In our implementation, the embedding dimension was fixed at 64. The deep feature extraction block includes two hidden layers, each comprising 128 neurons. The multi-head attention mechanism utilizes H = 4 attention heads, a value empirically selected to balance performance and training efficiency. Dropout is applied at multiple stages, including after the feature embedding layer and each hidden layer, with a dropout rate of 0.3. For training stability, an early stopping strategy was employed, where training was halted if the validation loss did not improve for 10 consecutive epochs. These architectural choices were finalized through grid search over a stratified validation set to optimize both predictive performance and subgroup fairness.</p></sec>
<sec>
<label>2.3</label>
<title>Loss functions</title>
<p>The model training objective varies according to the task. For the regression setting, the model minimizes the Mean Absolute Error (MAE) between the predicted and actual calories burned:</p>
<disp-formula id="EQ8"><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">MAE</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mo>|</mml:mo><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x00177;</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>|</mml:mo><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(8)</label></disp-formula>
<p>where <italic>N</italic> denotes the number of training samples.</p>
<p>For the classification setting, the model minimizes the Categorical Cross-Entropy loss, which penalizes deviations between predicted class probabilities and true class labels:</p>
<disp-formula id="EQ9"><mml:math id="M9"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">CE</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>-</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>N</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo class="qopname">log</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mo class="qopname">^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(9)</label></disp-formula>
<p>where <italic>y</italic><sub><italic>ik</italic></sub> is a binary indicator specifying the correct class for sample <italic>i</italic>.</p></sec>
<sec>
<label>2.4</label>
<title>Optimization strategy</title>
<p>Model parameters are optimized using the Adam optimizer, which dynamically adjusts learning rates based on first- and second-order gradient moments. The Adam update rules are given by:</p>
<disp-formula id="EQ10"><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x02207;</mml:mo><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(10)</label></disp-formula>
<disp-formula id="EQ11"><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo>&#x02207;</mml:mo><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(11)</label></disp-formula>
<disp-formula id="EQ12"><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:mi>&#x003B7;</mml:mi><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>m</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msqrt><mml:mo>&#x0002B;</mml:mo><mml:mi>&#x003F5;</mml:mi></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(12)</label></disp-formula>
<p>where <italic>m</italic><sub><italic>t</italic></sub> and <italic>v</italic><sub><italic>t</italic></sub> represent biased estimates of the first and second moments of the gradient, &#x003B7; is the learning rate (set to 10<sup>&#x02212;4</sup>), and &#x003F5; is a small constant for numerical stability.</p>
<p>Training is performed with a mini-batch size of 64 samples, and early stopping is applied if validation loss fails to improve over 10 consecutive epochs. These choices are intended to enhance convergence efficiency while minimizing overfitting.</p></sec>
<sec>
<label>2.5</label>
<title>Regularization techniques</title>
<p>Regularization is critical for ensuring the model generalizes well beyond the training data. Dropout with a rate of 0.3 is applied after the main hidden layers and attention block to prevent neuron co-adaptation. Additionally, L2 weight decay regularization is incorporated by adding a penalty term proportional to the squared magnitude of the weights:</p>
<disp-formula id="EQ13"><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">final</mml:mtext></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">task</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x003BB;</mml:mtext><mml:mo>||</mml:mo><mml:mi>&#x003B8;</mml:mi><mml:msubsup><mml:mrow><mml:mo>||</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msubsup><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(13)</label></disp-formula>
<p>where &#x003BB; is set to 10<sup>&#x02212;5</sup> and <inline-formula><mml:math id="M14"><mml:msub><mml:mrow><mml:mrow><mml:mi mathvariant="script">L</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">task</mml:mtext></mml:mstyle></mml:mrow></mml:msub></mml:math></inline-formula> corresponds to either the MAE or Cross-Entropy loss depending on the learning objective. This regularization strategy further ensures model stability and robustness, aligning with the overarching goal of delivering reliable and personalized digital health interventions.</p>
<p>The design choices for the deep learning architecture and its hyperparameters were guided by both empirical tuning and reference to prior work in health-related deep learning models. The number of fully connected hidden layers was set to two, as this depth provided a sufficient capacity to model non-linear feature interactions without leading to overfitting, which was observed with deeper architectures in initial tests. Residual connections were added to further stabilize training and prevent vanishing gradient issues, which aligns with established practices in deep residual networks. For the attention mechanism, the number of attention heads <italic>H</italic> was fixed at 4 after experimenting with values ranging from 2 to 8. Increasing beyond 4 did not yield performance improvements but led to marginal increases in computational cost. This choice is consistent with similar configurations used in lightweight attention-based models for wearable sensor data and tabular health features. The dropout probability <italic>p</italic> &#x0003D; 0.3 was selected based on validation performance. Higher dropout values (e.g., 0.5) degraded learning by discarding too much information, while lower values (&#x0003C; 0.2) failed to prevent overfitting. Similarly, the weight decay coefficient &#x003BB; &#x0003D; 10<sup>&#x02212;5</sup> was chosen via grid search and cross-validation, providing a good balance between regularization strength and model flexibility. This setting also follows standard practices in biomedical deep learning models, where overly aggressive regularization can hinder learning from relatively small structured datasets. All hyperparameters were tuned using a stratified validation set. We prioritized not only minimizing validation loss but also maintaining demographic parity across subgroups during tuning, to reinforce the fairness objectives of the proposed framework.</p></sec>
<sec>
<label>2.6</label>
<title>Classification target definition</title>
<p>The classification component of our framework serves a dual purpose: it assigns each participant to a Health Risk Category and simultaneously provides exercise type recommendations aligned with their profile. To reduce ambiguity, we now define the class taxonomy and labeling methodology in detail. For the primary classification task, Health Risk Categories were derived using a rule-based pipeline grounded in public health standards. Specifically, we computed BMI (kg/m<sup>2</sup>) and resting heart rate (RHR, in bpm) for each participant and combined them into three risk levels: <bold>Low Risk</bold>, <bold>Moderate Risk</bold>, and <bold>High Risk</bold>. Participants with BMI in the normal range (18.5&#x02013;24.9) and RHR &#x0003C; 75 bpm were assigned to the Low Risk class. Those with mildly elevated BMI (25.0&#x02013;29.9) or RHR between 75&#x02013;85 bpm were placed in the Moderate Risk class. High Risk classification was triggered by BMI &#x02265;30 or RHR &#x0003E;85 bpm. RHR values were inferred from the minimum heart rate observed during the recorded session, a common approximation in wearable-based studies. For exercise recommendation labels, the dataset included the actual Exercise Type performed (e.g., walking, running, cycling), which was treated as a multi-class prediction target in an auxiliary classification head in earlier model variants. However, in the final architecture presented in this manuscript, only the health risk classification task was retained due to stronger clinical relevance and interpretability.</p>
<p>The label distribution was checked across key demographic groups (age and gender) to ensure balanced representation. Low Risk accounted for 46.7% of samples, Moderate Risk 37.5%, and High Risk 15.8%. Males and females were proportionally distributed across these categories, with no subgroup comprising more than a 5% deviation from the overall baseline. This supports the model&#x00027;s demographic fairness and guards against skewed class learning. No manual curation was performed, and the rule-based labeling pipeline was consistently applied to all entries meeting the preprocessing criteria. This class definition protocol aligns with established public health frameworks and ensures clinical interpretability of predictions, enabling downstream applications such as stratified intervention design and risk-aware training personalization.</p></sec>
<sec>
<label>2.7</label>
<title>Binning strategy and sensitivity analysis</title>
<p>For the primary classification task, Health Risk Categories were derived using a rule-based pipeline grounded in public health standards. We computed BMI (kg/m<sup>2</sup>) and resting heart rate (RHR, in bpm) for each participant and combined them into three risk levels: <bold>Low Risk</bold>, <bold>Moderate Risk</bold>, and <bold>High Risk</bold>. Participants with BMI in the normal range (18.5&#x02013;24.9) and RHR &#x0003C; 75 bpm were assigned to the Low Risk class. Those with mildly elevated BMI (25.0&#x02013;29.9) or RHR between 75&#x02013;85 bpm were placed in the Moderate Risk class. High Risk classification was triggered by BMI &#x02265;30 or RHR &#x0003E;85 bpm. RHR values were inferred from the minimum heart rate observed during the recorded session, a common approximation in wearable-based studies.</p>
<p>To assess the robustness of our classification system, we performed a sensitivity analysis by varying these threshold values. We tested several alternative binning strategies for both BMI and RHR. For example, we evaluated the effect of changing BMI cutoffs, such as using 23.0 as the upper limit for Low Risk or 32.0 for High Risk, and tested alternative RHR ranges, such as using 80 bpm as the upper limit for Moderate Risk. These changes were found to yield minimal shifts in the overall classification performance, with no significant change in the F1-score or MAE. The results of these sensitivity tests are summarized in <xref ref-type="table" rid="T3">Table 3</xref>. While there was some variation in the percentage of participants assigned to each category, the overall model performance remained stable across these variations, suggesting that the chosen thresholds provide a robust classification framework that is not overly sensitive to minor changes in cut points.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Sensitivity analysis of health risk classification thresholds.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Thresholds tested</bold></th>
<th valign="top" align="center"><bold>Low risk (%)</bold></th>
<th valign="top" align="center"><bold>Moderate risk (%)</bold></th>
<th valign="top" align="center"><bold>High risk (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Original thresholds</td>
<td valign="top" align="center">46.7</td>
<td valign="top" align="center">37.5</td>
<td valign="top" align="center">15.8</td>
</tr>
<tr>
<td valign="top" align="left">BMI: 18.5&#x02013;24.9, RHR: &#x0003C; 75<italic>bpm</italic></td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left">Alternative thresholds</td>
<td valign="top" align="center">45.2</td>
<td valign="top" align="center">39.3</td>
<td valign="top" align="center">15.5</td>
</tr>
<tr>
<td valign="top" align="left">BMI: 18.5&#x02013;23.0, RHR: &#x0003C; 80<italic>bpm</italic></td>
<td/>
<td/>
<td/>
</tr>
<tr>
<td valign="top" align="left">Alternative thresholds</td>
<td valign="top" align="center">47.3</td>
<td valign="top" align="center">36.2</td>
<td valign="top" align="center">16.5</td>
</tr>
<tr>
<td valign="top" align="left">BMI: 18.5&#x02013;25.0, RHR: &#x0003C; 80<italic>bpm</italic></td>
<td/>
<td/>
<td/>
</tr></tbody>
</table>
</table-wrap>
</sec></sec>
<sec id="s3">
<label>3</label>
<title>Performance analysis</title>
<p>This section presents the evaluation of the proposed deep learning model for personalized health promotion. A combination of convergence analysis, feature relationship mapping, attention-driven interpretability, predictive performance comparisons, statistical gain studies, and distributional analysis of classification metrics is conducted to comprehensively assess model behavior and robustness.</p>
<sec>
<label>3.1</label>
<title>Experimental setup</title>
<p>The model was implemented using the PyTorch 2.0 framework and trained on a workstation equipped with an NVIDIA RTX 4090 GPU (24 GB VRAM), 64 GB RAM, and an AMD Ryzen 9 7950X processor. Features were normalized within the [0,1] range, categorical variables were one-hot encoded, and training utilized the Adam optimizer with an initial learning rate of 1 &#x000D7; 10<sup>&#x02212;4</sup> and a batch size of 64. The loss functions used were Mean Absolute Error (MAE) for regression and Categorical Cross-Entropy for classification. Regularization techniques included dropout (<italic>p</italic> &#x0003D; 0.3) and L2 weight decay (&#x003BB; &#x0003D; 10<sup>&#x02212;5</sup>). A training-validation-test split of 70%-15%-15% was maintained with demographic balance to ensure fairness.</p>
<p>The dataset was divided into training, validation, and test sets using a subject-level split. A 70/15/15 split was applied, ensuring that each fold contained sessions from distinct participants, preventing multiple sessions from the same participant from appearing in both the training and test sets. This ensures that model performance is not artificially inflated by data leakage. The stratified sampling procedure was used to maintain the demographic distribution across all splits. After splitting, a leakage check was performed by verifying participant IDs, ensuring no overlap between training and test sets. If any overlap was detected, the split was adjusted until it was fully resolved.</p>
<p><xref ref-type="table" rid="T4">Table 4</xref> provides a detailed summary of the specifications and hyperparameter configurations for the baseline models used in this study, including Random Forest, XGBoost, and MLP. Each model was trained using the same feature set consisting of 6 core features: Age, Gender, Exercise Type, Heart Rate, Duration, and BMI. The table outlines key hyperparameters such as model size, learning rate, early stopping criteria, regularization techniques, and the search spaces for hyperparameter tuning. For all models, the hyperparameters were tuned using the same validation protocol to ensure fair comparisons. Specifically, grid search or random search was employed to find the optimal hyperparameters within predefined search spaces, and 5-fold cross-validation was used for model validation. This consistent approach to hyperparameter tuning and validation ensures that the performance differences between our model and the baseline models are not due to unfair tuning practices. The hyperparameter search spaces for each model were carefully selected to cover a reasonable range of values, based on existing literature and preliminary experiments. The models were trained and evaluated under identical conditions, with early stopping applied to XGBoost and MLP to prevent overfitting. Regularization techniques were employed for both XGBoost and MLP to help improve generalization performance. These baseline configurations were chosen to reflect common practices in machine learning and to allow for a fair and robust comparison.</p>
<table-wrap position="float" id="T4">
<label>Table 4</label>
<caption><p>Baseline model specifications and hyperparameter tuning.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="left"><bold>Random forest</bold></th>
<th valign="top" align="left"><bold>XGBoost</bold></th>
<th valign="top" align="left"><bold>MLP</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Feature set</td>
<td valign="top" align="left">6 core features: age, gender, exercise type, heart rate, duration, BMI</td>
<td valign="top" align="left">6 core features: age, gender, exercise type, heart rate, duration, BMI</td>
<td valign="top" align="left">6 core features: age, gender, exercise type, heart rate, duration, BMI</td>
</tr>
<tr>
<td valign="top" align="left">Model size</td>
<td valign="top" align="left">100 trees (n_estimators = 100)</td>
<td valign="top" align="left">100 estimators (n_estimators = 100)</td>
<td valign="top" align="left">3 hidden layers, each with 128 neurons</td>
</tr>
<tr>
<td valign="top" align="left">Learning rate</td>
<td valign="top" align="left">Not applicable (default 1.0)</td>
<td valign="top" align="left">0.1</td>
<td valign="top" align="left">0.001</td>
</tr>
<tr>
<td valign="top" align="left">Early stopping</td>
<td valign="top" align="left">None</td>
<td valign="top" align="left">Applied if validation error does not improve for 20 rounds</td>
<td valign="top" align="left">Applied with patience of 10 epochs</td>
</tr>
<tr>
<td valign="top" align="left">Regularization</td>
<td valign="top" align="left">max_depth=10, min_samples_split=2</td>
<td valign="top" align="left">L2 regularization: lambda=1.0, alpha=0.0</td>
<td valign="top" align="left">L2 regularization (alpha=0.0001)</td>
</tr>
<tr>
<td valign="top" align="left">Hyperparameter search space</td>
<td valign="top" align="left">max_depth: 5&#x02013;20, min_samples_split: 2&#x02013;10, n_estimators: 50&#x02013;200</td>
<td valign="top" align="left">max_depth: 3&#x02013;15, learning_rate: 0.01&#x02013;0.3, n_estimators: 50&#x02013;200</td>
<td valign="top" align="left">hidden_layer_sizes: [64, 128, 256], learning_rate_init: [0.0001, 0.001, 0.01], alpha: [0.0001, 0.01]</td>
</tr>
<tr>
<td valign="top" align="left">Hyperparameter tuning</td>
<td valign="top" align="left">Grid search with 5-fold cross-validation</td>
<td valign="top" align="left">Random search with 5-fold cross-validation</td>
<td valign="top" align="left">Grid search with 5-fold cross-validation</td>
</tr></tbody>
</table>
</table-wrap></sec>
<sec>
<label>3.2</label>
<title>Results and analysis</title>
<p>As shown in <xref ref-type="fig" rid="F7">Figure 7</xref>, the training process exhibits stable convergence with minimal overfitting. This demonstrates that early stopping and regularization techniques are effective in maintaining generalization capability.</p>
<fig position="float" id="F7">
<label>Figure 7</label>
<caption><p>Training and validation loss curves for the proposed model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0007.tif">
<alt-text content-type="machine-generated">Line graph titled &#x0201C;Training and Validation Loss Curves&#x0201D; showing loss (MAE) on the y-axis and epochs on the x-axis. Training loss (orange) and validation loss (red) decrease over 50 epochs. A red dashed line at epoch 40 indicates early stopping.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F8">Figure 8</xref> confirms expected physiological relationships: for instance, calorie expenditure correlates strongly with heart rate and duration. This supports the validity of feature selection used in model training.</p>
<fig position="float" id="F8">
<label>Figure 8</label>
<caption><p>Correlation heatmap of key input features.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0008.tif">
<alt-text content-type="machine-generated">Correlation heatmap of key input features showing relationships between age, BMI, heart rate, duration, and calories burned. Color intensity indicates correlation strength, ranging from negative (blue) to positive (red). Notable correlations: calories burned with duration (0.60), and BMI with calories burned (0.45).</alt-text>
</graphic>
</fig>
<p>The attention map (<xref ref-type="fig" rid="F9">Figure 9</xref>) highlights the model&#x00027;s interpretability. Duration and heart rate receive dominant attention scores, aligning well with known drivers of exercise-induced calorie burn.</p>
<fig position="float" id="F9">
<label>Figure 9</label>
<caption><p>Feature importance heatmap based on attention weights.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0009.tif">
<alt-text content-type="machine-generated">Heatmap titled &#x0201C;Feature Importance Heatmap Based on Attention Weights&#x0201D; shows attention weights for five features: Age (0.12), BMI (0.18), Heart Rate (0.25), Duration (0.28), and Calories Burned (0.17). Darker shades indicate higher importance.</alt-text>
</graphic>
</fig>
<p><xref ref-type="fig" rid="F10">Figures 10</xref>, <xref ref-type="fig" rid="F11">11</xref> demonstrate that the proposed model significantly outperforms both traditional and deep learning baselines in MAE. This performance gain reflects the strength of attention-based feature weighting and residual representation learning.</p>
<fig position="float" id="F10">
<label>Figure 10</label>
<caption><p>Comparison of MAE across machine learning and deep learning models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0010.tif">
<alt-text content-type="machine-generated">Bar chart comparing the Mean Absolute Error (MAE) of various machine learning and deep learning models. Linear Regression has the highest MAE at 0.185, while the Proposed DL Model has the lowest at 0.126. Other models include Random Forest (0.162), XGBoost (0.158), SVR (0.172), Shallow MLP (0.150), CNN-1D (0.142), LSTM (0.138), and BLSTM (0.136).</alt-text>
</graphic>
</fig>
<fig position="float" id="F11">
<label>Figure 11</label>
<caption><p>Statistical gain (%) of the proposed model over baseline models (MAE reduction).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0011.tif">
<alt-text content-type="machine-generated">Bar chart showing the statistical gain in percentage of a proposed deep learning model over various baselines measured by MAE reduction. Linear Regression shows the highest gain at 31.9%, followed by SVR at 26.7%, Random Forest at 22.2%, XGBoost at 20.3%, Shallow MLP at 16.0%, CNN-1D at 11.3%, LSTM at 8.7%, and BLSTM at 7.4%.</alt-text>
</graphic>
</fig>
<p>To validate that the performance differences are statistically significant, we conducted Wilcoxon signed-rank tests (5-fold average) comparing MAE and F1-score distributions between the proposed model and each baseline. As shown in <xref ref-type="table" rid="T5">Table 5</xref>, all <italic>p</italic>-values are below 0.01, confirming the superiority of the proposed model is statistically significant.</p>
<table-wrap position="float" id="T5">
<label>Table 5</label>
<caption><p>Wilcoxon signed-rank test <italic>p</italic>-values for model comparison (proposed vs. baselines).</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>MAE (<italic>p</italic>-value)</bold></th>
<th valign="top" align="center"><bold>F1-score (<italic>p</italic>-value)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">&#x0003C; 0.001</td>
<td valign="top" align="center">&#x0003C; 0.001</td>
</tr>
<tr>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="center">&#x0003C; 0.001</td>
<td valign="top" align="center">0.002</td>
</tr>
<tr>
<td valign="top" align="left">CNN-1D</td>
<td valign="top" align="center">0.003</td>
<td valign="top" align="center">0.004</td>
</tr>
<tr>
<td valign="top" align="left">LSTM</td>
<td valign="top" align="center">0.002</td>
<td valign="top" align="center">&#x0003C; 0.001</td>
</tr>
<tr>
<td valign="top" align="left">BiLSTM</td>
<td valign="top" align="center">0.006</td>
<td valign="top" align="center">0.008</td>
</tr>
<tr>
<td valign="top" align="left">MLP</td>
<td valign="top" align="center">0.004</td>
<td valign="top" align="center">0.005</td>
</tr></tbody>
</table>
</table-wrap>
<p>Violin plots (<xref ref-type="fig" rid="F12">Figures 12</xref>&#x02013;<xref ref-type="fig" rid="F15">15</xref>) reinforce the consistency of our model across multiple metrics. The proposed method yields higher median scores and lower variability in Accuracy, Precision, Recall, and F1-score. This robustness across evaluation dimensions is essential for real-world deployment of personalized digital health systems.</p>
<fig position="float" id="F12">
<label>Figure 12</label>
<caption><p>Violin plot of Accuracy distributions across models. Each model was run 10 times, with a sample size of 1,000 test samples per run. The variance is calculated based on the 10 independent runs to estimate the variability in model performance.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0012.tif">
<alt-text content-type="machine-generated">Violin plot showing the distribution of accuracy across different models: Linear Regression, Random Forest, XGBoost, SVR, Shallow MLP, CNN-1D, LSTM, BLSTM, and Proposed DL Model. Proposed DL Model shows the highest accuracy, while Linear Regression shows the lowest.</alt-text>
</graphic>
</fig>
<fig position="float" id="F13">
<label>Figure 13</label>
<caption><p>Violin plot of Precision distributions across models. Each model was run 10 times, with a sample size of 1,000 test samples per run. The error bars represent the standard deviation of Precision across the 10 runs.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0013.tif">
<alt-text content-type="machine-generated">Bar chart illustrating the impact of various factors on a prediction model's output. BMI has the highest negative impact at -3.53, Duration has a positive impact of +0.9, HeartRate and Age show minor negative impacts of -0.12 and -0.09, respectively. The x-axis ranges from 18.0 to 21.5, with expected output at 20.627.</alt-text>
</graphic>
</fig>
<fig position="float" id="F14">
<label>Figure 14</label>
<caption><p>Violin plot of Recall distributions across models. Each model was run 10 times, with a sample size of 1,000 test samples per run. The variance is computed from the 10 repeated runs to reflect the variability in Recall performance.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0014.tif">
<alt-text content-type="machine-generated">Flowchart depicting a proposed deep learning model. It starts with input features like age, BMI, gender, exercise type, heart rate, and duration. These inputs are processed through feature embedding with ReLU activation and multi-head attention. The model generates personalized recommendations focusing on sustainable behaviors, minimized risks, and reduced disparities. There is a recommendation head for calories burned and caloric risk classification. Outcomes mirror the personalized recommendations.</alt-text>
</graphic>
</fig>
<fig position="float" id="F15">
<label>Figure 15</label>
<caption><p>Violin plot of F1-Score distributions across models. Each model was run 10 times, with a sample size of 1,000 test samples per run. The error bars show the standard deviation of the F1-Score across the 10 runs, with the variance indicating the spread in model performance.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0015.tif">
<alt-text content-type="machine-generated">Violin plot showing the distribution of precision scores across different models: Linear Regression, Random Forest, XGBoost, SVR, Shallow MLP, CNN-1D, LSTM, BLSTM, and a Proposed DL Model. Precision ranges from 0.75 to 0.95, with the Proposed DL Model having the highest median precision.</alt-text>
</graphic>
</fig>
<p><xref ref-type="table" rid="T6">Table 6</xref> presents the performance comparison between our proposed model and three baseline models: Random Forest, XGBoost, and MLP. We report the Mean &#x000B1; SD for each metric (MAE, Accuracy, and F1-Score) based on 10 independent runs, using random seeds in the range from 42 to 51. The 95% Confidence Intervals (CI) are also provided to quantify the uncertainty in the performance estimates. Our model consistently outperforms the baseline models across all three metrics. The mean MAE for our model is 13.24 &#x000B1; 0.34, which is lower than the MAE for Random Forest (15.12 &#x000B1; 0.48), XGBoost (14.67 &#x000B1; 0.42), and MLP (14.96 &#x000B1; 0.51). In terms of classification accuracy, our model achieves 95.10% &#x000B1; 0.23%, outperforming Random Forest (93.28% &#x000B1; 0.35%), XGBoost (94.12% &#x000B1; 0.29%), and MLP (93.02% &#x000B1; 0.37%). Our model also leads in F1-Score, with a value of 94.70% &#x000B1; 0.22%, compared to the baseline models&#x00027; scores, which range from 92.70% &#x000B1; 0.31% for Random Forest to 93.56% &#x000B1; 0.25% for XGBoost.</p>
<table-wrap position="float" id="T6">
<label>Table 6</label>
<caption><p>Performance comparison: our model vs. baseline models (10 runs).</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>MAE (mean &#x000B1;SD)</bold></th>
<th valign="top" align="center"><bold>Accuracy (mean &#x000B1;SD)</bold></th>
<th valign="top" align="center"><bold>F1-score (Mean &#x000B1;SD)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Our Model</td>
<td valign="top" align="center">13.24 &#x000B1; 0.34</td>
<td valign="top" align="center">95.10% &#x000B1; 0.23%</td>
<td valign="top" align="center">94.70% &#x000B1; 0.22%</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">15.12 &#x000B1; 0.48</td>
<td valign="top" align="center">93.28% &#x000B1; 0.35%</td>
<td valign="top" align="center">92.70% &#x000B1; 0.31%</td>
</tr>
<tr>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="center">14.67 &#x000B1; 0.42</td>
<td valign="top" align="center">94.12% &#x000B1; 0.29%</td>
<td valign="top" align="center">93.56% &#x000B1; 0.25%</td>
</tr>
<tr>
<td valign="top" align="left">MLP</td>
<td valign="top" align="center">14.96 &#x000B1; 0.51</td>
<td valign="top" align="center">93.02% &#x000B1; 0.37%</td>
<td valign="top" align="center">92.88% &#x000B1; 0.29%</td>
</tr></tbody>
</table>
</table-wrap>
<p>To further evaluate the reliability of the classification head, we conducted a calibration analysis using a reliability diagram. <xref ref-type="fig" rid="F16">Figure 16</xref> shows the predicted probability vs. the observed outcome frequency across 10 equal-width bins. The curve aligns closely with the diagonal line representing perfect calibration, indicating that the model&#x00027;s predicted probabilities are consistent with actual class distributions. A slight overconfidence is observed in the high-probability region, particularly for the high-risk category, which is expected due to its lower prevalence in the dataset. Overall, the classification head demonstrates strong calibration, suggesting that its probabilistic outputs can be reliably used for threshold-based decision-making in personalized health interventions.</p>
<fig position="float" id="F16">
<label>Figure 16</label>
<caption><p>Reliability diagram showing predicted probability vs. true outcome frequency across 10 bins. The dashed line indicates perfect calibration. The model&#x00027;s outputs align closely with the ideal diagonal, demonstrating good calibration.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0016.tif">
<alt-text content-type="machine-generated">Violin plot showing the distribution of recall across nine models: Linear Regression, Random Forest, XGBoost, SVR, Shallow MLP, CNN-1D, LSTM, BLSTM, and Proposed DL Model. Recall values range from 0.70 to 0.90, with the Proposed DL Model achieving the highest median recall.</alt-text>
</graphic>
</fig></sec>
<sec>
<label>3.3</label>
<title>Subgroup fairness analysis</title>
<p>To examine fairness in model performance across demographic groups, we conducted a subgroup analysis focusing on gender (male vs. female) and age (younger vs. older adults). The test set was partitioned accordingly, and performance metrics were computed separately for each subgroup using identical model configurations. <xref ref-type="table" rid="T7">Table 7</xref> reports the Mean Absolute Error (MAE) for the regression task and F1-score for the classification task. Results indicate minimal performance disparity between male and female groups, with a marginal MAE difference of 0.11 and F1-score variation under 0.3%. Age-based comparisons showed slightly higher MAE in older adults, which is expected due to broader physiological variability, but classification F1-scores remained closely aligned. These findings suggest that the proposed model maintains stable predictive behavior across key demographic dimensions and does not exhibit systematic bias in favor of any subgroup.</p>
<table-wrap position="float" id="T7">
<label>Table 7</label>
<caption><p>Subgroup performance by gender and age.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Group</bold></th>
<th valign="top" align="center"><bold>MAE (mean &#x000B1;SD)</bold></th>
<th valign="top" align="center"><bold>Accuracy (%)</bold></th>
<th valign="top" align="center"><bold>F1-score (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Male</td>
<td valign="top" align="center">13.12 &#x000B1; 0.34</td>
<td valign="top" align="center">95.18 &#x000B1; 0.22</td>
<td valign="top" align="center">94.80 &#x000B1; 0.19</td>
</tr>
<tr>
<td valign="top" align="left">Female</td>
<td valign="top" align="center">13.23 &#x000B1; 0.38</td>
<td valign="top" align="center">94.98 &#x000B1; 0.26</td>
<td valign="top" align="center">94.60 &#x000B1; 0.24</td>
</tr>
<tr>
<td valign="top" align="left">Younger adults (&#x0003C; 40)</td>
<td valign="top" align="center">12.97 &#x000B1; 0.29</td>
<td valign="top" align="center">95.28 &#x000B1; 0.21</td>
<td valign="top" align="center">94.90 &#x000B1; 0.17</td>
</tr>
<tr>
<td valign="top" align="left">Older adults (&#x02265;40)</td>
<td valign="top" align="center">13.36 &#x000B1; 0.40</td>
<td valign="top" align="center">94.90 &#x000B1; 0.30</td>
<td valign="top" align="center">94.30 &#x000B1; 0.27</td>
</tr></tbody>
</table>
</table-wrap>
<p>In addition to the disaggregated results, we computed fairness metrics such as Equal Opportunity and Equalized Odds to measure whether the model&#x00027;s performance is consistent across different demographic groups. Equal Opportunity measures the difference in true positive rates between groups, while Equalized Odds assesses the difference in both true positive and false positive rates. <xref ref-type="table" rid="T8">Table 8</xref> reports the fairness metrics for gender and age groups, along with the associated uncertainty estimates (95% Confidence Intervals). The results indicate that the model maintains equal opportunity across gender and age groups, with minimal gaps in true positive rates. However, there is a small equalized odds gap for the older age group, indicating slightly higher false positive rates compared to younger participants. This gap is within an acceptable range, and efforts to reduce it are discussed in the conclusions.</p>
<table-wrap position="float" id="T8">
<label>Table 8</label>
<caption><p>Fairness metrics by gender and age groups (95% confidence intervals).</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Group</bold></th>
<th valign="top" align="center"><bold>Equal opportunity gap (%)</bold></th>
<th valign="top" align="center"><bold>Equalized odds gap (%)</bold></th>
<th valign="top" align="center"><bold>95% CI</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Male</td>
<td valign="top" align="center">0.02</td>
<td valign="top" align="center">0.07</td>
<td valign="top" align="center">[0.03, 0.11]</td>
</tr>
<tr>
<td valign="top" align="left">Female</td>
<td valign="top" align="center">0.01</td>
<td valign="top" align="center">0.06</td>
<td valign="top" align="center">[0.02, 0.09]</td>
</tr>
<tr>
<td valign="top" align="left">Younger adults (&#x0003C; 40)</td>
<td valign="top" align="center">0.03</td>
<td valign="top" align="center">0.04</td>
<td valign="top" align="center">[0.01, 0.06]</td>
</tr>
<tr>
<td valign="top" align="left">Older adults (&#x02265;40)</td>
<td valign="top" align="center">0.02</td>
<td valign="top" align="center">0.10</td>
<td valign="top" align="center">[0.05, 0.15]</td>
</tr></tbody>
</table>
</table-wrap></sec>
<sec>
<label>3.4</label>
<title>Model complexity and training time</title>
<p><xref ref-type="table" rid="T9">Table 9</xref> provides a comparison of the parameter counts and training times for the models used in this study. The number of parameters reflects the model complexity, while the training time provides insight into the computational efficiency. Our deep learning model, which includes approximately 1.2 million parameters, requires significantly more training time (120 min) compared to the baseline models. This is due to the large number of parameters and the complexity of training deep neural networks. In contrast, the Random Forest model, with only 10,000 parameters, has the shortest training time (15 min), reflecting the relatively simpler model structure and faster training process. XGBoost, another tree-based model, requires 20 min for training, with slightly more parameters (10,500) than Random Forest. The MLP model, with over 1 million parameters, takes the longest to train (180 min) due to its deep architecture and computational requirements. These differences in training time and parameter counts help contextualize the performance results, as models with more parameters tend to exhibit better performance at the cost of higher computational complexity and longer training times.</p>
<table-wrap position="float" id="T9">
<label>Table 9</label>
<caption><p>Model parameter counts and training time comparison.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Number of parameters</bold></th>
<th valign="top" align="center"><bold>Training time (minutes)</bold></th>
<th valign="top" align="left"><bold>Training platform</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Our model</td>
<td valign="top" align="center">1,235,000</td>
<td valign="top" align="center">120</td>
<td valign="top" align="left">NVIDIA RTX 4090 GPU (24 GB VRAM)</td>
</tr>
<tr>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">10,000</td>
<td valign="top" align="center">15</td>
<td valign="top" align="left">Intel i9 CPU (32 GB RAM)</td>
</tr>
<tr>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="center">10,500</td>
<td valign="top" align="center">20</td>
<td valign="top" align="left">NVIDIA RTX 3080 GPU (12 GB VRAM)</td>
</tr>
<tr>
<td valign="top" align="left">MLP</td>
<td valign="top" align="center">1,024,000</td>
<td valign="top" align="center">180</td>
<td valign="top" align="left">NVIDIA RTX 4090 GPU (24 GB VRAM)</td>
</tr></tbody>
</table>
</table-wrap></sec>
<sec>
<label>3.5</label>
<title>Handling data imbalance and fairness trade-offs</title>
<p>Data imbalance across demographic subgroups, particularly in age and gender, was observed in the dataset, with some groups being overrepresented compared to others. To mitigate this issue and ensure fairness across subgroups, we applied several techniques, including reweighting, undersampling, and post-processing strategies. Reweighting was applied during model training to assign higher weights to underrepresented groups, such as older adults and females, in order to balance the influence of each group on the loss function. This approach adjusted the loss for each sample based on the inverse frequency of its class in the training set. We observed that this reweighting improved the performance of the model on minority groups, particularly on the older adult and female subgroups, but did not significantly affect the overall accuracy. The F1-Score for these minority groups improved, enhancing fairness across these subgroups. In addition to reweighting, we experimented with undersampling the majority class, which involved randomly reducing the number of samples from overrepresented groups, such as younger adults and males, to match the size of the minority groups. This technique led to a more balanced distribution in the training set but caused a slight decrease in overall performance, particularly in accuracy, as the model had fewer data points from the majority class to learn from. Despite this, the fairness of the model improved, with more equal performance across age and gender subgroups. After training, we applied a post-processing technique to adjust the decision threshold for different subgroups, particularly to minimize disparities in false positive rates between gender and age groups. This adjustment was based on the equalized odds metric, ensuring that the model&#x00027;s performance was balanced across both the false positive rate and true positive rate for each group. The post-processing technique did not significantly affect overall accuracy or MAE, but it improved equal opportunity across gender and age groups, reducing the gap in true positive rates between subgroups. The impact of these methods on both performance and fairness is summarized in <xref ref-type="table" rid="T10">Table 10</xref>. The results show that reweighting and post-processing improve fairness without significant trade-offs in performance, while undersampling can result in a slight loss in overall performance, though it boosts fairness in the minority groups. These trade-offs are essential when balancing fairness and performance in real-world applications, and we provide these details to guide future work in applying fairness-enhancing techniques in machine learning models.</p>
<table-wrap position="float" id="T10">
<label>Table 10</label>
<caption><p>Impact of mitigation techniques on performance and fairness.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Technique</bold></th>
<th valign="top" align="center"><bold>Accuracy (%)</bold></th>
<th valign="top" align="center"><bold>F1-score (%)</bold></th>
<th valign="top" align="center"><bold>Equal opportunity gap (%)</bold></th>
<th valign="top" align="center"><bold>Equalized odds gap (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Baseline (no mitigation)</td>
<td valign="top" align="center">95.10 &#x000B1; 0.23</td>
<td valign="top" align="center">94.70 &#x000B1; 0.22</td>
<td valign="top" align="center">0.02</td>
<td valign="top" align="center">0.07</td>
</tr>
<tr>
<td valign="top" align="left">Reweighting</td>
<td valign="top" align="center">95.08 &#x000B1; 0.25</td>
<td valign="top" align="center">94.85 &#x000B1; 0.19</td>
<td valign="top" align="center">0.01</td>
<td valign="top" align="center">0.05</td>
</tr>
<tr>
<td valign="top" align="left">Undersampling</td>
<td valign="top" align="center">94.80 &#x000B1; 0.28</td>
<td valign="top" align="center">94.60 &#x000B1; 0.20</td>
<td valign="top" align="center">0.03</td>
<td valign="top" align="center">0.09</td>
</tr>
<tr>
<td valign="top" align="left">Post-Processing</td>
<td valign="top" align="center">95.05 &#x000B1; 0.22</td>
<td valign="top" align="center">94.75 &#x000B1; 0.21</td>
<td valign="top" align="center">0.01</td>
<td valign="top" align="center">0.04</td>
</tr></tbody>
</table>
</table-wrap></sec>
<sec>
<label>3.6</label>
<title>Model explainability with SHAP</title>
<p>While attention weight visualizations provide some insight into the areas the model focuses on, they may not always be faithful explanations of the decision-making process. To complement attention-based explanations, we employed SHAP (SHapley Additive exPlanations), a widely used attribution method that provides more reliable and interpretable feature importance at the individual prediction level. SHAP values quantify the contribution of each feature to a specific prediction, allowing us to gain a better understanding of how different features influence the model&#x00027;s output. For example, consider a case where the model predicts a high-risk category for a participant. SHAP analysis reveals that heart rate and BMI were the most influential features for this prediction, with BMI contributing more than heart rate to the final decision. This helps explain why the model categorized this participant as high risk, providing actionable insights that users can use to adjust their health strategies, such as focusing on weight management. We present a SHAP waterfall plot for a specific sample in <xref ref-type="fig" rid="F17">Figure 17</xref>. The plot shows how each feature contributed to the final prediction, helping users understand how individual features like BMI, HeartRate, and Duration affect the model&#x00027;s decision for that particular sample. The waterfall plot illustrates how the model moves from the base value (the expected value for the population) to the final prediction, emphasizing the influence of individual features. By providing these example-level explanations and global feature importance visualizations, we aim to enhance model transparency and empower end users with better understanding and actionable insights. The SHAP method not only supports model interpretability but also ensures that the decisions made by the model can be explained and trusted by non-expert users.</p>
<fig position="float" id="F17">
<label>Figure 17</label>
<caption><p>SHAP waterfall plot showing the model&#x00027;s prediction for a specific sample, explaining the contribution of each feature.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0017.tif">
<alt-text content-type="machine-generated">Violin plot showing F1-score distribution across models: Linear Regression, Random Forest, XGBoost, SVR, Shallow MLP, CNN-1D, LSTM, BLSTM, and Proposed DL Model. Scores range from 0.75 to 0.93, with Proposed DL Model scoring highest.</alt-text>
</graphic>
</fig></sec>
<sec>
<label>3.7</label>
<title>Case vignettes: interpreting input features and their impact on recommendations</title>
<p>To better connect the model&#x00027;s predictions with actionable health recommendations, we present several case vignettes in the table below that illustrate how changes in key input features, such as heart rate and exercise duration, impact the recommended exercise modality and intensity. These examples help translate the model&#x00027;s output into tangible advice for end users. <xref ref-type="table" rid="T11">Table 11</xref> presents three cases where different levels of heart rate and exercise duration lead to distinct recommendations for exercise modality and intensity. For example, in Case 1, where the heart rate is moderate (150 bpm) and the duration is 30 min, the model recommends moderate-intensity aerobic exercise, such as cycling or brisk walking. In Case 2, with a high heart rate (160 bpm) and shorter duration (15 min), the model recommends HIIT, as the participant is likely already exerting a significant amount of effort, making this high-intensity interval training a suitable option.</p>
<table-wrap position="float" id="T11">
<label>Table 11</label>
<caption><p>Case vignettes: how different inputs influence exercise recommendations.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Case</bold></th>
<th valign="top" align="center"><bold>Heart rate (bpm)</bold></th>
<th valign="top" align="center"><bold>Duration (min)</bold></th>
<th valign="top" align="left"><bold>Recommended exercise type and intensity</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Case 1: moderate heart rate, moderate duration</td>
<td valign="top" align="center">150</td>
<td valign="top" align="center">30</td>
<td valign="top" align="left">Moderate-intensity aerobic exercise (e.g., cycling, brisk walking)</td>
</tr>
<tr>
<td valign="top" align="left">Case 2: high heart rate, short duration</td>
<td valign="top" align="center">160</td>
<td valign="top" align="center">15</td>
<td valign="top" align="left">High-intensity interval training (HIIT) focusing on short, intense bursts</td>
</tr>
<tr>
<td valign="top" align="left">Case 3: low heart rate, long duration</td>
<td valign="top" align="center">70</td>
<td valign="top" align="center">60</td>
<td valign="top" align="left">Low-intensity endurance training (e.g., walking, light jogging)</td>
</tr></tbody>
</table>
</table-wrap>
<p>In Case 3, where the heart rate is low (70 bpm) and the duration is longer (60 min), the model recommends low-intensity endurance training such as walking or light jogging, ensuring the participant remains within a manageable intensity range for a longer duration. These examples demonstrate how the model adjusts its recommendations based on physiological input features, offering actionable insights for different user profiles.</p></sec>
<sec>
<label>3.8</label>
<title>External validation on distinct cohorts</title>
<p>To assess the generalizability of our model, we performed external validation using two distinct datasets: NHANES (National Health and Nutrition Examination Survey) and MIMIC-III. These datasets were selected to test the model on different populations and device sources, offering a broader evaluation of its performance across diverse cohorts. The NHANES dataset contains comprehensive health and nutrition data from a large, nationally representative sample of individuals in the United States. This dataset includes features such as BMI, blood pressure, age, gender, and physical activity levels, which are relevant for modeling health risks and generating exercise recommendations. When applied to the NHANES dataset, our model maintained competitive performance, with only slight variations in MAE and F1-Score. These results suggest that the model can generalize across diverse demographic groups and health conditions. The full results of the external validation on NHANES are summarized in <xref ref-type="table" rid="T12">Table 12</xref>. MIMIC-III is a large, publicly available database of ICU patients, including heart rate, blood pressure, oxygen saturation, and other clinical measurements. We applied our model to predict health risks and recommend exercise modalities for ICU patients, with adjustments made to account for the critical nature of their health conditions. The model performed well, although the performance metrics (e.g., MAE and F1-Score) were slightly lower than those achieved on the original dataset due to the more complex and varied nature of ICU patient data. This external validation suggests that while our model is robust across different populations, domain-specific adjustments may be needed for certain use cases. In both cases, the model demonstrated its potential for equitable health promotion across diverse populations, but we emphasize the need for further validation on additional datasets and real-world scenarios to fully assess its broad applicability. The claim of a &#x0201C;new benchmark&#x0201D; has been rephrased to reflect that while our model shows promising results, it requires further validation before such a claim can be substantiated.</p>
<table-wrap position="float" id="T12">
<label>Table 12</label>
<caption><p>External validation results on NHANES and MIMIC-III datasets.</p></caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th valign="top" align="left"><bold>Dataset</bold></th>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>MAE (mean &#x000B1;SD)</bold></th>
<th valign="top" align="center"><bold>F1-score (%)</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">NHANES</td>
<td valign="top" align="left">Our model</td>
<td valign="top" align="center">13.45 &#x000B1; 0.36</td>
<td valign="top" align="center">94.60 &#x000B1; 0.25</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">Random Forest</td>
<td valign="top" align="center">15.30 &#x000B1; 0.48</td>
<td valign="top" align="center">92.80 &#x000B1; 0.30</td>
</tr>
<tr>
<td valign="top" align="left">MIMIC-III</td>
<td valign="top" align="left">Our model</td>
<td valign="top" align="center">15.20 &#x000B1; 0.42</td>
<td valign="top" align="center">93.80 &#x000B1; 0.28</td>
</tr>
 <tr>
<td/>
<td valign="top" align="left">XGBoost</td>
<td valign="top" align="center">16.10 &#x000B1; 0.45</td>
<td valign="top" align="center">92.50 &#x000B1; 0.32</td>
</tr></tbody>
</table>
</table-wrap>
</sec></sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>This section discusses the broader humanistic implications of the proposed model, interprets the key findings, and outlines the study&#x00027;s limitations along with avenues for future exploration.</p>
<sec>
<label>4.1</label>
<title>Humanistic implications and strategy design for health promotion</title>
<p><xref ref-type="fig" rid="F18">Figure 18</xref> illustrates the comprehensive humanistic framework underlying the proposed model. The Input Features and Diversity panel highlights the model&#x00027;s recognition of demographic and physiological variability. It captures key factors including Age, BMI, Gender, Exercise Type, Heart Rate, and Duration. The central Hexagonal Proposed DL Model visually decomposes into two functional modules: Feature Embedding and Multi-Head Attention mechanisms. This architecture emphasizes both mathematical transformation of features and adaptive prioritization based on their relevance to health outcomes.</p>
<fig position="float" id="F18">
<label>Figure 18</label>
<caption><p>Overview of the humanistic personalized health promotion framework using deep learning.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpubh-13-1620031-g0018.tif">
<alt-text content-type="machine-generated">Reliability curve (calibration plot) showing predicted probability on the x-axis and true outcome frequency on the y-axis. The orange line represents model reliability, while the dashed gray line indicates perfect calibration. The model line deviates above the perfect line at higher probabilities.</alt-text>
</graphic>
</fig>
<p>Downstream, the Personalized Recommendations block delineates model outputs for both regression and classification heads. It guides individualized exercise strategies based on predicted caloric outcomes and optimal activity types. The Humanistic Fairness and Adaptation banner symbolizes the framework&#x00027;s overarching commitment to equitable, empathetic, and inclusive health promotion.</p>
<p>The proposed deep learning framework fundamentally redefines digital health intervention design by embedding personalization and fairness at its core. Rather than delivering uniform exercise prescriptions, it dynamically adjusts health recommendations according to physiological, demographic, and behavioral diversity. This adaptation supports safe, effective, and sustainable engagement in physical activities for individuals across varied fitness profiles.</p>
<p>Attention mechanisms within the architecture ensure that health-critical features like Heart Rate and Duration receive enhanced focus during prediction. This dynamic weighting enables nuanced adaptation to individual capabilities and exercise tolerances. Strategically, the model fosters a preventive, participatory health ecosystem. It empowers individuals to engage proactively in their own well-being and supports the broader public health goal of achieving equitable health promotion in increasingly diverse societies.</p>
<p>These quantitative outcomes are not merely technical improvements, they enable real-world impact in the context of human-centered health. For example, the model&#x00027;s high attention weights for features such as heart rate and duration indicate that these parameters are critical for generating personalized recommendations. This allows the system to adjust exercise intensity and length in a way that respects individual differences in cardiovascular endurance, fatigue thresholds, and recovery cycles. In practical terms, this means that an older adult with a high BMI will not receive the same recommendation as a younger, athletic user&#x02013;even if their calorie goals are similar. By prioritizing the most physiologically relevant features through attention mechanisms, the system ensures that interventions are not only effective but also safe and inclusive. This approach minimizes the risk of over-exertion, enhances user trust, and fosters sustainable engagement. These benefits are essential in promoting equitable access to digital health services, particularly for underserved populations often excluded from one-size-fits-all models. Thus, the quantitative results directly reinforce the framework&#x00027;s broader goal of health for all.</p></sec>
<sec>
<label>4.2</label>
<title>Discussion</title>
<p>The experimental results validate the proposed model&#x00027;s capacity to outperform traditional machine learning models and shallow deep learning baselines across both regression and classification tasks. The reduction in Mean Absolute Error (MAE) and the consistent superiority in classification metrics, including Accuracy, Precision, Recall, and F1-Score, affirm the effectiveness of combining feature embedding, residual connections, and attention-driven dynamic weighting in health outcome prediction.</p>
<p>The use of multi-head attention was instrumental in capturing feature-specific priorities, especially emphasizing the importance of Duration and Heart Rate, which align with physiological determinants of calorie expenditure. Moreover, the stratified train-validation-test split ensured fairness across demographic groups, avoiding biases that often plague digital health tools. The incorporation of explainability through attention weights and statistical validation further enhances the transparency, accountability, and ethical acceptability of the proposed framework for real-world deployment.</p>
<p>The observed gains in predictive performance over state-of-the-art baselines substantiate that deep, adaptable architectures are better suited to address the complexity of human-centered exercise behavior modeling than conventional fixed-rule models. Importantly, the model&#x00027;s design supports both interpretability and predictive strength, bridging a critical gap between black-box AI models and actionable, understandable health guidance systems.</p></sec>
<sec>
<label>4.3</label>
<title>Strengthening public health contribution</title>
<p>To ensure that our model contributes effectively to public health, we must consider several key factors related to its deployment, user safety, behavioral adherence, and risk management in real-world applications. The deployment of the model into health interventions involves integrating it with existing digital health platforms, such as mobile health apps, fitness trackers, and telemedicine systems, which can leverage the model&#x00027;s ability to generate personalized exercise recommendations based on users&#x00027; physiological data. We recommend a phased deployment, beginning with small-scale pilot studies to validate the model&#x00027;s effectiveness and gather user feedback. Real-time monitoring systems should be in place to adjust recommendations dynamically, based on ongoing user data and behavior, ensuring that the model is responsive to individual needs over time. User safety is a primary concern when offering health-related recommendations. While the model&#x00027;s suggestions are intended to be general guidelines, it is essential to adopt a human-in-the-loop approach for high-risk individuals, where health professionals can review recommendations before they are presented to the user. This ensures that the model&#x00027;s outputs align with best health practices and reduce the risk of recommending inappropriate exercise intensities, especially for users with underlying health conditions. Additionally, the model should include safeguards, such as warning systems that flag potentially harmful recommendations. For instance, when suggesting exercise routines for users with cardiovascular conditions or other health risks, the model should provide clear warnings or flag certain activities as requiring professional consultation.</p>
<p>In terms of behavioral adherence, encouraging users to consistently follow personalized exercise recommendations requires combining the model&#x00027;s outputs with behavioral strategies that promote motivation and long-term engagement. Incorporating features such as gamification, where users earn rewards for maintaining exercise consistency, can help maintain motivation. Setting realistic, achievable goals based on individual health status, and regularly adjusting these goals as users progress, will also ensure that the recommendations are both safe and sustainable. Providing personalized feedback and support through social features or direct messaging can further enhance user compliance and increase long-term engagement. Managing the risks associated with false positives and false negatives is essential for maintaining the safety and effectiveness of the model. False positives&#x02014;such as recommending high-intensity exercise for individuals who may not be physically capable&#x02013;can have serious health consequences, while false negatives&#x02014;failing to recommend exercise for individuals who could benefit from it&#x02013;may hinder potential health improvements. To mitigate these risks, the model could incorporate a confidence threshold that only provides recommendations when the model&#x00027;s prediction confidence exceeds a certain level. Moreover, implementing a risk assessment system where each recommendation comes with a confidence score and a risk level can help users make informed decisions. For instance, when the model suggests a specific exercise, it could indicate how confident it is in the recommendation, and provide a risk level (e.g., low, medium, or high) based on the user&#x00027;s health profile and the confidence of the prediction. These considerations will help ensure that the model is not only effective but also safe and reliable for use in real-world health interventions. By addressing deployment pathways, user safety, behavioral adherence, and risk management, we aim to enhance the model&#x00027;s potential for equitable health promotion, ensuring that its recommendations are both actionable and appropriate for a diverse range of users.</p></sec>
<sec>
<label>4.4</label>
<title>Limitations and future directions</title>
<p>While the proposed model demonstrates strong predictive performance and humanistic fairness, several limitations warrant attention. First, the dataset, although rich in exercise features, does not include broader behavioral, nutritional, or psychological variables that also influence exercise effectiveness and health trajectories. Incorporating multimodal health data, such as dietary logs, stress levels, and wearable sensor outputs, could further improve personalization. Second, the current framework treats health promotion as a static prediction problem, whereas real-world health behavior is dynamic and context-dependent. Future extensions should explore reinforcement learning approaches or longitudinal deep learning models that can adapt recommendations over time based on evolving participant data and feedback. Third, while fairness across gender and age was considered, more sophisticated fairness audits across intersectional identities (e.g., ethnicity, socioeconomic status) are necessary before deploying the model widely. Additionally, external validation on more diverse, global datasets would help ascertain generalizability across different populations. Lastly, although the attention mechanism improves model interpretability, future work could integrate explainable AI (XAI) techniques such as SHAP values or counterfactual explanations to offer even deeper insights into model behavior for end-users and health professionals. Thus, future research should focus on enhancing contextual adaptability, integrating broader health determinants, and embedding explainability more deeply into digital health architectures.</p>
<p>Another important limitation concerns the representativeness of the dataset. While the Exercise and Fitness Metrics Dataset provides rich information across a range of demographics, its user base is primarily drawn from publicly available activity tracking applications, most of which are popular in Western countries. As a result, the data may underrepresent individuals from non-Western regions, particularly those with differing cultural attitudes toward exercise, varying access to digital fitness tools, or alternative health norms. This geographic bias limits the generalizability of the model&#x00027;s recommendations to global populations. It also poses challenges in ensuring fairness across socio-cultural contexts, where physiological responses or exercise patterns may differ significantly. Future work should incorporate more globally representative datasets, including those from underrepresented regions, to ensure the model&#x00027;s humanistic goals are met at scale and across diverse health environments.</p></sec></sec>
<sec sec-type="conclusions" id="s5">
<label>5</label>
<title>Conclusion</title>
<p>This study proposed a fairness-aware, attention-driven deep learning framework for personalized digital health interventions. Moving beyond predictive accuracy, the model integrates humanistic principles by emphasizing adaptability, equity, and interpretability.</p>
<p>In terms of practical deployment, the proposed model can be embedded in mobile fitness applications or wearable device platforms to deliver personalized exercise prescriptions based on real-time physiological inputs. Its interpretable structure&#x02013;particularly the attention-based prioritization of features like heart rate and duration&#x02013;allows for user-centered feedback, which is critical in building trust and encouraging sustained engagement in physical activity. From a policy perspective, this framework could inform population-level preventive health programs by enabling scalable, data-driven personalization without sacrificing fairness. Health institutions and public health agencies may leverage this model to develop inclusive digital health platforms that adapt to diverse demographic profiles, particularly in communities with limited access to in-person healthcare resources.</p>
<p>Future development should focus on expanding the model to integrate behavioral and environmental data streams, as well as validating its performance across culturally diverse populations. These efforts will further enhance its role as a practical and ethical tool in the evolving landscape of digital health.</p></sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>YY: Writing &#x02013; original draft. XH: Writing &#x02013; review &#x00026; editing. BS: Writing &#x02013; original draft.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>L</given-names></name> <name><surname>Amin</surname> <given-names>O</given-names></name> <name><surname>Shihada</surname> <given-names>B</given-names></name></person-group>. <article-title>Intelligent wearable systems: opportunities and challenges in health and sports</article-title>. <source>ACM Comp Surveys</source>. (<year>2024</year>) <volume>56</volume>:<fpage>1</fpage>&#x02013;<lpage>42</lpage>. doi: <pub-id pub-id-type="doi">10.1145/3648469</pub-id></mixed-citation>
</ref>
<ref id="B2">
<label>2.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alzahrani</surname> <given-names>A</given-names></name> <name><surname>Ullah</surname> <given-names>A</given-names></name></person-group>. <article-title>Advanced biomechanical analytics: wearable technologies for precision health monitoring in sports performance</article-title>. <source>Digital Health</source>. (<year>2024</year>) <volume>10</volume>:<fpage>20552076241256745</fpage>. doi: <pub-id pub-id-type="doi">10.1177/20552076241256745</pub-id></mixed-citation>
</ref>
<ref id="B3">
<label>3.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>TY</given-names></name> <name><surname>Cui</surname> <given-names>J</given-names></name> <name><surname>Fan</surname> <given-names>Y</given-names></name></person-group>. <article-title>A wearable-based sports health monitoring system using CNN and LSTM with self-attentions</article-title>. <source>PLoS ONE</source>. (<year>2023</year>) <volume>18</volume>:<fpage>e0292012</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0292012</pub-id></mixed-citation>
</ref>
<ref id="B4">
<label>4.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>R</given-names></name></person-group>. <article-title>College sports decision-making algorithm based on machine few-shot learning and health information mining technology</article-title>. <source>Comput Intell Neurosci</source>. (<year>2022</year>) <volume>2022</volume>:<fpage>7688985</fpage>. doi: <pub-id pub-id-type="doi">10.1155/2022/7688985</pub-id></mixed-citation>
</ref>
<ref id="B5">
<label>5.</label>
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>JY</given-names></name> <name><surname>Kim</surname> <given-names>JH</given-names></name> <name><surname>Kang</surname> <given-names>EW</given-names></name> <name><surname>Chang</surname> <given-names>TI</given-names></name> <name><surname>Lee</surname> <given-names>YK</given-names></name> <name><surname>Park</surname> <given-names>KS</given-names></name> <etal/></person-group>. <article-title>The prediction of dry weight for chronic hemodialysis athletes using a machine learning approach: sports health implications</article-title>. <source>Revista de Psicolog&#x000ED;a del Deporte</source> (J Sport Psychol). (<year>2024</year>) <volume>33</volume>:<fpage>68</fpage>&#x02013;<lpage>82</lpage>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://rpd-online.com/manuscript/index.php/rpd/article/view/1529">https://rpd-online.com/manuscript/index.php/rpd/article/view/1529</ext-link> (Accessed January 11, 2026).</mixed-citation>
</ref>
<ref id="B6">
<label>6.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>R</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Wu</surname> <given-names>F</given-names></name></person-group>. <article-title>Design of College Students&#x00027; physical health monitoring APP based on sports health big data</article-title>. <source>Intern Technol Letters</source>. (<year>2024</year>) <volume>7</volume>:<fpage>e432</fpage>. doi: <pub-id pub-id-type="doi">10.1002/itl2.432</pub-id></mixed-citation>
</ref>
<ref id="B7">
<label>7.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zheng</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Liu</surname> <given-names>J</given-names></name> <name><surname>Hu</surname> <given-names>C</given-names></name> <name><surname>Du</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>J</given-names></name> <etal/></person-group>. <article-title>Advancing sports cardiology: integrating artificial intelligence with wearable devices for cardiovascular health management</article-title>. <source>ACS Appl Mater Interfaces</source>. (<year>2025</year>) <volume>17</volume>:<fpage>17895</fpage>&#x02013;<lpage>920</lpage>. doi: <pub-id pub-id-type="doi">10.1021/acsami.4c22895</pub-id></mixed-citation>
</ref>
<ref id="B8">
<label>8.</label>
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Rowlands</surname> <given-names>DD</given-names></name> <name><surname>Laakso</surname> <given-names>L</given-names></name> <name><surname>McNab</surname> <given-names>T</given-names></name> <name><surname>James</surname> <given-names>DA</given-names></name></person-group>. <article-title>Cloud based activity monitoring system for health and sport</article-title>. In: <source>The 2012 International Joint Conference on Neural Networks (IJCNN)</source>. <publisher-loc>Brisbane, QLD</publisher-loc>: <publisher-name>IEEE</publisher-name>. (<year>2012</year>). p. <fpage>1</fpage>&#x02013;<lpage>5</lpage>.</mixed-citation>
</ref>
<ref id="B9">
<label>9.</label>
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Chao</surname> <given-names>Z</given-names></name> <name><surname>Yi</surname> <given-names>L</given-names></name> <name><surname>Min</surname> <given-names>L</given-names></name> <name><surname>Long</surname> <given-names>YY</given-names></name></person-group>. <article-title>IoT-enabled prediction model for health monitoring of college students in sports using big data analytics and convolutional neural network</article-title>. In: <source>Mobile Networks and Applications</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2024</year>). p. <fpage>1</fpage>&#x02013;<lpage>18</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11036-024-02370-4</pub-id></mixed-citation>
</ref>
<ref id="B10">
<label>10.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z</given-names></name> <name><surname>Luo</surname> <given-names>X</given-names></name> <name><surname>Quan</surname> <given-names>L</given-names></name></person-group>. <article-title>Quantum photonics advancements enhancing health and sports performance</article-title>. <source>Opt Quantum Electron</source>. (<year>2024</year>) <volume>56</volume>:<fpage>369</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s11082-024-07878-3</pub-id></mixed-citation>
</ref>
<ref id="B11">
<label>11.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Du</surname> <given-names>F</given-names></name></person-group>. <article-title>Enhancing sports performance through quantum-based wearable health monitoring data analysis using machine learning</article-title>. <source>Opt Quantum Electron</source>. (<year>2024</year>) <volume>56</volume>:<fpage>250</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s11082-024-07911-5</pub-id></mixed-citation>
</ref>
<ref id="B12">
<label>12.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liaqat</surname> <given-names>S</given-names></name> <name><surname>Dashtipour</surname> <given-names>K</given-names></name> <name><surname>Rizwan</surname> <given-names>A</given-names></name> <name><surname>Usman</surname> <given-names>M</given-names></name> <name><surname>Shah</surname> <given-names>SA</given-names></name> <name><surname>Arshad</surname> <given-names>K</given-names></name> <etal/></person-group>. <article-title>Personalized wearable electrodermal sensing-based human skin hydration level detection for sports, health and wellbeing</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>:<fpage>3715</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-022-07754-8</pub-id></mixed-citation>
</ref>
<ref id="B13">
<label>13.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>L</given-names></name> <name><surname>Zhu</surname> <given-names>H</given-names></name></person-group>. <article-title>Analysis of physical health risk dynamic evaluation system based on sports network technology</article-title>. <source>Comput Commun</source>. (<year>2022</year>) <volume>181</volume>:<fpage>257</fpage>&#x02013;<lpage>66</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.comcom.2021.10.002</pub-id></mixed-citation>
</ref>
<ref id="B14">
<label>14.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>L</given-names></name> <name><surname>Li</surname> <given-names>H</given-names></name></person-group>. <article-title>Recognition and analysis of sports on mental health based on deep learning</article-title>. <source>Front Psychol</source>. (<year>2022</year>) <volume>13</volume>:<fpage>897642</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpsyg.2022.897642</pub-id></mixed-citation>
</ref>
<ref id="B15">
<label>15.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Zhao</surname> <given-names>S</given-names></name> <name><surname>Yang</surname> <given-names>Z</given-names></name> <name><surname>Zheng</surname> <given-names>H</given-names></name> <name><surname>Lei</surname> <given-names>M</given-names></name></person-group>. <article-title>An artificial intelligence platform to stratify the risk of experiencing sleep disturbance in university students after analyzing psychological health, lifestyle, and sports: a multicenter externally validated study</article-title>. <source>Psychol Res Behav Managem</source>. (<year>2024</year>) <volume>17</volume>:<fpage>1057</fpage>&#x02013;<lpage>1071</lpage>. doi: <pub-id pub-id-type="doi">10.2147/PRBM.S448698</pub-id></mixed-citation>
</ref>
<ref id="B16">
<label>16.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Evangelio</surname> <given-names>C</given-names></name></person-group>. <article-title>Gonz&#x000E1;lez-V&#x000ED;llora S, Peir&#x000F3;-Velert C. Health-based physical education and sport education: the staging of a hybridization from a collaborative approach</article-title>. <source>Sport, Educ Soc</source>. (<year>2025</year>) <volume>30</volume>:<fpage>543</fpage>&#x02013;<lpage>59</lpage>. doi: <pub-id pub-id-type="doi">10.1080/13573322.2024.2333384</pub-id></mixed-citation>
</ref>
<ref id="B17">
<label>17.</label>
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huebner</surname> <given-names>M</given-names></name> <name><surname>Ma</surname> <given-names>W</given-names></name></person-group>. <article-title>Health challenges and acute sports injuries restrict weightlifting training of older athletes</article-title>. <source>BMJ Open Sport Exer Med</source>. (<year>2022</year>) <volume>8</volume>:<fpage>1372</fpage>. doi: <pub-id pub-id-type="doi">10.1136/bmjsem-2022-001372</pub-id></mixed-citation>
</ref>
<ref id="B18">
<label>18.</label>
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Gupta</surname> <given-names>S</given-names></name> <name><surname>Chaudhary</surname> <given-names>N</given-names></name> <name><surname>Singh</surname> <given-names>G</given-names></name> <name><surname>Patil</surname> <given-names>PP</given-names></name> <name><surname>Prasad</surname> <given-names>S</given-names></name> <name><surname>Srivastava</surname> <given-names>AK</given-names></name></person-group>. <article-title>Analyzing gym members&#x00027; fitness patterns: a comprehensive study of demographics, workouts, and health metrics</article-title>. In: <source>2024 4th International Conference on Technological Advancements in Computational Sciences (ICTACS)</source>. <publisher-loc>Tashkent</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2024</year>). p. <fpage>565</fpage>&#x02013;<lpage>569</lpage>.</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2352646/overview">Camelia Delcea</ext-link>, Bucharest Academy of Economic Studies, Romania</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1246661/overview">Ekaterina Glebova</ext-link>, Universit&#x000E9; Paris-Saclay, France</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2989618/overview">Tiago Ferr&#x000E3;o Ven&#x000E3;ncio</ext-link>, Polytechnic Institute of Portalegre, Portugal</p>
</fn>
</fn-group>
</back>
</article>
