<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Genet.</journal-id>
<journal-title>Frontiers in Genetics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Genet.</abbrev-journal-title>
<issn pub-type="epub">1664-8021</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1538544</article-id>
<article-id pub-id-type="doi">10.3389/fgene.2025.1538544</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Genetics</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Auto-branch multi-task learning for simultaneous prediction of multiple correlated traits associated with Alzheimer&#x2019;s disease</article-title>
<alt-title alt-title-type="left-running-head">Liang et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fgene.2025.1538544">10.3389/fgene.2025.1538544</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Liang</surname>
<given-names>Jiaqi</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2805859/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Xue</surname>
<given-names>Zhao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>&#x2020;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2948698/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhou</surname>
<given-names>Wenchao</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2336398/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Guo</surname>
<given-names>Xiangjie</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1995275/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wen</surname>
<given-names>Yalu</given-names>
</name>
<xref ref-type="aff" rid="aff5">
<sup>5</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/311671/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/Writing - review &#x26; editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Academy of Medical Sciences</institution>, <institution>Shanxi Medical University</institution>, <addr-line>Taiyuan</addr-line>, <addr-line>Shanxi</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Department of Health Statistics</institution>, <institution>School of Public Health</institution>, <institution>Shanxi Medical University</institution>, <addr-line>Taiyuan</addr-line>, <addr-line>Shanxi</addr-line>, <country>China</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Translational Medicine Research Center</institution>, <institution>Shanxi Medical University</institution>, <addr-line>Taiyuan</addr-line>, <addr-line>Shanxi</addr-line>, <country>China</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>School of Forensic Medicine</institution>, <institution>Shanxi Medical University</institution>, <addr-line>Jinzhong</addr-line>, <addr-line>Shanxi</addr-line>, <country>China</country>
</aff>
<aff id="aff5">
<sup>5</sup>
<institution>Department of Statistics</institution>, <institution>University of Auckland</institution>, <addr-line>Auckland</addr-line>, <country>New Zealand</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/43295/overview">Ettore Mosca</ext-link>, National Research Council (CNR), Italy</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/215357/overview">Alessandro Orro</ext-link>, National Research Council (CNR), Italy</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3002453/overview">Nicolas Derus</ext-link>, University of Bologna, Italy</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Yalu Wen, <email>y.wen@auckland.ac.nz</email>; Xiangjie Guo, <email>13994207779@163.com</email>
</corresp>
<fn fn-type="equal" id="fn001">
<label>
<sup>&#x2020;</sup>
</label>
<p>These authors have contributed equally to this work and share first authorship</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>10</day>
<month>06</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>16</volume>
<elocation-id>1538544</elocation-id>
<history>
<date date-type="received">
<day>04</day>
<month>12</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>05</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Liang, Xue, Zhou, Guo and Wen.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Liang, Xue, Zhou, Guo and Wen</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Correlated phenotypes may have both shared and unique causal factors, and jointly modeling these phenotypes can enhance prediction performance by enabling efficient information transfer.</p>
</sec>
<sec>
<title>Methods</title>
<p>We propose an auto-branch multi-task learning model within a deep learning framework for the simultaneous prediction of multiple correlated phenotypes. This model dynamically branches from a hard parameter sharing structure to prevent negative information transfer, ensuring that parameter sharing among phenotypes is beneficial.</p>
</sec>
<sec>
<title>Results</title>
<p>Through simulation studies and analysis of seven Alzheimer&#x0027;s disease-related phenotypes, our method consistently outperformed Multi-Lasso model, single-task learning approaches, and commonly used hard parameter sharing models with predefine shared layers. These analyses also reveal that while genetic contributions across phenotypes are similar, the relative influence of each genetic factor varies substantially among phenotypes.</p>
</sec>
</abstract>
<kwd-group>
<kwd>alzheimer&#x2019;s disease</kwd>
<kwd>multi-task learning</kwd>
<kwd>phenotype prediction</kwd>
<kwd>deep learning</kwd>
<kwd>autobranch method</kwd>
<kwd>genetic analysis</kwd>
</kwd-group>
<contract-num rid="cn001">82173632</contract-num>
<contract-sponsor id="cn001">National Natural Science Foundation of China<named-content content-type="fundref-id">10.13039/501100001809</named-content>
</contract-sponsor>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Computational Genomics</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Alzheimer&#x2019;s disease (AD) is a progressive neurodegenerative disorder, with its prevalence increasing annually (<xref ref-type="bibr" rid="B1">AuthorAnonymous, 2023</xref>). Approximately 50 million individuals worldwide are affected by dementia with approximately 60%&#x2013;70% being AD cases, and this figure is projected to rise to 152 million by 2050 (<xref ref-type="bibr" rid="B35">Santiago et al., 2019</xref>; <xref ref-type="bibr" rid="B42">Zhang et al., 2021</xref>). AD is a multifactorial condition manifested through various traits, such as cognitive decline and functional changes (<xref ref-type="bibr" rid="B23">L&#xf6;ffler et al., 2014</xref>; <xref ref-type="bibr" rid="B15">Jabir et al., 2021</xref>). Genetic risk prediction models have been developed for various AD-related traits, but these models usually only focus on one trait, ignoring their inter-relationships (<xref ref-type="bibr" rid="B17">Jung et al., 2020</xref>; <xref ref-type="bibr" rid="B47">Zhu et al., 2024</xref>). Although each AD-related trait provides valuable information on the genetic risk of AD, none of them alone can capture the full complexity of the disease and a comprehensive model that can jointly model multiple traits is needed.</p>
<p>Cognitive and functional changes commonly observed in AD patients can be assessed using several tools, including the Mini-Mental State Examination (MMSE), Montreal Cognitive Assessment (MoCA), Clinical Dementia Rating-Sum of Boxes (CDRSB), Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), and the Functional Activities Questionnaire (FAQ). MMSE and MoCA assess general cognitive impairment, with MoCA being more sensitive in detecting early AD (<xref ref-type="bibr" rid="B31">Pinto et al., 2019</xref>; <xref ref-type="bibr" rid="B11">Duc et al., 2020</xref>). ADAS13 and CDRSB are designed for tracking AD progression. However, ADAS13 measures the severity of cognitive symptoms (<xref ref-type="bibr" rid="B5">Bucholc et al., 2019</xref>), whereas CDRSB assesses both cognitive and functional domains, offering a more comprehensive view of how AD affects a patient&#x2019;s daily life (<xref ref-type="bibr" rid="B8">Cullen et al., 2020</xref>). FAQ focuses on assessing functional ability in daily activities (<xref ref-type="bibr" rid="B30">Petersen et al., 2021</xref>). Neuroimaging is also used in AD diagnosis and monitoring (<xref ref-type="bibr" rid="B4">Besson et al., 2015</xref>; <xref ref-type="bibr" rid="B40">Winer et al., 2018</xref>). For example, florbetapir (AV45) detects the amyloid-beta plaque in the brain (<xref ref-type="bibr" rid="B26">Mattson, 2004</xref>; <xref ref-type="bibr" rid="B16">Johnson et al., 2013</xref>). Fluorodeoxyglucose (FDG) measures brain glucose metabolism and identifies regions of hypometabolism (<xref ref-type="bibr" rid="B10">de Paula Faria et al., 2022</xref>). While AD assessment tools provide valuable information, none of them alone can be treated as a gold standard for AD diagnosis, especially for early-stage cases. For example, although amyloid plaque is a hallmark feature of AD, some individuals with such manifestations never develop into AD (<xref ref-type="bibr" rid="B33">Reinitz et al., 2022</xref>).</p>
<p>PET-imaging, cognitive, and functional changes provide confirmatory and complementary information regarding AD risk. Simultaneous modelling of them can leverage information across traits, which facilitates the detection of new biomarkers and improves the overall prediction. However, existing prediction models mainly focus on a single trait. For example, traditional models such as gBLUP build separate prediction models for each trait (<xref ref-type="bibr" rid="B9">de Los Campos et al., 2013</xref>). This trait-specific focus persists even within the deep learning domain. For example, <xref ref-type="bibr" rid="B11">Duc et al. (2020)</xref> employed a single task learning (STL) model to automatically diagnose AD (<xref ref-type="bibr" rid="B11">Duc et al., 2020</xref>). <xref ref-type="bibr" rid="B21">Liu et al. (2022)</xref> introduced an interpretable STL model to assess the risk of AD based on high-dimensional genomic data, where PET imaging outcomes were predicted separately (<xref ref-type="bibr" rid="B21">Liu et al., 2022</xref>). Notably, some studies have explored modeling multiple tasks simultaneously. For example, a classic multi-task model, Multi-Lasso, has shown promising performance when applied on SNPs data (<xref ref-type="bibr" rid="B38">Wang et al., 2012</xref>; <xref ref-type="bibr" rid="B3">Bee et al., 2024</xref>). This method, originally proposed by <xref ref-type="bibr" rid="B29">Obozinski et al. (2006)</xref>, applies joint sparse regularization through <italic>&#x2113;</italic>
<sub>2</sub>,<sub>1</sub>-norm across tasks, enabling feature sharing among related tasks. It has been applied in several studies (<xref ref-type="bibr" rid="B38">Wang et al., 2012</xref>; <xref ref-type="bibr" rid="B3">Bee et al., 2024</xref>), including the recent application on Alzheimer&#x2019;s Disease Neuroimaging Initiative (ADNI) dataset (<xref ref-type="bibr" rid="B7">Cheng et al., 2019</xref>). These studies indicate the potential of multi-task modeling strategies in the context of AD prediction, although their practical application remains limited.</p>
<p>In recent years, Multi-task learning (MTL) has been widely applied in the field of deep learning as an effective strategy to improve model performance. It has been successfully applied to model multiple correlated outcomes, particularly in typical deep learning scenarios such as natural language processing (<xref ref-type="bibr" rid="B43">Zhang et al., 2020</xref>; <xref ref-type="bibr" rid="B19">Li et al., 2022</xref>) and image classification (<xref ref-type="bibr" rid="B20">Liu et al., 2021</xref>). Current deep MTL approaches can be broadly categorized into hard and soft parameter sharing models. Hard parameter sharing models use shared layers across tasks, typically sharing all layers except the last to learn a common representation while capturing task-specific characteristics (<xref ref-type="bibr" rid="B37">Vandenhende et al., 2022</xref>). Recent advances enable these models to automatically determine which layers to share. For example, the Fully Adaptive Feature Sharing method dynamically widens layers based on similarities among tasks (<xref ref-type="bibr" rid="B25">Lu et al., 2017</xref>). The Multilinear Relationship Network discovers inter-task relationships and alleviates the dilemma of negative transfer by jointly training transferable features (<xref ref-type="bibr" rid="B24">Long et al., 2017</xref>). The Task Affinity Grouping optimizes layer sharing by branching the network according to inter-task affinity scores (<xref ref-type="bibr" rid="B12">Fifty et al., 2021</xref>). Unlike hard sharing, soft parameter sharing models allow each task to maintain independent parameters and control the levels of sharedness using additional parameters. They offer greater flexibility, but at the cost of increased computational demands. The cross-stitch network represents a classic example, where additional parameters are introduced in the cross-stitch unit to ascertain the optimal degree of sharedness (<xref ref-type="bibr" rid="B27">Misra et al., 2016</xref>). AD-related traits, such as cognitive scores, functional assessments, and neuroimaging findings are interconnected. MTL approaches, especially those computationally efficient hard parameter sharing models, have great potential to enhance generalization, learning efficiency and overall prediction accuracy for genetic risk predictions. However, existing MTLs have been rarely used in such applications, partially due to the low signal-to-noise ratio and unclear levels of genetic relatedness.</p>
<p>We here developed an auto-branch multi-task learning model for the prediction analyses of multiple correlated traits using genetic data. Our method can distinguish and integrate commonalities and unique characteristics across multiple traits, leading to improved prediction performance, as measured by both Pearson correlation and root mean squared error (RMSE). In the following sections, we first provided the technical details of our method and then conducted extensive simulation studies to evaluate its performance. Finally, we built genetic risk prediction models for multiple AD-related traits, including cognitive and functional assessments and PET imaging outcomes, using data sourced from the ADNI (<xref ref-type="bibr" rid="B28">Mueller et al., 2005</xref>).</p>
</sec>
<sec sec-type="methods" id="s2">
<title>2 Methods</title>
<p>Our method is developed using the idea originally proposed by <xref ref-type="bibr" rid="B12">Fifty et al. (2021)</xref> in the analyses of facial image dataset CelebA (<xref ref-type="bibr" rid="B22">Liu et al., 2015</xref>) and computer vision dataset Taskonomy (<xref ref-type="bibr" rid="B41">Zamir et al., 2018</xref>). In hard parameter sharing models, the gradient update of one task can influence others. If the gradient update of one task reduces the loss of another, a &#x201c;synergistic effect&#x201d; between the two tasks is observed. Conversely, an &#x201c;antagonistic effect&#x201d; occurs when the update negatively impacts the other task. Jointly training can enhance model performance for synergistic tasks by leveraging their positive correlations, but it may reduce performance for antagonistic tasks. We proposed to quantify the &#x201c;synergistic effect&#x201d; among correlated traits and branch the network where traits are considered antagonistic. Specifically, we first constructed a hard parameter sharing model with all layers except the last shared to predict multiple traits using genetic data. We then quantified trait similarities and grouped traits using the inter-trait affinity (<xref ref-type="bibr" rid="B12">Fifty et al., 2021</xref>). Finally, we branched the hard parameter sharing model for traits that are deemed antagonistic. Unlike <xref ref-type="bibr" rid="B12">Fifty et al. (2021)</xref>, who used separate models for traits in each group, we proposed to use a hard-parameter sharing strategy and branch the network when phenotypes are &#x201c;antagonistic.&#x201d; This is mainly because correlated phenotypes are likely to have shared genetic determinants (<xref ref-type="bibr" rid="B2">Badr&#xe9; and Pan, 2023</xref>), and we hypothesized that this can be exploited to improve predictions. The overview of our workflow is in <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>The overview of the study design.</p>
</caption>
<graphic xlink:href="fgene-16-1538544-g001.tif"/>
</fig>
<sec id="s2-1">
<title>2.1 Train all traits together in a hard parameter sharing model</title>
<p>We proposed to use a hard parameter sharing model, where all layers except for the last one are shared, to train prediction models for all traits and evaluate the effect of gradient update from one task on another. We utilized the shared layers to capture the common representations among these traits and used the last trait-specific layer to handle their uniqueness. To consider biological information and improve model interpretation, we added a customized layer right after the input layer (Step 1 in <xref ref-type="fig" rid="F1">Figure 1</xref>), where predictors from the same genes are first grouped together and then fed to the downstream networks. This customized layer acts similarly to those set-based analyses (<xref ref-type="bibr" rid="B39">Wang et al., 2016</xref>; <xref ref-type="bibr" rid="B6">Chen et al., 2019</xref>), where weak signals within a gene are aggregated, enhancing the overall performance of the models. Note that although we aggregated the signals within each gene, a similar layer can be designed based on other biological information (e.g., pathways). As opposed to image classification, the signals in genetic data are weak and can lead to poor prediction models without signal enhancement, which can have a profound impact on gauging the trait similarities.</p>
</sec>
<sec id="s2-2">
<title>2.2 Branch network based on inter-trait affinity</title>
<p>Within the hard parameter-sharing framework, traits transfer information to each other through successive gradient updates to the shared parameters. For traits that are intrinsically similar, the update in one trait&#x2019;s gradient on the shared parameters would lead to a reduction in loss for the others. On the contrary, this update can lead to a negligible or an increase in loss for independent traits. Therefore, we used inter-task affinity scores, calculated based on loss changes during parameter training (<xref ref-type="bibr" rid="B12">Fifty et al., 2021</xref>), to evaluate the pairwise similarity among traits. These scores are further used to determine whether traits should be trained together or branched. Specifically, we gathered gradient information during the training of the model outlined in session 2.1. We calculated the pairwise trait affinity during each parameter update, where the affinity was defined as the extent to which the gradient update of shared parameters by trait <italic>i</italic> impacted the loss of trait <italic>j</italic> (<xref ref-type="bibr" rid="B12">Fifty et al., 2021</xref>). At step <italic>t</italic>, following a gradient update of the shared parameters based on the loss of trait <italic>i</italic>, the affinity of trait <italic>i</italic> to trait <italic>j</italic> is calculated as shown in <xref ref-type="disp-formula" rid="e1">Equation 1</xref>:<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:msubsup>
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf1">
<mml:math id="m2">
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:msubsup>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> respectively denote the loss of trait <italic>j</italic> before and after shared parameter updates based on task <italic>i</italic> at step <italic>t</italic>. If the affinity is greater than 0, it signifies that the gradient update of task <italic>i</italic> positively influences trait <italic>j</italic>. We defined the overall affinity between trait <italic>i</italic> and trait <italic>j</italic> as an average over <italic>T</italic> epochs, as shown in <xref ref-type="disp-formula" rid="e2">Equation 2</xref>:<disp-formula id="e2">
<mml:math id="m4">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:msubsup>
<mml:msubsup>
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mi>t</mml:mi>
</mml:msubsup>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>
</p>
<p>To enable efficient information transfer, traits within the same branch are expected to have high pairwise affinities (<inline-formula id="inf3">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>). Let <inline-formula id="inf4">
<mml:math id="m6">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> be the set of traits that are in the same branch as trait <italic>i</italic> and <inline-formula id="inf5">
<mml:math id="m7">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="|" close="|" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> be the cardinality of the set. The inter-trait affinity for trait <italic>i</italic> can determine the synergistic effects among all traits in the same branch (i.e., set <inline-formula id="inf6">
<mml:math id="m8">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) and is defined in <xref ref-type="disp-formula" rid="e3">Equation 3</xref>:<disp-formula id="e3">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:msub>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mstyle>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mfenced open="|" close="|" separators="|">
<mml:mrow>
<mml:msub>
<mml:mi>S</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
</p>
<p>Obviously, <inline-formula id="inf7">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is large when all traits in the branch are similar to trait <italic>i</italic> (i.e., <inline-formula id="inf8">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is large), and it is small when traits are independent. We define the total inter-trait affinity for <italic>K</italic> traits as shown in <xref ref-type="disp-formula" rid="e4">Equation 4</xref>:<disp-formula id="e4">
<mml:math id="m12">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>K</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mo>&#x2211;</mml:mo>
<mml:mi>i</mml:mi>
</mml:munder>
</mml:mstyle>
<mml:msub>
<mml:mover accent="true">
<mml:mi mathvariant="script">Z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
</p>
<p>Identifying the best parameter sharing (i.e., branching) strategy is equivalent to finding the optimal number of branches and the partition of the traits that maximize the total inter-trait affinity defined in <xref ref-type="disp-formula" rid="e4">Equation 4</xref>. However, this problem is an NP-hard problem (<xref ref-type="bibr" rid="B12">Fifty et al., 2021</xref>). From the practical perspective, we treat the optimal number of branches as a prior and find the best partition of traits that maximize <inline-formula id="inf9">
<mml:math id="m13">
<mml:mrow>
<mml:mi mathvariant="script">Z</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</sec>
<sec id="s2-3">
<title>2.3 Joint prediction of multiple traits</title>
<p>Given a pre-specified number of branches and its corresponding best partition of traits, we constructed a hard parameter sharing model with traits in different partitions branched (Step 3 in <xref ref-type="fig" rid="F1">Figure 1</xref>). As phenotypes are likely to have shared genetic determinants, we set the top few hidden layers as shared among all phenotypes. We then branched the networks for phenotypes that are deemed dissimilar. Our basic rationale for such a design is to use 1) common shared layers to learn the pan-representations across multiple traits, 2) layers shared by branch to capture common characteristics among traits that are similar, and 3) task-specific layers to capture the uniqueness of each trait. This network architecture enables efficient information transfer among similar phenotypes while avoiding negative impacts on dissimilar ones.</p>
</sec>
</sec>
<sec id="s3">
<title>3 Simulation studies</title>
<p>We conducted extensive simulations to assess the performance of our method. As outlined in the method, we pre-set the number of branches to be 2, 3, and 4, and then found the trait partition accordingly. We compared our model with a typical hard parameter sharing model where all parameters except the last are shared (denoted as HPS), a single-task model with each trait modeled independently (denoted as STL), and Multi-lasso that is a classic method for modeling multiple traits (<xref ref-type="bibr" rid="B7">Cheng et al., 2019</xref>). As we aim to improve prediction for multiple AD-related traits (i.e., FDG, AV45, FAQ, CDRSB, ADAS13, MMSE, and MoCA), we make our simulation settings like the ADNI dataset. Specifically, we directly extracted genomic data from ADNI and simulated seven phenotypes based on causal variants, which are harbored on six randomly selected genes and comprised of 10% of the total variants. We randomly split the samples into training, validation, and testing sets with a ratio of 8:1:1. For all models, we set 2 hidden layers with 128 and 32 hidden nodes, respectively (<xref ref-type="sec" rid="s12">Supplementary Figure S1</xref>).</p>
<p>Since all simulated phenotypes are continuous, we used mean squared error (MSE) as the loss function during the model training process for all deep learning models. We evaluated the prediction performance based on the testing set and reported the average prediction Pearson correlations and root of MSE (RMSE) based on 100 Monte Carlo simulations under each setting. Pearson correlation captures the linear consistency between predicted and true values and is widely used in continuous phenotype prediction. RMSE complements this by quantifying the average prediction error in the same unit as the original data, providing a comprehensive assessment of model performance.</p>
<sec id="s3-1">
<title>3.1 Scenario 1: the impact of different numbers of underlying groups among traits</title>
<p>In this scenario, we evaluated the impact of different numbers of underlying groups among the traits. We started with the case where all traits shared the same genetic causes (i.e., the underlying group is 1) and gradually split the traits into different groups until no traits shared any genetic causes (i.e., the underlying number of groups is 7).</p>
<p>Let <inline-formula id="inf10">
<mml:math id="m14">
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denote the set of traits in the <inline-formula id="inf11">
<mml:math id="m15">
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-th group and <inline-formula id="inf12">
<mml:math id="m16">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is a <inline-formula id="inf13">
<mml:math id="m17">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> matrix of the corresponding causal variants for the <inline-formula id="inf14">
<mml:math id="m18">
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-th group, where n is the sample size. Note that in this set of simulations, there is no overlap between causal variants among different groups. When the number of underlying groups among the seven traits is one (i.e., <inline-formula id="inf15">
<mml:math id="m19">
<mml:mrow>
<mml:mfenced open="" close=")" separators="|">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="{" close="}" separators="|">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>4</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>6</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>, all traits share identical causes and are simulated as <inline-formula id="inf16">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:msub>
</mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf17">
<mml:math id="m21">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf18">
<mml:math id="m22">
<mml:mrow>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">G</mml:mi>
</mml:msub>
<mml:mn mathvariant="bold">1</mml:mn>
</mml:msub>
</mml:msub>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf19">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:msub>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. When the number of underlying groups is seven, each trait has its own causes (i.e., <inline-formula id="inf20">
<mml:math id="m24">
<mml:mrow>
<mml:mfenced open="" close=")" separators="|">
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mfenced open="{" close="}" separators="|">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> and is simulated using <inline-formula id="inf21">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf22">
<mml:math id="m26">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> with <inline-formula id="inf23">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">G</mml:mi>
</mml:msub>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:msub>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> When the number of underlying groups is between two and six (i.e., some traits share identical causes), we simulated each trait as <inline-formula id="inf24">
<mml:math id="m28">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> for <inline-formula id="inf25">
<mml:math id="m29">
<mml:mrow>
<mml:mo>&#x2200;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msub>
<mml:mi>G</mml:mi>
<mml:mi>j</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> with <inline-formula id="inf26">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">G</mml:mi>
<mml:mi mathvariant="bold-italic">j</mml:mi>
</mml:msub>
</mml:msub>
</mml:msub>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. We split the seven traits given the specific number of groups and the details of grouping are summarized in <xref ref-type="sec" rid="s12">Supplementary Table S1</xref>.</p>
<p>We varied effect sizes by ranging <inline-formula id="inf27">
<mml:math id="m31">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> from 1:3 to 1:6. <xref ref-type="table" rid="T1">Table 1</xref> presents the average Pearson correlations across the seven simulated traits, while the specific values for each trait are provided in <xref ref-type="sec" rid="s12">Supplementary Tables S2&#x2013;S5</xref>. When all traits have the same underlying causal factors, our model, regardless of the number of pre-specified number of branches, has similar performance to HPS, and it has much better performance than STL and Multi-Lasso model. For example, when <inline-formula id="inf28">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:5, the average Pearson correlations for Multi-Lasso, HPS, and STL are 0.266, 0.356, and 0.259, respectively. For our proposed method, they are 0.366, 0.368, and 0.355 when the respective pre-specified numbers of branches are two, three and four. <xref ref-type="table" rid="T2">Table 2</xref> presents the average RMSE values for the seven simulated traits, with trait-specific results shown in <xref ref-type="sec" rid="s12">Supplementary Tables S6&#x2013;S9</xref>. These results indicate that our model also performs well in terms of RMSE. For example, under the <inline-formula id="inf29">
<mml:math id="m33">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:5 setting, the RMSE of Multi-Lasso reaches 1.192, which is substantially higher than those of other methods; HPS and STL yield RMSEs of 0.918 and 0.976, respectively. In contrast, our model yields lower RMSEs of 0.914, 0.912, and 0.914 for branch numbers of two, three, and four, respectively, demonstrating lower prediction errors and confirming the robustness of our model in controlling estimation error.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Average Pearson correlations for seven traits under different numbers of underlying groups among traits.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Sharing situations</th>
<th align="center">Multi-lasso</th>
<th align="center">HPS<xref ref-type="table-fn" rid="Tfn1">
<sup>a</sup>
</xref>
</th>
<th align="center">2 Groups<xref ref-type="table-fn" rid="Tfn2">
<sup>b</sup>
</xref>
</th>
<th align="center">3 Groups<xref ref-type="table-fn" rid="Tfn3">
<sup>c</sup>
</xref>
</th>
<th align="center">4 Groups<xref ref-type="table-fn" rid="Tfn4">
<sup>d</sup>
</xref>
</th>
<th align="center">STL<xref ref-type="table-fn" rid="Tfn5">
<sup>e</sup>
</xref>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="6" align="center">
<inline-formula id="inf30">
<mml:math id="m34">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:3</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">0.463</td>
<td align="center">0.531</td>
<td align="center">0.543</td>
<td align="center">0.538</td>
<td align="center">0.539</td>
<td align="center">0.433</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">0.425</td>
<td align="center">0.478</td>
<td align="center">0.495</td>
<td align="center">0.491</td>
<td align="center">0.494</td>
<td align="center">0.377</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">0.404</td>
<td align="center">0.420</td>
<td align="center">0.448</td>
<td align="center">0.451</td>
<td align="center">0.447</td>
<td align="center">0.416</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">0.402</td>
<td align="center">0.402</td>
<td align="center">0.417</td>
<td align="center">0.417</td>
<td align="center">0.422</td>
<td align="center">0.398</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">0.393</td>
<td align="center">0.322</td>
<td align="center">0.332</td>
<td align="center">0.331</td>
<td align="center">0.328</td>
<td align="center">0.395</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf31">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:4</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">0.349</td>
<td align="center">0.438</td>
<td align="center">0.445</td>
<td align="center">0.443</td>
<td align="center">0.444</td>
<td align="center">0.343</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">0.313</td>
<td align="center">0.377</td>
<td align="center">0.394</td>
<td align="center">0.394</td>
<td align="center">0.392</td>
<td align="center">0.319</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">0.291</td>
<td align="center">0.320</td>
<td align="center">0.338</td>
<td align="center">0.340</td>
<td align="center">0.337</td>
<td align="center">0.290</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">0.293</td>
<td align="center">0.294</td>
<td align="center">0.310</td>
<td align="center">0.312</td>
<td align="center">0.313</td>
<td align="center">0.300</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">0.289</td>
<td align="center">0.220</td>
<td align="center">0.234</td>
<td align="center">0.235</td>
<td align="center">0.226</td>
<td align="center">0.291</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf32">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:5</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">0.266</td>
<td align="center">0.356</td>
<td align="center">0.366</td>
<td align="center">0.368</td>
<td align="center">0.355</td>
<td align="center">0.259</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">0.239</td>
<td align="center">0.274</td>
<td align="center">0.305</td>
<td align="center">0.301</td>
<td align="center">0.304</td>
<td align="center">0.254</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">0.215</td>
<td align="center">0.218</td>
<td align="center">0.251</td>
<td align="center">0.262</td>
<td align="center">0.253</td>
<td align="center">0.229</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">0.220</td>
<td align="center">0.199</td>
<td align="center">0.228</td>
<td align="center">0.237</td>
<td align="center">0.235</td>
<td align="center">0.230</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">0.218</td>
<td align="center">0.140</td>
<td align="center">0.147</td>
<td align="center">0.138</td>
<td align="center">0.155</td>
<td align="center">0.219</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf33">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:6</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">0.207</td>
<td align="center">0.290</td>
<td align="center">0.295</td>
<td align="center">0.295</td>
<td align="center">0.308</td>
<td align="center">0.210</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">0.183</td>
<td align="center">0.218</td>
<td align="center">0.239</td>
<td align="center">0.242</td>
<td align="center">0.235</td>
<td align="center">0.186</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">0.163</td>
<td align="center">0.153</td>
<td align="center">0.192</td>
<td align="center">0.198</td>
<td align="center">0.190</td>
<td align="center">0.174</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">0.169</td>
<td align="center">0.129</td>
<td align="center">0.159</td>
<td align="center">0.172</td>
<td align="center">0.174</td>
<td align="center">0.173</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">0.171</td>
<td align="center">0.078</td>
<td align="center">0.098</td>
<td align="center">0.088</td>
<td align="center">0.096</td>
<td align="center">0.169</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn1">
<label>
<sup>a</sup>
</label>
<p>Hard parameter sharing across all layers except the last one.</p>
</fn>
<fn id="Tfn2">
<label>
<sup>b</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn3">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn4">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is four.</p>
</fn>
<fn id="Tfn5">
<label>
<sup>e</sup>
</label>
<p>Each trait is modeled independently without accounting for trait correlations.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T2" position="float">
<label>TABLE 2</label>
<caption>
<p>Average RMSEs for seven traits under different numbers of underlying groups among traits.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">Sharing situations</th>
<th align="center">Multi-lasso</th>
<th align="center">HPS<xref ref-type="table-fn" rid="Tfn6">
<sup>a</sup>
</xref>
</th>
<th align="center">2 Groups<xref ref-type="table-fn" rid="Tfn7">
<sup>b</sup>
</xref>
</th>
<th align="center">3 Groups<xref ref-type="table-fn" rid="Tfn8">
<sup>c</sup>
</xref>
</th>
<th align="center">4 Groups<xref ref-type="table-fn" rid="Tfn9">
<sup>d</sup>
</xref>
</th>
<th align="center">STL<xref ref-type="table-fn" rid="Tfn10">
<sup>e</sup>
</xref>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="6" align="center">
<inline-formula id="inf34">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:3</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">1.001</td>
<td align="center">0.832</td>
<td align="center">0.829</td>
<td align="center">0.829</td>
<td align="center">0.831</td>
<td align="center">0.893</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">1.036</td>
<td align="center">0.857</td>
<td align="center">0.851</td>
<td align="center">0.853</td>
<td align="center">0.850</td>
<td align="center">0.910</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">1.054</td>
<td align="center">0.890</td>
<td align="center">0.878</td>
<td align="center">0.880</td>
<td align="center">0.879</td>
<td align="center">0.908</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">1.056</td>
<td align="center">0.892</td>
<td align="center">0.886</td>
<td align="center">0.885</td>
<td align="center">0.887</td>
<td align="center">0.913</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">1.061</td>
<td align="center">0.941</td>
<td align="center">0.932</td>
<td align="center">0.936</td>
<td align="center">0.934</td>
<td align="center">0.917</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf35">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:4</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">1.115</td>
<td align="center">0.882</td>
<td align="center">0.879</td>
<td align="center">0.879</td>
<td align="center">0.876</td>
<td align="center">0.937</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">1.145</td>
<td align="center">0.908</td>
<td align="center">0.900</td>
<td align="center">0.902</td>
<td align="center">0.900</td>
<td align="center">0.947</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">1.162</td>
<td align="center">0.936</td>
<td align="center">0.930</td>
<td align="center">0.927</td>
<td align="center">0.927</td>
<td align="center">0.960</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">1.161</td>
<td align="center">0.935</td>
<td align="center">0.930</td>
<td align="center">0.931</td>
<td align="center">0.929</td>
<td align="center">0.963</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">1.163</td>
<td align="center">0.979</td>
<td align="center">0.973</td>
<td align="center">0.973</td>
<td align="center">0.969</td>
<td align="center">0.958</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf36">
<mml:math id="m40">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:5</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">1.192</td>
<td align="center">0.918</td>
<td align="center">0.914</td>
<td align="center">0.912</td>
<td align="center">0.914</td>
<td align="center">0.976</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">1.215</td>
<td align="center">0.939</td>
<td align="center">0.933</td>
<td align="center">0.935</td>
<td align="center">0.933</td>
<td align="center">0.986</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">1.232</td>
<td align="center">0.967</td>
<td align="center">0.958</td>
<td align="center">0.958</td>
<td align="center">0.961</td>
<td align="center">0.974</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">1.230</td>
<td align="center">0.963</td>
<td align="center">0.957</td>
<td align="center">0.957</td>
<td align="center">0.958</td>
<td align="center">0.990</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">1.229</td>
<td align="center">0.995</td>
<td align="center">0.992</td>
<td align="center">0.995</td>
<td align="center">0.992</td>
<td align="center">0.985</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf37">
<mml:math id="m41">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:6</td>
<td align="left"/>
</tr>
<tr>
<td align="center">Complete Sharing</td>
<td align="center">1.246</td>
<td align="center">0.942</td>
<td align="center">0.936</td>
<td align="center">0.937</td>
<td align="center">0.939</td>
<td align="center">0.993</td>
</tr>
<tr>
<td align="center">Two-group Sharing</td>
<td align="center">1.266</td>
<td align="center">0.961</td>
<td align="center">0.958</td>
<td align="center">0.957</td>
<td align="center">0.957</td>
<td align="center">0.998</td>
</tr>
<tr>
<td align="center">Three-group Sharing</td>
<td align="center">1.278</td>
<td align="center">0.983</td>
<td align="center">0.979</td>
<td align="center">0.976</td>
<td align="center">0.977</td>
<td align="center">1.000</td>
</tr>
<tr>
<td align="center">Four-group Sharing</td>
<td align="center">1.276</td>
<td align="center">0.975</td>
<td align="center">0.973</td>
<td align="center">0.974</td>
<td align="center">0.974</td>
<td align="center">0.997</td>
</tr>
<tr>
<td align="center">No Sharing</td>
<td align="center">1.274</td>
<td align="center">1.001</td>
<td align="center">1.000</td>
<td align="center">1.001</td>
<td align="center">1.001</td>
<td align="center">1.004</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn6">
<label>
<sup>a</sup>
</label>
<p>Hard parameter sharing across all layers except the last one.</p>
</fn>
<fn id="Tfn7">
<label>
<sup>b</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn8">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn9">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is four.</p>
</fn>
<fn id="Tfn10">
<label>
<sup>e</sup>
</label>
<p>Each trait is modeled independently without accounting for trait correlations.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>When neither all traits share identical causes nor are completely independent of one another, our branching method tends to perform better than Multi-Lasso, HPS, and STL models, regardless of the pre-specified number of branches. For instance, when <inline-formula id="inf38">
<mml:math id="m42">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:5 and the number of underlying groups of traits is two, the average Pearson correlations for Multi-Lasso, HPS, and STL models are 0.239, 0.274, and 0.254, respectively. For our auto-branch model, they are 0.305, 0.301 and 0.304 when the pre-specified branches are two, three and four, respectively. In terms of RMSE, our method also demonstrates superior performance. The average RMSEs for our model are 0.933, 0.935, and 0.933 for two, three, and four branches, respectively, all of which are lower than those of Multi-Lasso (1.215), HPS (0.939), and STL (0.986). This demonstrates that moderate grouping can effectively leverage shared information between traits, thereby enhancing prediction performance. It is worth noting that although our proposed method performs the best when the pre-specified number of branches matches the underlying disease model, the difference resulting from this is relatively small, indicating the robustness of our method in most practical applications.</p>
<p>Although it is highly unlikely that disease-relevant correlated traits have no shared genetic causes, we still assessed this situation for completeness. As expected, when each trait has its own causes, jointly training cannot benefit model performance. The single-task STL has the best performance, followed by the Multi-Lasso model. Our method performs similarly to that of HPS, as our model assumes that traits share some common causes and sets the pre-specified number of branches smaller than the true underlying groups. Although we can specify the number of branches to be exactly the same as the number of traits, we considered this unnecessary in practice, as only traits that are expected to share some underlying causes should be dealt with using the MTL approach.</p>
<p>In summary, models assuming either identical causes for all traits or complete independence perform poorly when these assumptions are violated. HPS performs well when all traits share the same causes, but its performance drops sharply as the degrees of sharedness among traits decreases. For STL that assumes traits are independent of each other, it has outperformed the models where grouping is not needed, but it suffers greatly when indeed some phenotypes share identical causes. Multi-Lasso allows for independent feature selection for each trait, which enables competitive performance when traits are largely unrelated. However, in scenarios where traits share underlying causal factors, the model may not fully exploit such shared structures due to the lack of an explicit trait grouping mechanism. For our proposed method, it can have robust and better performance, provided the traits are not completely independent of each other. We consider this property important for the prediction of multiple disease-related traits. For example, within the AD prediction domain, it is highly unlikely that AD-related traits are completely independent and thus STL that fails to account for their relatedness can have sub-optimal performance. Similarly, as the amount of sharedness among these AD-related traits is unknown in advance, HPS is also unlikely to achieve the best performance. A method that can flexibly account for different degrees of sharedness can be of great use for these studies.</p>
</sec>
<sec id="s3-2">
<title>3.2 Scenario 2: the impact of the relative contributions of shared causal variants</title>
<p>In practice, correlated traits can not only have shared causes, but also their own unique risk factors. In this set of simulations, we evaluated the relative contributions of shared causal factors on the model performance. We used <inline-formula id="inf39">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf40">
<mml:math id="m44">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to respectively denote <inline-formula id="inf41">
<mml:math id="m45">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf42">
<mml:math id="m46">
<mml:mrow>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> matrices, representing causal factors specific to trait <italic>i</italic> and shared across traits. Similar to scenario 1, we considered seven traits, and simulated the outcome under the additive model as <inline-formula id="inf43">
<mml:math id="m47">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">X</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf44">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
</mml:msub>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf45">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3b2;</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:msub>
<mml:mi mathvariant="bold-italic">p</mml:mi>
<mml:mi mathvariant="bold-italic">s</mml:mi>
</mml:msub>
</mml:msub>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf46">
<mml:math id="m50">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">&#x3f5;</mml:mi>
<mml:mi mathvariant="bold-italic">i</mml:mi>
</mml:msub>
<mml:mo>&#x223c;</mml:mo>
<mml:mi mathvariant="script">N</mml:mi>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="bold-italic">I</mml:mi>
<mml:mi mathvariant="bold-italic">n</mml:mi>
</mml:msub>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. We gradually varied the relative importance of the unique risk factors to the shared ones and set <inline-formula id="inf47">
<mml:math id="m51">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> at 1:9, 3:7, 5:5, 7:3, and 9:1, indicating a gradual increase in the influence of the shared causal variants. We also varied the effect sizes of all causal variants, where the signal-to-noise ratio (i.e., <inline-formula id="inf48">
<mml:math id="m52">
<mml:mrow>
<mml:mfenced open="" close=")" separators="|">
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> is set at 1:9, 1:16, 1:25, and 1:36.</p>
<p>
<xref ref-type="table" rid="T3">Table 3</xref> presents the average Pearson correlations across the seven simulated traits as the amount of shared causal factors varies and the specific values for each trait are provided in <xref ref-type="sec" rid="s12">Supplementary Tables S10&#x2013;S13</xref>. <xref ref-type="table" rid="T4">Table 4</xref> shows the corresponding average RMSE values, with detailed results listed in <xref ref-type="sec" rid="s12">Supplementary Tables S14&#x2013;S17</xref>. When shared causal variants significantly contribute to the outcomes, our method substantially enhances the prediction. For example, when <inline-formula id="inf49">
<mml:math id="m53">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:9 and signal-to-noise ratio is 1:25, the average Pearson correlation achieved through the auto-branch method are 0.285, 0.293, and 0.286 when the pre-specified number of branches are 2, 3, and 4, respectively. The corresponding RMSEs are all 0.933. In contrast, Multi-Lasso, STL, and HPS obtain lower correlations (0.215, 0.232, and 0.271) and higher RMSEs (1.239, 0.967, and 0.935). When <inline-formula id="inf50">
<mml:math id="m54">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 3:7, a similar trend remains. This indicates that, when shared causal factors explain most of the variabilities in the traits, allowing for grouping to enable information sharing can enhance model performance. HPS generally performs worse than our proposed method, indicating allowing uniqueness for traits when they are not caused by the same factors is important for the improved prediction.</p>
<table-wrap id="T3" position="float">
<label>TABLE 3</label>
<caption>
<p>Average Pearson correlations for seven traits as the relative contributions between unique causal factors and shared causal factors increases.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">
<inline-formula id="inf51">
<mml:math id="m55">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Multi-lasso</th>
<th align="center">HPS<xref ref-type="table-fn" rid="Tfn11">
<sup>a</sup>
</xref>
</th>
<th align="center">2 Groups<xref ref-type="table-fn" rid="Tfn12">
<sup>b</sup>
</xref>
</th>
<th align="center">3 Groups<xref ref-type="table-fn" rid="Tfn13">
<sup>c</sup>
</xref>
</th>
<th align="center">4 Groups<xref ref-type="table-fn" rid="Tfn14">
<sup>d</sup>
</xref>
</th>
<th align="center">STL<xref ref-type="table-fn" rid="Tfn15">
<sup>e</sup>
</xref>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="6" align="center">
<inline-formula id="inf52">
<mml:math id="m56">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:9</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">0.444</td>
<td align="center">0.495</td>
<td align="center">0.501</td>
<td align="center">0.502</td>
<td align="center">0.504</td>
<td align="center">0.416</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">0.440</td>
<td align="center">0.428</td>
<td align="center">0.444</td>
<td align="center">0.450</td>
<td align="center">0.444</td>
<td align="center">0.411</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">0.429</td>
<td align="center">0.404</td>
<td align="center">0.418</td>
<td align="center">0.420</td>
<td align="center">0.414</td>
<td align="center">0.400</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">0.350</td>
<td align="center">0.390</td>
<td align="center">0.402</td>
<td align="center">0.403</td>
<td align="center">0.407</td>
<td align="center">0.395</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">0.357</td>
<td align="center">0.329</td>
<td align="center">0.349</td>
<td align="center">0.353</td>
<td align="center">0.348</td>
<td align="center">0.369</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf53">
<mml:math id="m57">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:16</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">0.324</td>
<td align="center">0.397</td>
<td align="center">0.405</td>
<td align="center">0.418</td>
<td align="center">0.416</td>
<td align="center">0.326</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">0.306</td>
<td align="center">0.335</td>
<td align="center">0.345</td>
<td align="center">0.345</td>
<td align="center">0.343</td>
<td align="center">0.297</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">0.314</td>
<td align="center">0.294</td>
<td align="center">0.319</td>
<td align="center">0.314</td>
<td align="center">0.313</td>
<td align="center">0.295</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">0.299</td>
<td align="center">0.245</td>
<td align="center">0.266</td>
<td align="center">0.264</td>
<td align="center">0.271</td>
<td align="center">0.280</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">0.299</td>
<td align="center">0.234</td>
<td align="center">0.247</td>
<td align="center">0.250</td>
<td align="center">0.245</td>
<td align="center">0.273</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf54">
<mml:math id="m58">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:25</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">0.215</td>
<td align="center">0.271</td>
<td align="center">0.285</td>
<td align="center">0.293</td>
<td align="center">0.286</td>
<td align="center">0.232</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">0.228</td>
<td align="center">0.268</td>
<td align="center">0.284</td>
<td align="center">0.281</td>
<td align="center">0.282</td>
<td align="center">0.222</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">0.217</td>
<td align="center">0.223</td>
<td align="center">0.243</td>
<td align="center">0.238</td>
<td align="center">0.234</td>
<td align="center">0.224</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">0.214</td>
<td align="center">0.151</td>
<td align="center">0.177</td>
<td align="center">0.192</td>
<td align="center">0.188</td>
<td align="center">0.203</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">0.216</td>
<td align="center">0.134</td>
<td align="center">0.155</td>
<td align="center">0.161</td>
<td align="center">0.158</td>
<td align="center">0.215</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf55">
<mml:math id="m59">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:36</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">0.201</td>
<td align="center">0.259</td>
<td align="center">0.269</td>
<td align="center">0.273</td>
<td align="center">0.269</td>
<td align="center">0.200</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">0.187</td>
<td align="center">0.212</td>
<td align="center">0.223</td>
<td align="center">0.227</td>
<td align="center">0.225</td>
<td align="center">0.177</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">0.176</td>
<td align="center">0.152</td>
<td align="center">0.177</td>
<td align="center">0.179</td>
<td align="center">0.179</td>
<td align="center">0.170</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">0.187</td>
<td align="center">0.131</td>
<td align="center">0.152</td>
<td align="center">0.152</td>
<td align="center">0.154</td>
<td align="center">0.177</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">0.168</td>
<td align="center">0.086</td>
<td align="center">0.107</td>
<td align="center">0.110</td>
<td align="center">0.111</td>
<td align="center">0.155</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn11">
<label>
<sup>a</sup>
</label>
<p>Hard parameter sharing across all layers except the last one.</p>
</fn>
<fn id="Tfn12">
<label>
<sup>b</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn13">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn14">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for auto-branch method is four.</p>
</fn>
<fn id="Tfn15">
<label>
<sup>e</sup>
</label>
<p>Each trait is modeled independently without accounting for trait correlations.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T4" position="float">
<label>TABLE 4</label>
<caption>
<p>Average RMSEs for seven traits as the relative contributions between unique causal factors and shared causal factors increases.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="center">
<inline-formula id="inf56">
<mml:math id="m60">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>
</th>
<th align="center">Multi-lasso</th>
<th align="center">HPS<xref ref-type="table-fn" rid="Tfn16">
<sup>a</sup>
</xref>
</th>
<th align="center">2 Groups<xref ref-type="table-fn" rid="Tfn17">
<sup>b</sup>
</xref>
</th>
<th align="center">3 Groups<xref ref-type="table-fn" rid="Tfn18">
<sup>c</sup>
</xref>
</th>
<th align="center">4 Groups<xref ref-type="table-fn" rid="Tfn19">
<sup>d</sup>
</xref>
</th>
<th align="center">STL<xref ref-type="table-fn" rid="Tfn20">
<sup>e</sup>
</xref>
</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="6" align="center">
<inline-formula id="inf57">
<mml:math id="m61">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:9</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">1.019</td>
<td align="center">0.852</td>
<td align="center">0.851</td>
<td align="center">0.851</td>
<td align="center">0.849</td>
<td align="center">0.897</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">1.028</td>
<td align="center">0.886</td>
<td align="center">0.882</td>
<td align="center">0.882</td>
<td align="center">0.882</td>
<td align="center">0.899</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">1.040</td>
<td align="center">0.905</td>
<td align="center">0.900</td>
<td align="center">0.903</td>
<td align="center">0.899</td>
<td align="center">0.906</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">1.038</td>
<td align="center">0.918</td>
<td align="center">0.912</td>
<td align="center">0.912</td>
<td align="center">0.913</td>
<td align="center">0.904</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">1.054</td>
<td align="center">0.933</td>
<td align="center">0.926</td>
<td align="center">0.928</td>
<td align="center">0.926</td>
<td align="center">0.921</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf58">
<mml:math id="m62">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:16</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">1.141</td>
<td align="center">0.899</td>
<td align="center">0.899</td>
<td align="center">0.894</td>
<td align="center">0.898</td>
<td align="center">0.937</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">1.149</td>
<td align="center">0.930</td>
<td align="center">0.925</td>
<td align="center">0.923</td>
<td align="center">0.924</td>
<td align="center">0.944</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">1.146</td>
<td align="center">0.950</td>
<td align="center">0.944</td>
<td align="center">0.944</td>
<td align="center">0.948</td>
<td align="center">0.951</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">1.155</td>
<td align="center">0.965</td>
<td align="center">0.961</td>
<td align="center">0.961</td>
<td align="center">0.960</td>
<td align="center">0.945</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">1.156</td>
<td align="center">0.971</td>
<td align="center">0.966</td>
<td align="center">0.961</td>
<td align="center">0.967</td>
<td align="center">0.960</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf59">
<mml:math id="m63">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:25</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">1.239</td>
<td align="center">0.935</td>
<td align="center">0.933</td>
<td align="center">0.933</td>
<td align="center">0.933</td>
<td align="center">0.967</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">1.222</td>
<td align="center">0.953</td>
<td align="center">0.948</td>
<td align="center">0.948</td>
<td align="center">0.951</td>
<td align="center">0.971</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">1.216</td>
<td align="center">0.961</td>
<td align="center">0.959</td>
<td align="center">0.959</td>
<td align="center">0.959</td>
<td align="center">0.976</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">1.230</td>
<td align="center">0.978</td>
<td align="center">0.975</td>
<td align="center">0.975</td>
<td align="center">0.977</td>
<td align="center">0.978</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">1.231</td>
<td align="center">0.990</td>
<td align="center">0.989</td>
<td align="center">0.988</td>
<td align="center">0.992</td>
<td align="center">0.987</td>
</tr>
<tr>
<td colspan="6" align="center">
<inline-formula id="inf60">
<mml:math id="m64">
<mml:mrow>
<mml:mrow>
<mml:mfenced open="(" close=")" separators="|">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>&#x2b;</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mi>s</mml:mi>
</mml:msub>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>&#x3f5;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1:36</td>
<td align="left"/>
</tr>
<tr>
<td align="center">1:9</td>
<td align="center">1.254</td>
<td align="center">0.952</td>
<td align="center">0.951</td>
<td align="center">0.948</td>
<td align="center">0.947</td>
<td align="center">0.991</td>
</tr>
<tr>
<td align="center">3:7</td>
<td align="center">1.263</td>
<td align="center">0.960</td>
<td align="center">0.957</td>
<td align="center">0.957</td>
<td align="center">0.957</td>
<td align="center">0.995</td>
</tr>
<tr>
<td align="center">5:5</td>
<td align="center">1.271</td>
<td align="center">0.975</td>
<td align="center">0.975</td>
<td align="center">0.973</td>
<td align="center">0.973</td>
<td align="center">0.999</td>
</tr>
<tr>
<td align="center">7:3</td>
<td align="center">1.257</td>
<td align="center">0.982</td>
<td align="center">0.979</td>
<td align="center">0.981</td>
<td align="center">0.978</td>
<td align="center">0.994</td>
</tr>
<tr>
<td align="center">9:1</td>
<td align="center">1.272</td>
<td align="center">0.995</td>
<td align="center">0.994</td>
<td align="center">0.993</td>
<td align="center">0.994</td>
<td align="center">1.009</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn16">
<label>
<sup>a</sup>
</label>
<p>Hard parameter sharing across all layers except the last one.</p>
</fn>
<fn id="Tfn17">
<label>
<sup>b</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn18">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn19">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for auto-branch method is four.</p>
</fn>
<fn id="Tfn20">
<label>
<sup>e</sup>
</label>
<p>Each trait is modeled independently without accounting for trait correlations.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>When shared and non-shared causal factors contribute equally, a well-considered branching strategy can still help the model better capture the distinctions between traits, thereby improving predictions as compared to both STL and HPS. In this scenario, although Multi-Lasso shows comparable performance to our method in terms of Pearson correlation, it exhibits higher RMSE values, indicating its relative limitation in error control. However, when trait-specific factors make major contributions to the outcome variability, while the prediction performance after branching increases as compared to HPS, it does not surpass those in the STL or Multi-Lasso models. This suggests that, in such cases, the independence between traits renders group training less effective and negative information transfer may occur. In contrast, methods like STL or joint models with independent feature selection pathways, such as Multi-Lasso, are more appropriate in scenarios with minimal shared information. Therefore, when traits are expected to be largely independent, STL or similar strategies should be considered as the first choice and multi-task learning is not expected to benefit model performance.</p>
<p>Overall, joint modeling benefits more when the shared factors substantially contributed to correlated traits, whereas separate modeling and Multi-Lasso would be preferred if trait-specific factors explain most of the variability. In practical MTL applications, traits often share a moderate amount of common causes. Under these circumstances, our method has a notable advantage over HPS and STL by effectively utilizing shared information while preserving trait-specific focus, and it also outperforms Multi-Lasso, which fails to fully exploit the shared structure, enhancing its practical utility.</p>
</sec>
</sec>
<sec id="s4">
<title>4 The prediction analyses for multiple AD-related traits</title>
<p>We are interested in predicting multiple AD-related traits, including cognitive scores (i.e., MMSE, MoCA, ADAS13), functional assessments (i.e., FAQ, CDRSB), and neuroimaging findings (i.e., AV45 and FDG), using genetic data obtained from ADNI. ADNI is a comprehensive longitudinal study aimed at identifying biomarkers associated with AD and enhancing its clinical diagnosis and early intervention. As all seven phenotypes are quantitatively measured continuous traits, they were modeled as regression tasks in our study.</p>
<p>Data were downloaded from the ADNI website (<ext-link ext-link-type="uri" xlink:href="https://www.adni.loni.usc.edu/">https://www.adni.loni.usc.edu/</ext-link>). We excluded individuals without genomic data or have missing phenotypes. Only autosome SNPs were considered in our analysis. We adopted a candidate gene approach, where 57 AD susceptibility genes identified based on existing literature were included (<xref ref-type="sec" rid="s12">Supplementary Table S18</xref>). For quality control, we excluded SNPs if they met any of the subsequent criteria: 1) missing rate &#x3e;1%; 2) minor allele frequency (MAF) &#x3c; 5%; 3) Hardy-Weinberg equilibrium test with <inline-formula id="inf61">
<mml:math id="m65">
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-value &#x3c;1 &#xd7; 10<sup>&#x2212;6</sup>; 4) linkage disequilibrium (LD) analysis with an LD threshold of R<sup>2</sup> &#x3e; 0.9. Additionally, individuals with missing rate of SNPs larger than 1% were excluded. All selected SNPs were directly genotyped in the ADNI dataset. For individuals who passed the quality control, we utilized the plink2R package to impute the missing for the included SNPs. A total of 463 participants and 3,797 SNPs harbored on the 57 AD susceptibility genes were included in the final analyses. The distributions of seven phenotypes are shown in <xref ref-type="sec" rid="s12">Supplementary Figure S2</xref>.</p>
<p>We randomly split the data into an 8:1:1 ratio for training, validation and testing to mitigate overfitting, and repeated the random sampling 100 times for robustness. Given that all seven traits are AD-related and some even focus on similar aspect of AD (e.g., cognitive changes), it is unlikely they are fully independent. Likewise, while these traits provide complementary insights into AD, they are unlikely to share identical genetic causes. Therefore, for our method, we set the pre-specified number of branches to be 2, 3, and 4, excluding cases where all traits are independent or share identical causes. For comparison, we included the Multi-Lasso model, the hard sharing model HPS, where all layers are shared except for the last, and the STL model, where each trait is trained separately. We kept the network architecture consistent with that used in the simulations. We reported Pearson correlation and RMSE for all methods and further used Wilcoxon signed-rank test to compare our method to the others.</p>
<p>The prediction performance, measured by Pearson correlation and RMSE, is illustrated in <xref ref-type="fig" rid="F2">Figures 2</xref>, <xref ref-type="fig" rid="F3">3</xref>, respectively. Our auto-branch method performs the best when the pre-specified number of branches is set to three, though the performance differences across various branch numbers are minimal. This method demonstrates superior prediction across multiple traits, not only in terms of higher average Pearson correlations but also by maintaining lower average RMSEs. Among all compared methods, Multi-Lasso demonstrates weaker performance in several key traits (<xref ref-type="table" rid="T5">Table 5</xref>). Specifically, when the number of branches is set to three, the average Pearson correlations for FDG, AV45, FAQ, CDRSB, and ADAS13 show relative improvements of 29.45%, 33.96%, 98.60%, 17.48%, and 7.51% over Multi-Lasso, with corresponding absolute increases of 0.045, 0.059, 0.061, 0.022, and 0.011, respectively. In addition, Multi-Lasso yields consistently higher RMSEs, particularly for FAQ and ADAS13, reflecting its limited capacity to model complex inter-trait relationships. The Wilcoxon signed-rank test indicated that the increase is statistically significant for FDG, AV45, and FAQ (<xref ref-type="sec" rid="s12">Supplementary Table S19</xref>).</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Comparison of prediction performance of multiple AD-related traits using Pearson correlation. The red dots in the figure represent the average Pearson correlations. HPS: Hard parameter sharing across all layers except the last one. 2 Groups: The pre-specified number of branches for auto-branch method is two. 3 Groups: The pre-specified number of branches for auto-branch method is three. 4 Groups: The pre-specified number of branches for auto-branch method is four. STL: Each trait is modeled independently without accounting for trait correlations. Phenotypes include fluorodeoxyglucose (FDG) and florbetapir (AV45) PET imaging, Functional Activities Questionnaire (FAQ), Clinical Dementia Rating-Sum of Boxes (CDRSB) Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), Mini-Mental State Examination (MMSE), and Montreal Cognitive Assessment (MoCA).</p>
</caption>
<graphic xlink:href="fgene-16-1538544-g002.tif"/>
</fig>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Comparison of prediction performance of multiple AD-related traits using RMSE. The red dots in the figure represent the average RMSEs. HPS: Hard parameter sharing across all layers except the last one. 2 Groups: The pre-specified number of branches for auto-branch method is two. 3 Groups: The pre-specified number of branches for auto-branch method is three. 4 Groups: The pre-specified number of branches for auto-branch method is four. STL: Each trait is modeled independently without accounting for trait correlations. Phenotypes include fluorodeoxyglucose (FDG) and florbetapir (AV45) PET imaging, Functional Activities Questionnaire (FAQ), Clinical Dementia Rating-Sum of Boxes (CDRSB) Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), Mini-Mental State Examination (MMSE), and Montreal Cognitive Assessment (MoCA).</p>
</caption>
<graphic xlink:href="fgene-16-1538544-g003.tif"/>
</fig>
<table-wrap id="T5" position="float">
<label>TABLE 5</label>
<caption>
<p>Predictive performance (average Pearson correlations and RMSEs) for seven Alzheimer&#x2019;s-related phenotypes on real-world data.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th rowspan="2" align="center">Different strategies</th>
<th colspan="7" align="center">Seven Alzheimer&#x2019;s related phenotypes<xref ref-type="table-fn" rid="Tfn21">
<sup>a</sup>
</xref>
</th>
</tr>
<tr>
<th align="center">FDG</th>
<th align="center">AV45</th>
<th align="center">FAQ</th>
<th align="center">CDRSB</th>
<th align="center">ADAS13</th>
<th align="center">MMSE</th>
<th align="center">MOCA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="8" align="center">Pearson correlation</td>
</tr>
<tr>
<td align="center">Multi-Lasso</td>
<td align="center">0.153</td>
<td align="center">0.174</td>
<td align="center">0.061</td>
<td align="center">0.125</td>
<td align="center">0.147</td>
<td align="center">0.152</td>
<td align="center">0.118</td>
</tr>
<tr>
<td align="center">HPS<xref ref-type="table-fn" rid="Tfn22">
<sup>b</sup>
</xref>
</td>
<td align="center">0.167</td>
<td align="center">0.177</td>
<td align="center">0.092</td>
<td align="center">0.111</td>
<td align="center">0.116</td>
<td align="center">0.121</td>
<td align="center">0.117</td>
</tr>
<tr>
<td align="center">2 Groups<xref ref-type="table-fn" rid="Tfn23">
<sup>c</sup>
</xref>
</td>
<td align="center">0.181</td>
<td align="center">0.226</td>
<td align="center">0.134</td>
<td align="center">0.152</td>
<td align="center">0.143</td>
<td align="center">0.165</td>
<td align="center">0.113</td>
</tr>
<tr>
<td align="center">3 Groups<xref ref-type="table-fn" rid="Tfn24">
<sup>d</sup>
</xref>
</td>
<td align="center">0.198</td>
<td align="center">0.233</td>
<td align="center">0.122</td>
<td align="center">0.147</td>
<td align="center">0.158</td>
<td align="center">0.151</td>
<td align="center">0.116</td>
</tr>
<tr>
<td align="center">4 Groups<xref ref-type="table-fn" rid="Tfn25">
<sup>e</sup>
</xref>
</td>
<td align="center">0.200</td>
<td align="center">0.226</td>
<td align="center">0.134</td>
<td align="center">0.142</td>
<td align="center">0.156</td>
<td align="center">0.153</td>
<td align="center">0.108</td>
</tr>
<tr>
<td align="center">STL<xref ref-type="table-fn" rid="Tfn26">
<sup>f</sup>
</xref>
</td>
<td align="center">0.112</td>
<td align="center">0.181</td>
<td align="center">0.090</td>
<td align="center">0.122</td>
<td align="center">0.130</td>
<td align="center">0.127</td>
<td align="center">0.112</td>
</tr>
<tr>
<td colspan="8" align="center">RMSE</td>
</tr>
<tr>
<td align="center">Multi-Lasso</td>
<td align="center">0.926</td>
<td align="center">0.285</td>
<td align="center">6.563</td>
<td align="center">1.892</td>
<td align="center">10.563</td>
<td align="center">2.906</td>
<td align="center">4.906</td>
</tr>
<tr>
<td align="center">HPS</td>
<td align="center">0.694</td>
<td align="center">0.225</td>
<td align="center">4.613</td>
<td align="center">1.399</td>
<td align="center">8.169</td>
<td align="center">2.184</td>
<td align="center">3.617</td>
</tr>
<tr>
<td align="center">2 Groups</td>
<td align="center">0.692</td>
<td align="center">0.222</td>
<td align="center">4.575</td>
<td align="center">1.381</td>
<td align="center">8.084</td>
<td align="center">2.170</td>
<td align="center">3.573</td>
</tr>
<tr>
<td align="center">3 Groups</td>
<td align="center">0.692</td>
<td align="center">0.221</td>
<td align="center">4.581</td>
<td align="center">1.374</td>
<td align="center">8.109</td>
<td align="center">2.179</td>
<td align="center">3.558</td>
</tr>
<tr>
<td align="center">4 Groups</td>
<td align="center">0.694</td>
<td align="center">0.222</td>
<td align="center">4.566</td>
<td align="center">1.378</td>
<td align="center">8.140</td>
<td align="center">2.181</td>
<td align="center">3.589</td>
</tr>
<tr>
<td align="center">STL</td>
<td align="center">0.710</td>
<td align="center">0.231</td>
<td align="center">4.693</td>
<td align="center">1.414</td>
<td align="center">8.262</td>
<td align="center">2.263</td>
<td align="center">3.695</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn21">
<label>
<sup>a</sup>
</label>
<p>Phenotypes include fluorodeoxyglucose (FDG) and florbetapir (AV45) PET, imaging, Functional Activities Questionnaire (FAQ), Clinical Dementia Rating-Sum of Boxes (CDRSB) Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), Mini-Mental State Examination (MMSE), and Montreal Cognitive Assessment (MoCA).</p>
</fn>
<fn id="Tfn22">
<label>
<sup>b</sup>
</label>
<p>Hard parameter sharing across all layers except the last one.</p>
</fn>
<fn id="Tfn23">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn24">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn25">
<label>
<sup>e</sup>
</label>
<p>The pre-specified number of branches for auto-branch method is four.</p>
</fn>
<fn id="Tfn26">
<label>
<sup>f</sup>
</label>
<p>Each trait is modeled independently without accounting for trait correlations.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Regardless of branch number, our method consistently performs similarly to or better than HPS (i.e., similar or higher Pearson correlations, and similar or lower RMSE). Specifically, the improvements in average Pearson correlations for FDG, AV45, FAQ, CDRSB, ADAS13, and MMSE with a three-branch setup are 18.56%, 31.64%, 32.61%, 32.43%, 36.21%, and 24.79%, with corresponding absolute increases of 0.031, 0.056, 0.030, 0.036, 0.042, and 0.030, respectively. The Wilcoxon signed-rank test indicated that the increase is statistically significant for AV45, CDRSB, and ADAS13 (<xref ref-type="sec" rid="s12">Supplementary Table S19</xref>).</p>
<p>Compared to the STL models, our method consistently performs similarly to or better than STL (i.e., similar or higher Pearson correlations and consistently lower RMSE). The improvements in average Pearson correlations for FDG, AV45, FAQ, CDRSB, ADAS13, MMSE, and MoCA are 76.18%, 29.07%, 35.78%, 20.36%, 21.62%, 19.03%, and 3.22%, respectively, with corresponding absolute increases of 0.086, 0.052, 0.032, 0.025, 0.028, 0.024, and 0.004. The Wilcoxon signed-rank test indicated a statistically significant increase for FDG, AV45, FAQ, and MMSE (<xref ref-type="sec" rid="s12">Supplementary Table S19</xref>).</p>
<p>Our analyses suggest that the seven AD-related traits neither share identical genetic causes nor are completely independent. PET-imaging traits FDG and AV45 benefit substantially from joint modeling, indicating cognitive and function tests provide auxiliary information. Information transfer between PET-imaging outcomes and cognitive as well as function tests help learn a better representation, leading to the improvement of prediction. In summary, our method improves prediction performance for most of the traits regardless of the pre-specified number of branches, highlighting its robustness and potential for broad practical applications, especially for phenotypes where identifying latent patterns is essential.</p>
<p>To further investigate our model, we calculated the predictive feature importance score for each gene using a permutation-based approach proposed by <xref ref-type="bibr" rid="B21">Liu et al. (2022)</xref>. The basic rationale is that if a gene is predictive, then the model accuracies with and without it would differ significantly. Following the procedure proposed by <xref ref-type="bibr" rid="B21">Liu et al. (2022)</xref>, we assessed the importance of each gene by quantifying the difference in accuracies while accounting for variability. Specifically, we first calculated the Pearson correlation between predicted and observed outcomes using the original data with the already trained auto-branch model. We then recalculated the Pearson correlations after randomly shuffling the SNPs located within the gene of interest, while preserving the genetic structure (e.g., linkage disequilibrium), to generate a null distribution. To empirically estimate the variance, we repeated the permutation process 100 times. The standardized difference in Pearson correlations between the original and shuffled data (also called the predictive feature importance score by Liu et al.) reflects the gene&#x2019;s contribution to predictive performance, with a larger difference indicating greater predictive importance. We then calculated the predictive feature importance score, which follows an asymptotic normal distribution under the null as shown by <xref ref-type="bibr" rid="B21">Liu et al. (2022)</xref>, to evaluate the predictive importance of each gene. Genes were ranked based on their predictive importance scores, and we focused on those with scores greater than 1.645, corresponding to a 5% significance threshold under the asymptotic normal distribution. It is important to note that we did not intend to perform hypothesis testing, rather we chose a cut-off value (i.e., 1.645) and focused on genes with predictive feature importance score larger than this cut-off. Therefore, no multiple testing correction was applied. <xref ref-type="table" rid="T6">Table 6</xref> presents the genes identified as having a probability greater than 75% of being significantly predictive at the 5% level for at least one trait, and the details for each gene are provided in <xref ref-type="sec" rid="s12">Supplementary Tables S20&#x2013;S22</xref>, which correspond to models with 2, 3, and 4 pre-specified numbers of branches, respectively. <italic>APOC1</italic>, <italic>APOE</italic>, and <italic>TOMM40</italic> demonstrate stable and significant predictive power. Even though the most significant genes are similar among phenotypes, their predictive power varies considerably (<xref ref-type="fig" rid="F4">Figure 4</xref>), which explains the superior performance of our branching method over HPS. In our model with three branches, <italic>APOE</italic> plays a crucial role in predicting FDG and AV45, with its removal leading to average Pearson correlations decreases of 0.10 and 0.14, respectively. For FAQ, CDRSB, ADAS13, MMSE and MoCA, the impact is small-to-moderate, with Pearson correlations reduced by 0.06, 0.08, 0.08, 0.08, and 0.04 respectively.</p>
<table-wrap id="T6" position="float">
<label>TABLE 6</label>
<caption>
<p>Probability of the gene being significantly predictive at 5% level with probably for at least one trait greater than 75%.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th rowspan="2" align="center">Different strategies</th>
<th rowspan="2" align="center">Gene</th>
<th colspan="7" align="center">Seven Alzheimer&#x2019;s related phenotypes<xref ref-type="table-fn" rid="Tfn27">
<sup>a</sup>
</xref>
</th>
</tr>
<tr>
<th align="center">FDG</th>
<th align="center">AV45</th>
<th align="center">FAQ</th>
<th align="center">CDRSB</th>
<th align="center">ADAS13</th>
<th align="center">MMSE</th>
<th align="center">MoCA</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td rowspan="3" align="center">2 Groups<xref ref-type="table-fn" rid="Tfn28">
<sup>b</sup>
</xref>
</td>
<td align="center">
<italic>APOC1</italic>
</td>
<td align="center">0.92</td>
<td align="center">0.98</td>
<td align="center">0.87</td>
<td align="center">0.87</td>
<td align="center">0.85</td>
<td align="center">0.85</td>
<td align="center">0.74</td>
</tr>
<tr>
<td align="center">
<italic>APOE</italic>
</td>
<td align="center">0.94</td>
<td align="center">0.97</td>
<td align="center">0.86</td>
<td align="center">0.87</td>
<td align="center">0.90</td>
<td align="center">0.85</td>
<td align="center">0.77</td>
</tr>
<tr>
<td align="center">
<italic>TOMM40</italic>
</td>
<td align="center">0.85</td>
<td align="center">0.93</td>
<td align="center">0.79</td>
<td align="center">0.82</td>
<td align="center">0.77</td>
<td align="center">0.83</td>
<td align="center">0.73</td>
</tr>
<tr>
<td rowspan="3" align="center">3 Groups<xref ref-type="table-fn" rid="Tfn29">
<sup>c</sup>
</xref>
</td>
<td align="center">
<italic>APOC1</italic>
</td>
<td align="center">0.93</td>
<td align="center">0.98</td>
<td align="center">0.77</td>
<td align="center">0.86</td>
<td align="center">0.88</td>
<td align="center">0.84</td>
<td align="center">0.68</td>
</tr>
<tr>
<td align="center">
<italic>APOE</italic>
</td>
<td align="center">0.95</td>
<td align="center">0.96</td>
<td align="center">0.81</td>
<td align="center">0.89</td>
<td align="center">0.90</td>
<td align="center">0.87</td>
<td align="center">0.72</td>
</tr>
<tr>
<td align="center">
<italic>TOMM40</italic>
</td>
<td align="center">0.91</td>
<td align="center">0.93</td>
<td align="center">0.77</td>
<td align="center">0.78</td>
<td align="center">0.76</td>
<td align="center">0.80</td>
<td align="center">0.71</td>
</tr>
<tr>
<td rowspan="3" align="center">4 Groups<xref ref-type="table-fn" rid="Tfn30">
<sup>d</sup>
</xref>
</td>
<td align="center">
<italic>APOC1</italic>
</td>
<td align="center">0.89</td>
<td align="center">0.96</td>
<td align="center">0.75</td>
<td align="center">0.87</td>
<td align="center">0.87</td>
<td align="center">0.84</td>
<td align="center">0.70</td>
</tr>
<tr>
<td align="center">
<italic>APOE</italic>
</td>
<td align="center">0.90</td>
<td align="center">0.95</td>
<td align="center">0.83</td>
<td align="center">0.84</td>
<td align="center">0.89</td>
<td align="center">0.90</td>
<td align="center">0.77</td>
</tr>
<tr>
<td align="center">
<italic>TOMM40</italic>
</td>
<td align="center">0.86</td>
<td align="center">0.93</td>
<td align="center">0.73</td>
<td align="center">0.78</td>
<td align="center">0.78</td>
<td align="center">0.78</td>
<td align="center">0.70</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn27">
<label>
<sup>a</sup>
</label>
<p>Phenotypes include fluorodeoxyglucose (FDG) and florbetapir (AV45) PET, imaging, Functional Activities Questionnaire (FAQ), Clinical Dementia Rating-Sum of Boxes (CDRSB) Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), Mini-Mental State Examination (MMSE), and Montreal Cognitive Assessment (MoCA).</p>
</fn>
<fn id="Tfn28">
<label>
<sup>b</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is two.</p>
</fn>
<fn id="Tfn29">
<label>
<sup>c</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is three.</p>
</fn>
<fn id="Tfn30">
<label>
<sup>d</sup>
</label>
<p>The pre-specified number of branches for the auto-branch method is four.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Distribution of predictive feature importance scores for the top significant genes. The red dots in the figure represent the average Pearson correlations. <bold>(A)</bold>: The pre-specified number of branches for the auto-branch method is two. <bold>(B)</bold>: The pre-specified number of branches for the auto-branch method is three. <bold>(C)</bold>: The pre-specified number of branches for the auto-branch method is four. Phenotypes include fluorodeoxyglucose (FDG) and florbetapir (AV45) PET imaging, Functional Activities Questionnaire (FAQ), Clinical Dementia Rating-Sum of Boxes (CDRSB) Alzheimer&#x2019;s Disease Assessment Scale-Cognitive Subscale 13 (ADAS13), Mini-Mental State Examination (MMSE), and Montreal Cognitive Assessment (MoCA).</p>
</caption>
<graphic xlink:href="fgene-16-1538544-g004.tif"/>
</fig>
</sec>
<sec sec-type="discussion" id="s5">
<title>5 Discussion</title>
<p>In this study, we proposed an efficient and robust auto-branch multi-task learning method for simultaneously predicting multiple correlated traits. Using total inter-task affinity, which quantifies the impact of gradient updates from one trait on the others, our method automatically determines the best partition of traits to enable efficient information transfer among similar traits, thereby enhancing prediction performance.</p>
<p>Through simulations, we found that our method has similar or better performance than that of the hard parameter sharing model where shared layers are pre-specified (<xref ref-type="bibr" rid="B44">Zhao et al., 2020</xref>). Our auto-branch model identifies the optimal phenotype partitioning that maximizes overall inter-trait affinity. Phenotypes grouped together share layers, while those in separate groups are assigned distinct branches. This data-driven approach to sharing layers enables efficient information transfer among inherently similar phenotypes and significantly reduces the risk of negative transfer. Therefore, our method can facilitate the capture of complex patterns in the data, which includes not only the shared representations but also the uniqueness of each phenotype. Our auto-branch method can also outperform the single-task models, when phenotypes shared moderate levels of common causes. This is mainly because single-task models fail to utilize information from auxiliary tasks and the efficient sample sizes are much less than those in the multi-task settings. Note that we designed our prediction model using a branch network architecture as opposed to training separate HPSs for each group of phenotypes. The rationale for such a design lies in the fact that most disease-related traits have shared genetic architecture (<xref ref-type="bibr" rid="B2">Badr&#xe9; and Pan, 2023</xref>), and by allowing all disease-related phenotypes to have some shared layers to enable the efficient modeling of these common patterns. In the unlikely event that the underlying genetic architectures differ significantly among traits in different groups, our proposed inter-trait affinity measure can still guide trait groupings. Separate HPS models can then be applied to each trait group for predictions (<xref ref-type="sec" rid="s12">Supplementary Figure S3</xref>). This network structure facilitates capturing the unique characteristics of each group, making it particularly powerful for analyzing phenotypes that are largely distinct (<xref ref-type="sec" rid="s12">Supplementary Tables S23, S24</xref>). Nevertheless, we recommend using branch network structure as illustrated in <xref ref-type="fig" rid="F1">Figure 1</xref> for genetic risk prediction of multiple correlated traits in most practical applications.</p>
<p>Our proposed method offers significant advantages in predicting seven AD-related phenotypes, including neuroimaging findings, cognitive scores, and functional assessments. For example, in the prediction of FDG, our method with 3 pre-determined branches increases the average Pearson correlations by 29.45%, 18.56% and 76.18% for Multi-Lasso, HPS, and STL, with corresponding absolute increase of 0.059, 0.031 and 0.086, respectively. Similarly, in the prediction of AV45, we have observed an increase of 33.96%, 31.64% and 29.07% for Multi-Lasso, HPS, and STL, with corresponding absolute increase of 0.061, 0.056 and 0.052, respectively. Notably, MoCA did not show significant improvement across the models, which may suggest weaker correlations with other traits or higher noise in its measurements that impacted model performance.</p>
<p>To verify the reliability and effectiveness of our proposed model, we conducted a comparative analysis with several related studies. For instance, <xref ref-type="bibr" rid="B46">Zhu et al. (2016)</xref> incorporated SNPs from the top 10 <italic>APOE</italic>-related genes into their MMSE prediction model and achieved a Pearson correlation of 0.150. In contrast, our method attained a higher correlation of 0.165 on the same task, demonstrating improved predictive performance. <xref ref-type="bibr" rid="B14">Hongmyeong-eup (2015)</xref> selected 39 SNPs, identified from approximately 1.5 million candidates as being closely associated with Alzheimer&#x2019;s disease (AD) progression, and achieved a Pearson correlation of 0.400 in MMSE prediction. To ensure a fair comparison, we also constructed a model using SNPs from three well-established AD-associated genes&#x2014;<italic>APOC1</italic>, <italic>APOE</italic>, and <italic>TOMM40</italic>&#x2014;and obtained a correlation of 0.390, indicating that our approach performs comparably when using similarly strong genetic signals. In addition, <xref ref-type="bibr" rid="B13">Hao et al. (2016)</xref> reported Pearson correlation values ranging from 0.02 to 0.25 and 0.03&#x2013;0.23 for traditional and improved methods on AV45 and FDG phenotypes, respectively. Our model also falls within these performance ranges, suggesting comparable accuracy in these tasks. In summary, our proposed method exhibited comparable predictive capabilities to existing state-of-the-art methods in multiple AD-related phenotypes, further demonstrating its potential in modeling Alzheimer&#x2019;s disease progression.</p>
<p>These performance gains can be largely attributed to the underlying design of our method. Our method is constructed within a deep learning framework, effectively parsing complex trait relationships, particularly in the context of multi-gene co-regulation. Additionally, it can dynamically determine shared layers and thus is more powerful in managing intricate relationships among traits. Evidence suggests that these AD-related traits reflect different aspects of AD, and they are neither entirely correlated nor completely independent. For example, cognitive decline and neuroimaging abnormalities often exhibit strong genetic correlations (<xref ref-type="bibr" rid="B42">Zhang et al., 2021</xref>), but they do not share identical genetic causes. Therefore, methods like Multi-Lasso that rely on fixed feature selection across all tasks are unable to capture these nuanced relationships, as they do not allow for the dynamic identification of shared and unique factors between traits. Similarly, both STL and HPS models are unlikely to achieve the optimal performance, as STL fails to exploit inter-task correlations and HPS pre-specifies shared layers that lead to limited flexibility. Therefore, for the prediction of AD-related traits, our auto-branch multi-task learning method can leverage shared signals among genes or biological pathways to gain a more comprehensive understanding of these phenotypes while allowing each phenotype to have their unique characteristics, and thus offers greater flexibility as compared to Multi-Lasso, STL, and HPS.</p>
<p>We found that the <italic>APOC1</italic>, <italic>APOE</italic>, and <italic>TOMM40</italic> exhibit stable and significant predictive abilities for all seven AD-related phenotypes, but their predictive capabilities vary substantially across traits, indicating HPS that forces most of the model parameter the same is unlikely to work well. All highly predictive genes are significantly associated with AD. For example, <italic>APOE4&#x2a;&#x3b5;4</italic> regulates neuronal metabolism and epigenetics, and is involved in the pathological processes of AD (<xref ref-type="bibr" rid="B32">Prasad and Rao, 2018</xref>). <italic>APOC1</italic> influences AD development through its role in cholesterol metabolism (<xref ref-type="bibr" rid="B18">Leduc et al., 2010</xref>), with the <italic>APOC1 H2</italic> allele potentially acting synergistically with <italic>APOE</italic> to increase the risk of cognitive decline (<xref ref-type="bibr" rid="B45">Zhou et al., 2014</xref>). <italic>TOMM40</italic> is strongly linked to <italic>APOE</italic> and contributes to the pathological changes in AD (<xref ref-type="bibr" rid="B36">Tasaki et al., 2019</xref>), including the formation of neurofibrillary tangles and neuritic plaques. <italic>TOMM40</italic> regulates oxidative stress and mitochondrial function, and is associated with late-onset AD (<xref ref-type="bibr" rid="B34">Roses, 2010</xref>). We noticed that the predictive ability of <italic>APOE</italic> for FDG and AV45 is especially significant, aligning with findings from previous research (<xref ref-type="bibr" rid="B32">Prasad and Rao, 2018</xref>). Future studies are needed to further decipher additional factors that contribute to the variability of each phenotype.</p>
<p>While this study offers valuable insights into the effectiveness of multi-task learning, several limitations remain. Due to the NP-hard nature of the problem, we pre-specified the optimal number of branches and then found the corresponding partitions. Future research can improve this by determining the optimal number of branches using a data-driven manner (e.g., cross-validation). Although our method outperforms baseline approaches, factors such as limited sample size, low-to-median heritability, and trait heterogeneity may still contribute to variability and should be further explored in future work. Additionally, the findings presented in this study are based on simulated data and ADNI dataset. Future work should validate these results across diverse datasets and applications.</p>
<p>In summary, we developed an efficient auto-branch multi-task learning framework for the prediction analyses of multiple correlated phenotypes. It can dynamically branch the network to allow for efficient information transfer and improve the overall prediction. Our method is available at <ext-link ext-link-type="uri" xlink:href="https://github.com/jiaqi69/TAB">https://github.com/jiaqi69/TAB</ext-link>.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="sec" rid="s12">Supplementary Material</xref>, further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>JL: Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Software, Visualization, Writing &#x2013; original draft, Writing &#x2013; review and editing, Validation. ZX: Writing &#x2013; original draft, Writing &#x2013; review and editing, Conceptualization, Data curation, Formal Analysis, Investigation, Methodology, Software, Validation, Visualization. WZ: Writing &#x2013; review and editing, Data curation, Funding acquisition, Resources, Validation, Visualization. XG: Writing &#x2013; review and editing, Conceptualization, Funding acquisition, Project administration, Resources, Supervision. YW: Writing &#x2013; original draft, Writing &#x2013; review and editing, Conceptualization, Funding acquisition, Methodology, Project administration, Resources, Supervision.</p>
</sec>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. This research was funded by the National Natural Science Foundation of China (Award Nos 82173632 and 82271925), the Marsden Fund from the Royal Society of New Zealand (Project No. 19-UOA-209), Basic Research Program of Shanxi (Project No. 20210302123314), and the Graduate Innovation Program of Shanxi Provincial Department of Education (Project No. 2024SJ219).</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The authors declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s12">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fgene.2025.1538544/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fgene.2025.1538544/full&#x23;supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<collab>AuthorAnonymous</collab> (<year>2023</year>). <article-title>2023 Alzheimer&#x27;s disease facts and figures</article-title>. <source>Alzheimers Dement.</source> <volume>19</volume> (<issue>4</issue>), <fpage>1598</fpage>&#x2013;<lpage>1695</lpage>. <pub-id pub-id-type="doi">10.1002/alz.13016</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Badr&#xe9;</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Explainable multi-task learning improves the parallel estimation of polygenic risk scores for many diseases through shared genetic basis</article-title>. <source>PLoS Comput. Biol.</source> <volume>19</volume> (<issue>7</issue>), <fpage>e1011211</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1011211</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bee</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Poole</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Worden</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Dervilis</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Bull</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Multitask feature selection within structural datasets</article-title>. <source>Data-Centric Eng.</source> <volume>5</volume>, <fpage>e4</fpage>. <pub-id pub-id-type="doi">10.1017/dce.2024.1</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Besson</surname>
<given-names>F. L.</given-names>
</name>
<name>
<surname>La Joie</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Doeuvre</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Gaubert</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>M&#xe9;zenge</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Egret</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2015</year>). <article-title>Cognitive and brain profiles associated with current neuroimaging biomarkers of preclinical Alzheimer&#x27;s disease</article-title>. <source>J. Neurosci.</source> <volume>35</volume> (<issue>29</issue>), <fpage>10402</fpage>&#x2013;<lpage>10411</lpage>. <pub-id pub-id-type="doi">10.1523/jneurosci.0150-15.2015</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bucholc</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ding</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Glass</surname>
<given-names>D. H.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Prasad</surname>
<given-names>G.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>A practical computerized decision support system for predicting the severity of Alzheimer&#x27;s disease of an individual</article-title>. <source>Expert Syst. Appl.</source> <volume>130</volume>, <fpage>157</fpage>&#x2013;<lpage>171</lpage>. <pub-id pub-id-type="doi">10.1016/j.eswa.2019.04.022</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Huffman</surname>
<given-names>J. E.</given-names>
</name>
<name>
<surname>Brody</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Efficient variant set mixed model association tests for continuous and binary traits in large-scale whole-genome sequencing studies</article-title>. <source>Am. J. Hum. Genet.</source> <volume>104</volume> (<issue>2</issue>), <fpage>260</fpage>&#x2013;<lpage>274</lpage>. <pub-id pub-id-type="doi">10.1016/j.ajhg.2018.12.012</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheng</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>D.</given-names>
</name>
</person-group>
<collab>Alzheimer&#x2019;s Disease Neuroimaging Initiative</collab> (<year>2019</year>). <article-title>Robust multi-label transfer feature learning for early diagnosis of Alzheimer&#x27;s disease</article-title>. <source>Brain Imaging Behav.</source> <volume>13</volume> (<issue>1</issue>), <fpage>138</fpage>&#x2013;<lpage>153</lpage>. <pub-id pub-id-type="doi">10.1007/s11682-018-9846-8</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cullen</surname>
<given-names>N. C.</given-names>
</name>
<name>
<surname>Zetterberg</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Insel</surname>
<given-names>P. S.</given-names>
</name>
<name>
<surname>Olsson</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Andreasson</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Blennow</surname>
<given-names>K.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Comparing progression biomarkers in clinical trials of early Alzheimer&#x27;s disease</article-title>. <source>Ann. Clin. Transl. Neurol.</source> <volume>7</volume> (<issue>9</issue>), <fpage>1661</fpage>&#x2013;<lpage>1673</lpage>. <pub-id pub-id-type="doi">10.1002/acn3.51158</pub-id>
</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>de Los Campos</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Vazquez</surname>
<given-names>A. I.</given-names>
</name>
<name>
<surname>Fernando</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Klimentidis</surname>
<given-names>Y. C.</given-names>
</name>
<name>
<surname>Sorensen</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Prediction of complex human traits using the genomic best linear unbiased predictor</article-title>. <source>PLoS Genet.</source> <volume>9</volume> (<issue>7</issue>), <fpage>e1003608</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pgen.1003608</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>de Paula Faria</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Estessi de Souza</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Duran</surname>
<given-names>F. L. S.</given-names>
</name>
<name>
<surname>Buchpiguel</surname>
<given-names>C. A.</given-names>
</name>
<name>
<surname>Britto</surname>
<given-names>L. R.</given-names>
</name>
<name>
<surname>Crippa</surname>
<given-names>J. A. S.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Cannabidiol treatment improves glucose metabolism and memory in streptozotocin-induced Alzheimer&#x27;s disease rat model: a proof-of-concept study</article-title>. <source>Int. J. Mol. Sci.</source> <volume>23</volume> (<issue>3</issue>), <fpage>1076</fpage>. <pub-id pub-id-type="doi">10.3390/ijms23031076</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Duc</surname>
<given-names>N. T.</given-names>
</name>
<name>
<surname>Ryu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Qureshi</surname>
<given-names>M. N. I.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>K. H.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>3D-deep learning based automatic diagnosis of Alzheimer&#x2019;s disease with joint MMSE prediction using resting-state fMRI</article-title>. <source>Neuroinformatics</source> <volume>18</volume>, <fpage>71</fpage>&#x2013;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1007/s12021-019-09419-w</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fifty</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Amid</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Anil</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Finn</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Efficiently identifying task groupings for multi-task learning</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>34</volume>, <fpage>27503</fpage>&#x2013;<lpage>27516</lpage>. <pub-id pub-id-type="doi">10.5555/3540261.3542367</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yao</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Risacher</surname>
<given-names>S. L.</given-names>
</name>
<name>
<surname>Saykin</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Identifying multimodal intermediate phenotypes between genetic risk factors and disease status in Alzheimer&#x2019;s disease</article-title>. <source>Neuroinformatics</source> <volume>14</volume>, <fpage>439</fpage>&#x2013;<lpage>452</lpage>. <pub-id pub-id-type="doi">10.1007/s12021-016-9307-8</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<collab>Hongmyeong-eup</collab> (<year>2015</year>). <article-title>Combining clinical and personal genetic characteristics to predict changes in the Mini-Mental State Examination (MMSE) after 24 months</article-title>. <source>Diss. Graduate Inst. Bio-Industry Mech. Eng. Natl. Taiwan Univ.</source>, <fpage>1</fpage>&#x2013;<lpage>84</lpage>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jabir</surname>
<given-names>N. R.</given-names>
</name>
<name>
<surname>Rehman</surname>
<given-names>M. T.</given-names>
</name>
<name>
<surname>Alsolami</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Shakil</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Zughaibi</surname>
<given-names>T. A.</given-names>
</name>
<name>
<surname>Alserihi</surname>
<given-names>R. F.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Concatenation of molecular docking and molecular simulation of BACE-1, &#x3b3;-secretase targeted ligands: in pursuit of Alzheimer&#x27;s treatment</article-title>. <source>Ann. Med.</source> <volume>53</volume> (<issue>1</issue>), <fpage>2332</fpage>&#x2013;<lpage>2344</lpage>. <pub-id pub-id-type="doi">10.1080/07853890.2021.2009124</pub-id>
</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Johnson</surname>
<given-names>K. A.</given-names>
</name>
<name>
<surname>Sperling</surname>
<given-names>R. A.</given-names>
</name>
<name>
<surname>Gidicsin</surname>
<given-names>C. M.</given-names>
</name>
<name>
<surname>Carmasin</surname>
<given-names>J. S.</given-names>
</name>
<name>
<surname>Maye</surname>
<given-names>J. E.</given-names>
</name>
<name>
<surname>Coleman</surname>
<given-names>R. E.</given-names>
</name>
<etal/>
</person-group> (<year>2013</year>). <article-title>Florbetapir (F18-AV-45) PET to assess amyloid burden in Alzheimer&#x27;s disease dementia, mild cognitive impairment, and normal aging</article-title>. <source>Alzheimer&#x27;s and Dementia</source> <volume>9</volume> (<issue>5</issue>), <fpage>S72</fpage>&#x2013;<lpage>S83</lpage>. <pub-id pub-id-type="doi">10.1016/j.jalz.2012.10.007</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jung</surname>
<given-names>S. H.</given-names>
</name>
<name>
<surname>Nho</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Won</surname>
<given-names>H. H.</given-names>
</name>
<name>
<surname>Initiative</surname>
<given-names>A. s.D. N.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Genetic risk prediction of late&#x2010;onset Alzheimer&#x2019;s disease based on tissue&#x2010;specific transcriptomic analysis and polygenic risk scores: genetics/genetic factors of Alzheimer&#x27;s disease</article-title>. <source>Alzheimer&#x27;s and Dementia</source> <volume>16</volume>, <fpage>e045184</fpage>. <pub-id pub-id-type="doi">10.1002/alz.045184</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Leduc</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Jasmin-B&#xe9;langer</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Poirier</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>APOE and cholesterol homeostasis in Alzheimer&#x27;s disease</article-title>. <source>Trends Mol. Med.</source> <volume>16</volume> (<issue>10</issue>), <fpage>469</fpage>&#x2013;<lpage>477</lpage>. <pub-id pub-id-type="doi">10.1016/j.molmed.2010.07.008</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>N. N.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Learning knowledge graph embedding with heterogeneous relation attention networks</article-title>. <source>IEEE Trans. Neural Netw. Learn Syst.</source> <volume>33</volume> (<issue>8</issue>), <fpage>3961</fpage>&#x2013;<lpage>3973</lpage>. <pub-id pub-id-type="doi">10.1109/tnnls.2021.3055147</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Nie</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.-F.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Anisotropic angle distribution learning for head pose estimation and attention understanding in human-computer interaction</article-title>. <source>Neurocomputing</source> <volume>433</volume>, <fpage>310</fpage>&#x2013;<lpage>322</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2020.09.068</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Meng</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Weng</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Wen</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Explainable deep transfer learning model for disease risk prediction using high-dimensional genomic data</article-title>. <source>PLoS Comput. Biol.</source> <volume>18</volume> (<issue>7</issue>), <fpage>e1010328</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pcbi.1010328</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Deep learning face attributes in the wild</article-title>,&#x201d; in <source>Proceedings of the IEEE international conference on computer vision</source>, <fpage>3730</fpage>&#x2013;<lpage>3738</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2015.425</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>L&#xf6;ffler</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Flunkert</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Havas</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Schweinzer</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Uger</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Windisch</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Neuroinflammation and related neuropathologies in APPSL mice: further value of this <italic>in vivo</italic> model of Alzheimer&#x27;s disease</article-title>. <source>J. Neuroinflammation</source> <volume>11</volume>, <fpage>84</fpage>. <pub-id pub-id-type="doi">10.1186/1742-2094-11-84</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Long</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>P. S.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Learning multiple tasks with multilinear relationship networks</article-title>. <source>Adv. neural Inf. Process. Syst.</source> <volume>30</volume>. <pub-id pub-id-type="doi">10.5555/3294771.3294923</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Zhai</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Javidi</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Feris</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Fully-adaptive feature sharing in multi-task networks with applications in person attribute classification</article-title>,&#x201d; in <source>Proceedings of the IEEE conference on computer vision and pattern recognition</source>, <fpage>5334</fpage>&#x2013;<lpage>5343</lpage>.</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mattson</surname>
<given-names>M. P.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Pathways towards and away from Alzheimer&#x27;s disease</article-title>. <source>Nature</source> <volume>430</volume> (<issue>7000</issue>), <fpage>631</fpage>&#x2013;<lpage>639</lpage>. <pub-id pub-id-type="doi">10.1038/nature02621</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Misra</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Shrivastava</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gupta</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hebert</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Cross-stitch networks for multi-task learning</article-title>,&#x201d; in <source>Proceedings of the IEEE conference on computer vision and pattern recognition</source>, <fpage>3994</fpage>&#x2013;<lpage>4003</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2016.433</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mueller</surname>
<given-names>S. G.</given-names>
</name>
<name>
<surname>Weiner</surname>
<given-names>M. W.</given-names>
</name>
<name>
<surname>Thal</surname>
<given-names>L. J.</given-names>
</name>
<name>
<surname>Petersen</surname>
<given-names>R. C.</given-names>
</name>
<name>
<surname>Jack</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Jagust</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2005</year>). <article-title>The Alzheimer&#x2019;s disease neuroimaging initiative</article-title>. <source>Neuroimaging Clin. N. Am.</source> <volume>15</volume> (<issue>4</issue>), <fpage>869</fpage>&#x2013;<lpage>877</lpage>. <pub-id pub-id-type="doi">10.1016/j.nic.2005.09.008</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Obozinski</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Taskar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Jordan</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Multi-task feature selection</article-title>. <source>Stat. Dep. UC Berkeley, Tech. Rep.</source> <volume>2</volume> (<issue>2.2</issue>), <fpage>2</fpage>.</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Petersen</surname>
<given-names>R. C.</given-names>
</name>
<name>
<surname>Wiste</surname>
<given-names>H. J.</given-names>
</name>
<name>
<surname>Weigand</surname>
<given-names>S. D.</given-names>
</name>
<name>
<surname>Fields</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Geda</surname>
<given-names>Y. E.</given-names>
</name>
<name>
<surname>Graff-Radford</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>NIA-AA Alzheimer&#x27;s disease framework: clinical characterization of stages</article-title>. <source>Ann. Neurol.</source> <volume>89</volume> (<issue>6</issue>), <fpage>1145</fpage>&#x2013;<lpage>1156</lpage>. <pub-id pub-id-type="doi">10.1002/ana.26071</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pinto</surname>
<given-names>T. C. C.</given-names>
</name>
<name>
<surname>Machado</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Bulgacov</surname>
<given-names>T. M.</given-names>
</name>
<name>
<surname>Rodrigues-J&#xfa;nior</surname>
<given-names>A. L.</given-names>
</name>
<name>
<surname>Costa</surname>
<given-names>M. L. G.</given-names>
</name>
<name>
<surname>Ximenes</surname>
<given-names>R. C. C.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Is the Montreal cognitive assessment (MoCA) screening superior to the mini-mental state examination (MMSE) in the detection of mild cognitive impairment (MCI) and Alzheimer&#x27;s disease (AD) in the elderly?</article-title> <source>Int. Psychogeriatr.</source> <volume>31</volume> (<issue>4</issue>), <fpage>491</fpage>&#x2013;<lpage>504</lpage>. <pub-id pub-id-type="doi">10.1017/s1041610218001370</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Prasad</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Rao</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Amyloid clearance defect in ApoE4 astrocytes is reversed by epigenetic correction of endosomal pH</article-title>. <source>Proc. Natl. Acad. Sci. U. S. A.</source> <volume>115</volume> (<issue>28</issue>), <fpage>E6640</fpage>&#x2013;<lpage>E6649</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1801612115</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Reinitz</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>E. Y.</given-names>
</name>
<name>
<surname>Nicolis di Robilant</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Chuluun</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Antony</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>R. C.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Inhibiting USP16 rescues stem cell aging and memory in an Alzheimer&#x27;s model</article-title>. <source>Elife</source> <volume>11</volume>, <fpage>e66037</fpage>. <pub-id pub-id-type="doi">10.7554/eLife.66037</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roses</surname>
<given-names>A. D.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>An inherited variable poly-T repeat genotype in TOMM40 in Alzheimer disease</article-title>. <source>Arch. Neurol.</source> <volume>67</volume> (<issue>5</issue>), <fpage>536</fpage>&#x2013;<lpage>541</lpage>. <pub-id pub-id-type="doi">10.1001/archneurol.2010.88</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Santiago</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Bottero</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Potashkin</surname>
<given-names>J. A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Transcriptomic and network analysis highlight the association of diabetes at different stages of Alzheimer&#x27;s disease</article-title>. <source>Front. Neurosci.</source> <volume>13</volume>, <fpage>1273</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2019.01273</pub-id>
</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tasaki</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gaiteri</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Petyuk</surname>
<given-names>V. A.</given-names>
</name>
<name>
<surname>Blizinsky</surname>
<given-names>K. D.</given-names>
</name>
<name>
<surname>De Jager</surname>
<given-names>P. L.</given-names>
</name>
<name>
<surname>Buchman</surname>
<given-names>A. S.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Genetic risk for Alzheimer&#x27;s dementia predicts motor deficits through multi-omic systems in older adults</article-title>. <source>Transl. Psychiatry</source> <volume>9</volume> (<issue>1</issue>), <fpage>241</fpage>. <pub-id pub-id-type="doi">10.1038/s41398-019-0577-4</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vandenhende</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Georgoulis</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Van Gansbeke</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Proesmans</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Dai</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Van Gool</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Multi-task learning for dense prediction tasks: a survey</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>44</volume> (<issue>7</issue>), <fpage>3614</fpage>&#x2013;<lpage>3633</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2021.3054719</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Nie</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Nho</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Risacher</surname>
<given-names>S. L.</given-names>
</name>
<etal/>
</person-group> (<year>2012</year>). <article-title>Identifying quantitative trait loci via group-sparse multitask regression and feature selection: an imaging genetics study of the ADNI cohort</article-title>. <source>Bioinformatics</source> <volume>28</volume> (<issue>2</issue>), <fpage>229</fpage>&#x2013;<lpage>237</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btr649</pub-id>
</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Gim</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Qiao</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Cho</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Elston</surname>
<given-names>R. C.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Family-based rare variant association analysis: a fast and efficient method of multivariate phenotype association analysis</article-title>. <source>Genet. Epidemiol.</source> <volume>40</volume> (<issue>6</issue>), <fpage>502</fpage>&#x2013;<lpage>511</lpage>. <pub-id pub-id-type="doi">10.1002/gepi.21985</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Winer</surname>
<given-names>J. R.</given-names>
</name>
<name>
<surname>Maass</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pressman</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Stiver</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Schonhaut</surname>
<given-names>D. R.</given-names>
</name>
<name>
<surname>Baker</surname>
<given-names>S. L.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Associations between tau, &#x3b2;-amyloid, and cognition in Parkinson disease</article-title>. <source>JAMA Neurol.</source> <volume>75</volume> (<issue>2</issue>), <fpage>227</fpage>&#x2013;<lpage>235</lpage>. <pub-id pub-id-type="doi">10.1001/jamaneurol.2017.3713</pub-id>
</citation>
</ref>
<ref id="B41">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zamir</surname>
<given-names>A. R.</given-names>
</name>
<name>
<surname>Sax</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Guibas</surname>
<given-names>L. J.</given-names>
</name>
<name>
<surname>Malik</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Savarese</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Taskonomy: disentangling task transfer learning</article-title>,&#x201d; in <source>Proceedings of the IEEE conference on computer vision and pattern recognition</source>, <fpage>3712</fpage>&#x2013;<lpage>3722</lpage>. <pub-id pub-id-type="doi">10.24963/ijcai.2019/871</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>X. X.</given-names>
</name>
<name>
<surname>Tian</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>Z. T.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>Y. H.</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Yu</surname>
<given-names>J. T.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>The epidemiology of Alzheimer&#x27;s disease modifiable risk factors and prevention</article-title>. <source>J. Prev. Alzheimers Dis.</source> <volume>8</volume> (<issue>3</issue>), <fpage>313</fpage>&#x2013;<lpage>321</lpage>. <pub-id pub-id-type="doi">10.14283/jpad.2021.15</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Xiong</surname>
<given-names>N. N.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Multi-scale dynamic convolutional network for knowledge graph embedding</article-title>. <source>IEEE Trans. Knowl. Data Eng.</source> <volume>34</volume> (<issue>5</issue>), <fpage>2335</fpage>&#x2013;<lpage>2347</lpage>. <pub-id pub-id-type="doi">10.1109/tkde.2020.3005952</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Multi-source domain adaptation with joint learning for cross-domain sentiment classification</article-title>. <source>Knowledge-Based Syst.</source> <volume>191</volume>, <fpage>105254</fpage>. <pub-id pub-id-type="doi">10.1016/j.knosys.2019.105254</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname>
<given-names>Q.</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Yuan</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Lv</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Pang</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>APOE and APOC1 gene polymorphisms are associated with cognitive impairment progression in Chinese patients with late-onset Alzheimer&#x27;s disease</article-title>. <source>Neural Regen. Res.</source> <volume>9</volume> (<issue>6</issue>), <fpage>653</fpage>&#x2013;<lpage>660</lpage>. <pub-id pub-id-type="doi">10.4103/1673-5374.130117</pub-id>
</citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Panwar</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Dodge</surname>
<given-names>H. H.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Hampstead</surname>
<given-names>B. M.</given-names>
</name>
<name>
<surname>Albin</surname>
<given-names>R. L.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>COMPASS: a computational model to predict changes in MMSE scores 24-months after initial assessment of Alzheimer&#x2019;s disease</article-title>. <source>Sci. Rep.</source> <volume>6</volume> (<issue>1</issue>), <fpage>34567</fpage>. <pub-id pub-id-type="doi">10.1038/srep34567</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Walker</surname>
<given-names>K. A.</given-names>
</name>
<name>
<surname>Zhong</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ghoneim</surname>
<given-names>D. H.</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Z.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>Associations between genetically predicted plasma protein levels and Alzheimer&#x27;s disease risk: a study using genetic prediction models</article-title>. <source>Alzheimers Res. Ther.</source> <volume>16</volume> (<issue>1</issue>), <fpage>8</fpage>. <pub-id pub-id-type="doi">10.1186/s13195-023-01378-4</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>