<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2025.1731062</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Deep multimodal learning for domain-level cognitive decline prediction in Alzheimer&#x00027;s disease</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Garc&#x00131;&#x00301;a-Guti&#x000E9;rrez</surname> <given-names>Fernando</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<uri xlink:href="https://loop.frontiersin.org/people/3253130"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Matias-Guiu</surname> <given-names>Jordi A.</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<uri xlink:href="https://loop.frontiersin.org/people/252696"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Ayala</surname> <given-names>Jos&#x000E9; L.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<uri xlink:href="https://loop.frontiersin.org/people/1604941"/>
</contrib>
<on-behalf-of>The Alzheimer&#x00027;s Disease Neuroimaging Initiative (ADNI)</on-behalf-of>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Computer Architecture and Automation, Universidad Complutense de Madrid</institution>, <city>Madrid</city>, <country country="es">Spain</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Neurology, Hospital Cl&#x00131;&#x00301;nico San Carlos, Instituto de Investigaci&#x000F3;n Sanitaria San Carlos (IdISSC)</institution>, <city>Madrid</city>, <country country="es">Spain</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Fernando Garc&#x00131;&#x00301;a-Guti&#x000E9;rrez, <email xlink:href="mailto:fegarc05@ucm.es">fegarc05@ucm.es</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-12-17">
<day>17</day>
<month>12</month>
<year>2025</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>8</volume>
<elocation-id>1731062</elocation-id>
<history>
<date date-type="received">
<day>23</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>15</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>24</day>
<month>11</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2025 Garc&#x00131;&#x00301;a-Guti&#x000E9;rrez, Matias-Guiu and Ayala.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Garc&#x00131;&#x00301;a-Guti&#x000E9;rrez, Matias-Guiu and Ayala</copyright-holder>
<license>
<ali:license_ref start_date="2025-12-17">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Alzheimer&#x00027;s disease (AD) is characterized by significant variability in clinical progression; however, few studies have focused on developing models to predict cognitive decline. Anticipating these trajectories is essential for patient management, care planning, and developing new treatments. This study explores the potential of artificial intelligence (AI) techniques to model neurocognitive trajectories from multimodal neuroimaging data and further investigates different data representation frameworks.</p>
</sec>
<sec>
<title>Methods</title>
<p>Using information from 653 participants from the Alzheimer&#x00027;s Disease Neuroimaging Initiative (ADNI), we developed models to predict future clinical diagnoses and cognitive decline, both quantitatively (rate of decline) and qualitatively (presence or absence of decline). Input features included structural T1-weighted magnetic resonance imaging (MRI), [<sup>18</sup>F]-fluorodeoxyglucose positron emission tomography (FDG-PET), [<sup>18</sup>F]-florbetapir PET (AV45-PET), neuropsychological assessments, and demographic variables. Several information representation strategies were explored, including tabular data models, convolutional neural networks (CNNs), and graph neural networks (GNNs). Furthermore, to maximize the use of all available information, we proposed a modeling framework that performed modality-specific pre-training to learn feature embeddings, which were then integrated through a late-fusion layer to produce a unified representation for downstream prediction.</p>
</sec>
<sec>
<title>Results</title>
<p>The modeling strategies demonstrated good predictive performance for future clinical diagnoses, consistent with previous studies (F1 = 0.779). Quantitative models explained approximately 29.4%&#x02013;36.0% of the variance in cognitive decline. In the qualitative analysis, the models achieved AUC values above 0.83 when predicting cognitive deterioration in the memory, language, and executive function domains. Architecturally, CNN- and GNN-based models yielded the best performance, and the proposed pre-training strategy consistently improved predictive accuracy.</p>
</sec>
<sec>
<title>Conclusions</title>
<p>This study demonstrates that AI techniques can capture patterns of cognitive decline by exploiting multimodal neuroimaging data. These findings contribute to the development of more precise phenotyping approaches for neurodegenerative patterns in AD.</p>
</sec></abstract>
<kwd-group>
<kwd>automated pattern recognition</kwd>
<kwd>artificial intelligence</kwd>
<kwd>machine learning</kwd>
<kwd>neuroimaging</kwd>
<kwd>neurodegenerative diseases</kwd>
<kwd>Alzheimer&#x00027;s disease</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declare that financial support was received for the research and/or publication of this article. Jordi A. Matias-Guiu was supported by the Instituto de Salud Carlos III, Spain through the project INT20/00079 and INT23/00017 (co-funded by European Regional Development Fund &#x0201C;A way to make Europe&#x0201D;).</funding-statement>
</funding-group>
<counts>
<fig-count count="5"/>
<table-count count="3"/>
<equation-count count="7"/>
<ref-count count="83"/>
<page-count count="18"/>
<word-count count="12535"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Medicine and Public Health</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Alzheimer&#x00027;s disease (AD) represents the predominant cause of dementia worldwide (<xref ref-type="bibr" rid="B4">Alzheimer&#x00027;s Association, 2024</xref>). In its typical course, AD progresses gradually, beginning with a prolonged prodromal phase characterized by subtle memory impairments. As the disease advances, cognitive deficits become more pronounced and extend to other domains, eventually resulting in significant functional decline (<xref ref-type="bibr" rid="B59">Pe&#x000F1;a-Casanova et al., 2012</xref>).</p>
<p>Nevertheless, the progression of AD exhibits considerable interindividual variability. This heterogeneity is influenced by a combination of environment, genetic predisposition, and levels of cognitive reserve, among other factors (<xref ref-type="bibr" rid="B33">Hersi et al., 2017</xref>; <xref ref-type="bibr" rid="B18">Duara and Barker, 2022</xref>). Moreover, the presence of copathology further increases the variability in neurocognitive trajectories (<xref ref-type="bibr" rid="B46">Lam et al., 2013</xref>).</p>
<p>In this context, despite advancements in diagnostic and early detection methods (<xref ref-type="bibr" rid="B5">Ansart et al., 2021</xref>; <xref ref-type="bibr" rid="B26">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al., 2022</xref>, <xref ref-type="bibr" rid="B25">2024a</xref>; <xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>; <xref ref-type="bibr" rid="B47">Lee et al., 2024</xref>), tools for estimating cognitive decline remain underdeveloped. This limitation is particularly critical, as accurate trajectory prediction is essential for clinical decision-making, therapeutic planning, and patient selection for clinical trials (<xref ref-type="bibr" rid="B2">Abdelnour et al., 2022</xref>; <xref ref-type="bibr" rid="B32">Hampel et al., 2022</xref>). In fact, anticipating disease progression can substantially reduce the number of participants needed in clinical trials, with corresponding reductions in cost and patient exposure (<xref ref-type="bibr" rid="B18">Duara and Barker, 2022</xref>).</p>
<p>Accordingly, much of the existing research on modeling cognitive decline has adopted a qualitative approach, primarily focusing on transitions between clinical stages, such as from mild cognitive impairment (MCI) to dementia, with reported classification accuracies ranging from 75% to 95% (<xref ref-type="bibr" rid="B5">Ansart et al., 2021</xref>; <xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>; <xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al., 2024b</xref>; <xref ref-type="bibr" rid="B80">Yang et al., 2025</xref>). However, predicting diagnostic transitions tends to oversimplify the complexity of cognitive decline, as it overlooks the finer-grained dynamics of domain-specific cognitive trajectories.</p>
<p>In contrast, a smaller subset of studies has aimed to model cognitive decline quantitatively. These efforts typically involve forecasting future scores on a limited number of neuropsychological assessments within defined time windows (<xref ref-type="bibr" rid="B14">Dansson et al., 2021</xref>; <xref ref-type="bibr" rid="B48">Lei et al., 2022</xref>; <xref ref-type="bibr" rid="B51">Maheux et al., 2023</xref>; <xref ref-type="bibr" rid="B17">Devanarayan et al., 2024</xref>). Commonly targeted measures include the Mini-Mental State Examination (MMSE), the Clinical Dementia Rating (CDR), and the cognitive sub-scale of the Alzheimer&#x00027;s Disease Assessment Scale (ADAS-Cog), often in the context of the TADPOLE challenge (<xref ref-type="bibr" rid="B52">Marinescu et al., 2019</xref>). Notably, recent studies have expanded the range of neuropsychological tests considered, incorporating measures such as the Rey Auditory Verbal Learning Test (RAVLT) (<xref ref-type="bibr" rid="B69">Seo et al., 2024</xref>), the Preclinical Alzheimer&#x00027;s Cognitive Composite (<xref ref-type="bibr" rid="B16">Devanarayan et al., 2025</xref>), and composite cognitive domains (<xref ref-type="bibr" rid="B54">Moradi et al., 2025</xref>). These advancements represent meaningful progress toward modeling cognitive decline with enhanced clinical relevance.</p>
<p>Furthermore, although these studies utilize a wide range of biomarkers for modeling, the representation of neuroimaging data remains restrictive. Typically, features are extracted from a single imaging modality, or occasionally from multiple modalities, and treated as tabular data. However, this ignores spatial and topological information inherent to brain structure (<xref ref-type="bibr" rid="B51">Maheux et al., 2023</xref>; <xref ref-type="bibr" rid="B17">Devanarayan et al., 2024</xref>).</p>
<p>This approach contrasts with other applications in which the superiority of artificial intelligence (AI) models incorporating three-dimensional or graph-based information has been well established (<xref ref-type="bibr" rid="B50">Li et al., 2021</xref>; <xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>). This is particularly relevant given the inherent complexity of modeling cognitive decline, a task considerably more challenging than diagnosis or conversion prediction. Enhancing predictive performance in this setting requires more effective utilization of neuroimaging data, one of the most informative biomarkers (<xref ref-type="bibr" rid="B57">Nordberg et al., 2010</xref>).</p>
<p>In this context, we aim to model domain-specific cognitive trajectories and to predict diagnostic transitions between clinical stages of the AD continuum. To this end, we introduce a multimodal learning framework that exploits complementary data representations and maximizes the use of available neuroimaging information.</p>
<p>Whereas many prior approaches rely primarily on global cognitive assessments such as the MMSE, our method specifically targets the core cognitive domains typically affected in AD, including memory, language, visuospatial abilities, and executive functions. Furthermore, to obtain a more robust quantification of cognitive decline and to mitigate the noise associated with single time-point measurements, we focus on modeling longitudinal trajectories of cognitive performance.</p>
<p>Methodologically, the proposed framework incorporates two key strategies to improve predictive performance. First, we introduce a pre-training stage to exploit the full dataset, thereby alleviating common limitations associated with high-dimensional neuroimaging data and relatively small sample sizes. Second, we systematically explore alternative data representations to enhance model accuracy and efficiency. Specifically, we evaluate tabular features derived from multiple neuroimaging modalities, leverage three-dimensional information through convolutional neural networks (CNNs), and incorporate brain connectivity patterns via graph neural networks (GNNs).</p>
<p>Through this approach, we aim not only to enhance the predictive accuracy of cognitive decline but also to advance toward more detailed and clinically relevant modeling. This approach has the potential to optimize intervention strategies, support treatment personalization, and improve patient selection for clinical trials.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Methodology</title>
<sec>
<label>2.1</label>
<title>Study cohort</title>
<p>This study used data from the Alzheimer&#x00027;s Disease Neuroimaging Initiative (ADNI) accessed on April 28, 2024 (<ext-link ext-link-type="uri" xlink:href="https://adni.loni.usc.edu">https://adni.loni.usc.edu</ext-link>). Initiated in 2003, ADNI represents a collaborative effort aimed at exploring whether a combination of serial magnetic resonance imaging (MRI) and positron emission tomography (PET) scans, along with other biomarkers and clinical/neuropsychological evaluations, can effectively track the progression of MCI and early AD.</p>
<p>Models were developed to predict future diagnoses at two and four years, as well as cognitive decline (as defined in Section 2.4). The dataset used to model cognitive decline included 653 subjects. In contrast, for predicting future diagnoses, not all participants had diagnostic information available within a window of at least six months relative to each prediction point. Consequently, 519 and 354 subjects were included in the two- and four-year diagnosis prediction tasks, respectively.</p>
<p>Selection criteria included the availability of neuropsychological assessments, MRI, [<sup>18</sup>F]-fluorodeoxyglucose PET (FDG), and [<sup>18</sup>F]-florbetapir (AV45) PET data acquired within a three-month window, as well as a minimum follow-up period of two years from the initial neuropsychological visit. Additionally, since dementia patients already exhibit severe cognitive impairment, only individuals clinically diagnosed as cognitively normal (CN) or as having MCI at baseline were included. The socio-demographic characteristics of the sample are presented in <xref ref-type="table" rid="T1">Table 1</xref>.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Clinical and socio-demographic variables of the sample used for modeling cognitive decline.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Diagnosis</bold></th>
<th valign="top" align="center"><bold>CN</bold></th>
<th valign="top" align="center"><bold>MCI</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Sample size (N, %)</td>
<td valign="top" align="center">267 (40.9)</td>
<td valign="top" align="center">386 (59.1)</td>
</tr>
<tr>
<td valign="top" align="left">Age (Mean, SD)</td>
<td valign="top" align="center">74.4 (6.3)</td>
<td valign="top" align="center">72.3 (7.4)</td>
</tr>
<tr>
<td valign="top" align="left">Sex (% Female)</td>
<td valign="top" align="center">51.7</td>
<td valign="top" align="center">40.9</td>
</tr>
<tr>
<td valign="top" align="left">Years of formal education (Mean, SD)</td>
<td valign="top" align="center">16.8 (2.4)</td>
<td valign="top" align="center">16.2 (2.7)</td>
</tr>
<tr>
<td valign="top" align="left">MMSE (Mean, SD)</td>
<td valign="top" align="center">29.0 (1.2)</td>
<td valign="top" align="center">27.9 (1.8)</td>
</tr>
<tr>
<td valign="top" align="left">Years follow-up (Mean, SD)</td>
<td valign="top" align="center">6.6 (3.6)</td>
<td valign="top" align="center">5.3 (3.2)</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>CN, cognitively normal; MCI, mild cognitive impairment; MMSE, Mini-Mental State Examination; SD, standard deviation.</p>
</table-wrap-foot>
</table-wrap>
<p>Moreover, the modeling framework presented in Section 2.5 includes a modality-specific pre-training step. For this pre-training, data from 1,668 subjects with associated neuroimaging information&#x02014;diagnosed as CN, MCI, or dementia&#x02014;were used.</p>
</sec>
<sec>
<label>2.2</label>
<title>Neuroimaging data</title>
<p>Raw high-resolution (1.5T and 3T) T1-weighted MRI magnetization-prepared rapid gradient echo (MPRAGE), FDG, and AV45 scans were processed. MRI and PET images were paired if acquired within three months of the corresponding MRI scan.</p>
<p>Firstly, brain extraction was performed on the MRI scans using <italic>SynthStrip</italic> (<xref ref-type="bibr" rid="B34">Hoopes et al., 2022</xref>), and bias field inhomogeneity was corrected using the N4 algorithm from Advanced Normalization Tools (ANTs) (<xref ref-type="bibr" rid="B73">Tustison et al., 2010</xref>). Then, gray matter (GM) and white matter (WM) were segmented using the <italic>New Segment</italic> tool in Statistical Parametric Mapping 12 (<ext-link ext-link-type="uri" xlink:href="https://www.fil.ion.ucl.ac.uk/spm/">SPM12</ext-link>), implemented in MATLAB R2023b (MathWorks Inc.). Since bias correction had already been performed with ANTs, no additional bias regularization was applied in <italic>New Segment</italic>. Segmentation was performed using the default SPM12 tissue probability maps, with all other parameters left unchanged.</p>
<p>Separately, <italic>dynamic</italic> PET acquisitions were converted to <italic>static</italic> images following ADNI preprocessing protocols (<xref ref-type="bibr" rid="B38">Jagust et al., 2015</xref>). Static PET images were first pre-aligned to their corresponding MRI scans using FSL, applying a rigid-body transformation with normalized mutual information as the cost function. The PET images were then co-registered to the MRI scans using the <italic>Co-register</italic> function in SPM12, again using normalized mutual information as the cost function and applying a 3rd-degree spline interpolation with all other parameters set to default.</p>
<p>Thereafter, MRI scans were normalized using SPM12 to the MNI152 template (isotropic 1 mm<sup>3</sup> voxel size) using a 7th-degree spline interpolation. The deformation fields obtained during normalization were applied to the PET acquisitions, along with GM and WM masks, transforming all images into the same reference space.</p>
<p>Afterwards, spatial smoothing with a full width at half maximum (FWHM) of 4 mm was performed using an isotropic Gaussian kernel. Additionally, for PET images, standardized uptake value ratios (SUVRs) were calculated using the whole cerebellum as the reference region (<xref ref-type="bibr" rid="B19">Dukart et al., 2010</xref>; <xref ref-type="bibr" rid="B71">Shekari et al., 2024</xref>).</p>
<p>Finally, outlier detection methods were applied to the processed images (see <xref ref-type="supplementary-material" rid="SM1">Appendix 1</xref>), followed by manual inspection to identify potential errors in the preprocessing pipeline. Cases exhibiting acquisition-related artifacts or processing failures were either reprocessed or excluded from the dataset. The final dataset used for modeling (Section 2.5) consisted of 3,861 MRI scans, 2,548 FDG scans, and 1,403 AV45 scans.</p>
</sec>
<sec>
<label>2.3</label>
<title>Neuropsychological data</title>
<p>Neurocognitive performance was evaluated using neuropsychological composite scores covering four cognitive domains: memory, language, executive function, and visuospatial abilities. These composites were derived via structural equation modeling, following established methodologies (<xref ref-type="bibr" rid="B58">Park et al., 2012</xref>; <xref ref-type="bibr" rid="B28">Gibbons et al., 2012</xref>; <xref ref-type="bibr" rid="B13">Crane et al., 2012</xref>; <xref ref-type="bibr" rid="B11">Choi et al., 2020</xref>).</p>
<p>For the memory domain, subscores from the RAVLT (trials 1, 2, and 6, 30-min delayed recall, and recognition) were included, along with word recall, delayed recall, and recognition sub-scores from the ADAS-Cog (<xref ref-type="bibr" rid="B64">Rey, 1958</xref>; <xref ref-type="bibr" rid="B53">Mohs et al., 1997</xref>). Language ability was assessed using the Category Fluency Test (<xref ref-type="bibr" rid="B55">Morris et al., 1989</xref>), the Naming Objects and Fingers item, and word-finding difficulty ratings from the ADAS-Cog, as well as the Boston Naming Test (BNT) (<xref ref-type="bibr" rid="B41">Kaplan et al., 2001</xref>). Executive function was characterized using scores from the Trail Making Test (log-transformed completion time in seconds) (<xref ref-type="bibr" rid="B63">Reitan and Wolfson, 2014</xref>) and the ADAS-Cog attention subscore (number cancellation). Visuospatial abilities were evaluated using the total score from the Clock Drawing Test (copy and draw) (<xref ref-type="bibr" rid="B29">Goodglass et al., 2001</xref>), and the constructive and ideational praxis subscores from the ADAS-Cog.</p>
<p>Missing data (&#x0003C; 4%, except for BNT) were imputed using multivariate imputation by chained equations (<xref ref-type="bibr" rid="B77">White et al., 2011</xref>), employing a Random Forest (RF) as a surrogate model. Given that the BNT total score was only available for ADNI 1/GO/2 cohorts, this resulted in 20.4% missingness (16.5% in the longitudinal subset), and a separate imputation strategy was employed. In this case, missing values were imputed using a RF model, incorporating the remaining neuropsychological tests, age, sex, and years of education as predictor variables. Experimental validation (<xref ref-type="supplementary-material" rid="SM1">Appendix 2</xref>) showed a correlation exceeding 0.75 with the observed values and a mean absolute error of 2.2 for this approach.</p>
<p>Composite scores were estimated by fixing the latent factor variance to one for model identification (<xref ref-type="bibr" rid="B31">Hair et al., 2021</xref>). The executive function, language, and visuospatial composites were computed using the weighted least squares mean and variance adjusted estimator, treating the ADAS-Cog attention number cancellation, the language-related items, and visuospatial sub-scores as ordinal categorical variables. For the memory composite, the robust maximum likelihood estimator was used.</p>
<p>To enhance interpretability, the resulting composite scores were scaled according to <xref ref-type="disp-formula" rid="EQ1">Equation 1</xref>, where a value of 0 corresponded to the median of individuals with mild dementia (<inline-formula><mml:math id="M1"><mml:msubsup><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>50</mml:mn></mml:mrow><mml:mrow><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>), and a value of 100 corresponded to the median of CN subjects (<inline-formula><mml:math id="M2"><mml:msubsup><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>50</mml:mn></mml:mrow><mml:mrow><mml:mi>c</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>):</p>
<disp-formula id="EQ1"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">adj</mml:mtext></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mfrac><mml:mrow><mml:mi>x</mml:mi><mml:mo>-</mml:mo><mml:msubsup><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>50</mml:mn></mml:mrow><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mtext class="textrm" mathvariant="normal">dem</mml:mtext></mml:mrow></mml:msubsup></mml:mrow><mml:mrow><mml:msubsup><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>50</mml:mn></mml:mrow><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mtext class="textrm" mathvariant="normal">cn</mml:mtext></mml:mrow></mml:msubsup><mml:mo>-</mml:mo><mml:msubsup><mml:mrow><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mn>50</mml:mn></mml:mrow><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mtext class="textrm" mathvariant="normal">dem</mml:mtext></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>&#x000B7;</mml:mo><mml:mn>100</mml:mn></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(1)</label></disp-formula>
<p>Lastly, all composite scores were adjusted for age and educational level using normative data. Normative subjects were defined as those with: a CN diagnosis sustained across all visits, a CDR of 0.0, a MMSE score greater than 26, and negative amyloid status based on the SUVR threshold defined in <xref ref-type="bibr" rid="B39">Johnson et al. (2013)</xref>.</p>
<p>Data imputation was performed using Python (v3.11.5) and the <italic>Scikit-Learn</italic> (v1.6.1) library (<xref ref-type="bibr" rid="B60">Pedregosa et al., 2011</xref>), while composite score calculations were conducted in R (v4.4.2) with the <italic>lavaan</italic> (v0.6-19) package (<xref ref-type="bibr" rid="B67">Rosseel, 2012</xref>). For additional details on composite score definitions, computations, and analysis, refer to <xref ref-type="supplementary-material" rid="SM1">Appendix 3</xref>.</p>
</sec>
<sec>
<label>2.4</label>
<title>Definition of cognitive decline</title>
<p>This study aimed to develop models for predicting cognitive decline across the cognitive domains described in Section 2.3. Cognitive decline was defined using two complementary criteria: (1) a quantitative measure, reflecting the annual rate of cognitive function deterioration, and (2) a qualitative measure, determined as the likelihood of an individual experiencing significant cognitive decline over time. These definitions were designed to reduce the variability inherent in single-time-point assessments of cognitive performance, by incorporating more robust longitudinal information on neurocognitive trajectories.</p>
<p>Quantitative cognitive decline was defined based on the slope of cognitive performance over time, measured from the neuroimaging acquisition used as model input (<xref ref-type="bibr" rid="B23">Franzmeier et al., 2020</xref>). Alternatively, the binary classification of cognitive decline (<italic>stable</italic> vs. <italic>decliner</italic>) was determined using domain-specific cut-off points. These thresholds were defined based on the 5th percentile of cognitive decline observed in individuals with a clinical diagnosis of CN and sustained amyloid-negative status over time. Accordingly, an individual was classified as exhibiting abnormal cognitive decline if the individual&#x00027;s rate of decline fell below that of 95% of control subjects.</p>
<p><xref ref-type="supplementary-material" rid="SM1">Supplementary Table A3</xref> summarizes the statistics of cognitive decline and their relationship to diagnostic changes.</p>
</sec>
<sec>
<label>2.5</label>
<title>Machine learning modeling</title>
<p><xref ref-type="fig" rid="F1">Figure 1</xref> provides an overview of the modeling framework. The following subsections describe the different strategies used for modeling. These include the use of multimodal models (Section 2.5.1), as commonly developed in the literature (<xref ref-type="bibr" rid="B23">Franzmeier et al., 2020</xref>; <xref ref-type="bibr" rid="B56">Nguyen et al., 2020</xref>; <xref ref-type="bibr" rid="B30">Gravina et al., 2024</xref>), as well as the main contribution of this paper, which incorporates a pre-training step and late fusion to maximize the use of the available information (Section 2.5.2). Finally, the different parameterizations used to exploit the neuroimaging data are described in Sections 2.5.3&#x02013;2.5.5.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Overview of the multimodal learning framework. <bold>(a)</bold> <italic>Standard multimodal training</italic>. Processed 3D neuroimaging data from all modalities (MRI, FDG-PET, and AV45-PET) are first mapped into structured representations through the mapping function &#x003D5; (e.g., tabular, 3D, or graph-based formats), and subsequently processed by a joint image-embedding model. In parallel, demographic and neuropsychological variables are encoded by an auxiliary embedding model. The resulting imaging and auxiliary embeddings are then concatenated and passed to a fusion network to produce the final outputs. <bold>(b)</bold> <italic>Multimodal training with pre-trained embedding models</italic>. Each imaging modality undergoes an independent multitask pre-training phase to learn modality-specific embedding models using all available data, including subjects with incomplete modalities or longitudinal follow-up data. The pre-trained modality-specific models are then frozen and used to generate fixed modality-specific embeddings which, together with embeddings derived from demographic and neuropsychological information, are concatenated and fed into a multimodal fusion network to generate the final outputs.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1731062-g0001.tif">
<alt-text content-type="machine-generated">Four confusion matrices showing binary classification results for cognitive functions. (a) Memory: 124 stable, 20 mislabeled, 13 missed decline, 34 correctly declined. (b) Language: 123 stable, 19 mislabeled, 11 missed decline, 38 correctly declined. (c) Visuospatial: 96 stable, 11 mislabeled, 44 missed decline, 40 correctly declined. (d) Executive: 117 stable, 23 mislabeled, 17 missed decline, 34 correctly declined. Each matrix compares real versus predicted stability and decline.</alt-text>
</graphic>
</fig>
<sec>
<label>2.5.1</label>
<title>Multimodal data modeling</title>
<p>The primary objective of this study was to approximate each target variable, <italic>y</italic>, from a set of three-dimensional images, <inline-formula><mml:math id="M4"><mml:mstyle mathvariant="bold"><mml:mtext>X</mml:mtext></mml:mstyle><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>M</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup></mml:math></inline-formula>, derived from <italic>M</italic> modalities. In general terms, the goal was to approximate:</p>
<disp-formula id="EQ2"><mml:math id="M5"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>y</mml:mi><mml:mo>&#x02248;</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003D5;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>X</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(2)</label></disp-formula>
<p>where &#x003D5; represents a mapping function, and &#x003B8; denotes the learnable parameters of the predictive model <italic>f</italic>. In this context, the role of &#x003D5; was to adapt the three-dimensional data to match the expected input dimensions of <italic>f</italic><sub>&#x003B8;</sub>. Furthermore, depending on the nature of each target variable <italic>y</italic>, continuous or categorical, models were trained to perform either regression or classification tasks.</p>
<p>In addition, to incorporate auxiliary information into the model, such as demographic or neuropsychological data, an information fusion layer was introduced. In this case, given tabular input data <inline-formula><mml:math id="M6"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mstyle></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msup></mml:math></inline-formula>, an auxiliary model <inline-formula><mml:math id="M7"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> was defined, and the original formulation <italic>f</italic><sub>&#x003B8;</sub> was partitioned into an output model <inline-formula><mml:math id="M8"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, and an embedding model, <inline-formula><mml:math id="M9"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, incorporating a multi-head attention mechanism (MHA) (<xref ref-type="bibr" rid="B74">Vaswani et al., 2017</xref>) to fuse the information (<xref ref-type="fig" rid="F1">Figure 1a</xref>):</p>
<disp-formula id="EQ3"><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>y</mml:mi><mml:mo>&#x02248;</mml:mo><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mtext>MHA</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo class="qopname">stack</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>&#x003D5;</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>X</mml:mtext></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(3)</label></disp-formula>
</sec>
<sec>
<label>2.5.2</label>
<title>Modeling with pre-training</title>
<sec>
<label>2.5.2.1</label>
<title>Pre-training strategy</title>
<p>Since not all subjects had data available for every modality or sufficient longitudinal follow-up with neuropsychological assessments, an alternative formulation to <xref ref-type="disp-formula" rid="EQ2">Equation 2</xref>, incorporating a pre-training step, was introduced. The goal of this strategy was to maximize the use of all available information.</p>
<p>At this stage, each modality <italic>m</italic> was considered independently, and a multitask model was trained to estimate a set of <italic>d</italic> target variables related to the main study objectives, denoted by <bold>y</bold>&#x02032; and approximated as follows:</p>
<disp-formula id="EQ4"><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>&#x02248;</mml:mo><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">pre-train</mml:mtext></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>X</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>.</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(4)</label></disp-formula>
<p>Here, <inline-formula><mml:math id="M12"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">pre-train</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> denotes a model used exclusively during the pre-training phase, based on a shallow feed-forward network (FFN) with multitask projection heads. Meanwhile, <inline-formula><mml:math id="M13"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> represents the modality-specific component of the model that undergoes pre-training. In this formulation, <inline-formula><mml:math id="M14"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is responsible for learning the representation associated with modality <italic>m</italic>, with &#x003D5;<sup><italic>m</italic></sup> denoting its corresponding mapping function.</p>
<p>Therefore, the pre-training loss function was formulated as follows:</p>
<disp-formula id="EQ5"><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>L</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">pre-train</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup><mml:mo>,</mml:mo><mml:msup><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>B</mml:mi></mml:mstyle></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>L</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">BCE</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;</mml:mtext><mml:mo>&#x0002B;</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>C</mml:mi></mml:mstyle></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>L</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">CE</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x000A0;&#x000A0;&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;&#x02003;</mml:mtext><mml:mo>&#x0002B;</mml:mo><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>R</mml:mi></mml:mstyle></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-script"><mml:mi>L</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">Huber</mml:mtext></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>r</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(5)</label></disp-formula>
<p>where &#x02112;<sup>BCE</sup> denotes the binary cross-entropy loss computed over the &#x0212C; binary classification tasks; &#x02112;<sup>CE</sup> corresponds to the categorical cross-entropy loss calculated for the &#x1D49E; multiclass classification problems; and &#x02112;<sup>Huber</sup> refers to the Huber loss used for the &#x0211B; regression tasks. The vector <bold>w</bold> contains the weights assigned to each task in <bold>y</bold>&#x02032;. Moreover, to handle missing values, a mask was applied to the entries of the predicted output vector &#x00177;&#x02032;, ignoring target variables for which ground truth data were unavailable for a given subject.</p>
</sec>
<sec>
<label>2.5.2.2</label>
<title>Fusion model for multimodal integration</title>
<p>After pre-training, the learned embedding representations&#x02014;defined as <inline-formula><mml:math id="M16"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>X</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula>&#x02014;from each imaging modality were integrated using a fusion model:</p>
<disp-formula id="EQ6"><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mi>y</mml:mi><mml:mo>&#x02248;</mml:mo><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mrow></mml:msubsup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mtext>MHA</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo class="qopname">stack</mml:mo><mml:mrow><mml:mo>[</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">mri</mml:mtext></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">mri</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">fdg</mml:mtext></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">fdg</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">amy</mml:mtext></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">amy</mml:mtext></mml:mrow></mml:msub></mml:mrow><mml:mo>]</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>,</mml:mo></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(6)</label></disp-formula>
<p>where <bold>W</bold><sub><italic>m</italic></sub> represented linear projections that mapped the embeddings to a common latent space of dimension <italic>D</italic>; and <inline-formula><mml:math id="M18"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> was the fine-tuned model for each of the addressed objectives. Specifically, <inline-formula><mml:math id="M19"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> consisted of a FFN, which incorporated either a softmax activation function for multiclass classification tasks or linear projections for regression tasks.</p>
<p>Similar to <xref ref-type="disp-formula" rid="EQ3">Equation 3</xref>, demographic and neuropsychological data were optionally integrated into the fusion models. However, due to the low dimensionality of these features, they were not considered during the pre-training phase. Instead, the information was processed through a simple FFN and subsequently incorporated into the fusion model (<xref ref-type="disp-formula" rid="EQ6">Equation 6</xref>), stacked with the rest of the embedding projections (<xref ref-type="fig" rid="F1">Figure 1b</xref>).</p>
</sec>
</sec>
<sec>
<label>2.5.3</label>
<title>Feed-forward network based models</title>
<p>Feed-forward network-based models represent the simplest neural network architecture. These models apply a sequence of linear transformations followed by nonlinear activation functions, optionally incorporating dropout and normalization layers to stabilize training.</p>
<p>Following the framework presented in previous sections, the mapping function &#x003D5; was designed to adapt the three-dimensional data to the FFNs. This function was defined to aggregate the information into regions of interest (ROIs) using the Automated Anatomical Labeling (AAL) brain atlas (<xref ref-type="bibr" rid="B66">Rolls et al., 2020</xref>). Specifically, for each ROI, the &#x003D5; function calculated the GM and WM volumes for the MRI data (<inline-formula><mml:math id="M20"><mml:msup><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">mri</mml:mtext></mml:mstyle></mml:mrow></mml:msup><mml:mo>:</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>&#x02192;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mn>232</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>) and computed the mean and standard deviation for the SUVR values of the PET data (<inline-formula><mml:math id="M21"><mml:msup><mml:mrow><mml:mi>&#x003D5;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fdg</mml:mtext></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">amy</mml:mtext></mml:mstyle></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>:</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>x</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>y</mml:mi></mml:mrow></mml:msub><mml:mo>&#x000D7;</mml:mo><mml:msub><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>z</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msup><mml:mo>&#x02192;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:mn>232</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula>).</p>
<p>In this parametrization, the models <inline-formula><mml:math id="M22"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M23"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> were implemented as stacks of multiple layers. For <inline-formula><mml:math id="M24"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, information from the different modalities was concatenated into a single vector. During experimentation, various hyperparameters were explored, including the number and width of the layers, the activation functions used, dropout rates, and the application of batch normalization. Details of the hyperparameter configurations are provided in <xref ref-type="supplementary-material" rid="SM1">Appendix 4</xref>.</p>
</sec>
<sec>
<label>2.5.4</label>
<title>3D convolutional network based models</title>
<p>Three-dimensional CNNs extend classical CNNs, originally designed for the two-dimensional imaging domain, by accommodating an additional dimension. The core idea behind these models is the use of convolution and pooling operations, which enable parameter sharing and the generation of equivariant representations. These properties make 3D-CNNs efficient and robust for processing high-dimensional neuroimaging data (<xref ref-type="bibr" rid="B72">Spasov et al., 2019</xref>; <xref ref-type="bibr" rid="B61">Peng et al., 2021</xref>; <xref ref-type="bibr" rid="B30">Gravina et al., 2024</xref>).</p>
<p>In this approach, to reduce the memory requirements associated with the models, the &#x003D5; functions were implemented by downsampling the image dimensions by a factor of 2 using a continuous interpolation. Additionally, to mitigate the heterogeneity of the MRI scans, the &#x003D5;<sup>mri</sup> function performed image normalization by dividing the intensity values by the median WM value.</p>
<p>For the model architectures evaluated for <inline-formula><mml:math id="M25"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M26"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, two distinct parameterizations were considered. The first followed the scaffold proposed in <xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al. (2024b)</xref>, while the second was based on a DenseNet architecture (<xref ref-type="bibr" rid="B35">Huang et al., 2017</xref>) adapted for processing 3D data. In the specific case of <inline-formula><mml:math id="M27"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> (model without pre-training, <xref ref-type="disp-formula" rid="EQ3">Equation 3</xref>), images from different modalities were provided as different input channels. During the experiments, various architectural configurations, normalization layers, and dropout rates were explored.</p>
</sec>
<sec>
<label>2.5.5</label>
<title>Graph neural network based models</title>
<p>Unlike the models from previous sections applied to data within a Euclidean domain, models based on GNNs operate on graphs. Formally, a graph &#x1D4A2;(&#x1D4B1;, &#x02130;) is defined as a set of &#x1D4B1; nodes containing <italic>d</italic> node features, <bold>X</bold>&#x02208;&#x0211D;<sup>|&#x1D4B1;| &#x000D7; <italic>d</italic></sup>, connected by a set of &#x02130; edges. The aim of the GNNs is to generate a series of node embeddings <bold>h</bold><sub><italic>u</italic></sub>&#x02200;<italic>u</italic>&#x02208;&#x1D4B1; by applying the following operations over <italic>k</italic> message-passing layers:</p>
<disp-formula id="EQ7"><mml:math id="M28"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>&#x0002B;</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mtext class="textrm" mathvariant="normal">UPDATE</mml:mtext></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mtext class="textrm" mathvariant="normal">AGGREGATE&#x000A0;</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>v</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mtext>&#x000A0;</mml:mtext><mml:mo>&#x02200;</mml:mo><mml:mi>v</mml:mi><mml:mo>&#x02208;</mml:mo><mml:mstyle mathvariant="script"><mml:mi>N</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(7)</label></disp-formula>
<p>where UPDATE and AGGREGATE are differentiable functions that determine how the information is aggregated and updated, &#x1D4A9;(<italic>u</italic>) represents the neighborhood of node <italic>u</italic>, and <inline-formula><mml:math id="M29"><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>h</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>X</mml:mi></mml:mrow><mml:mrow><mml:mi>u</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula>. Thus, in each iteration, each node aggregates information from its neighborhood, capturing both the structural properties of &#x1D4A2; and the features of neighboring nodes (<xref ref-type="bibr" rid="B12">Corso et al., 2024</xref>).</p>
<p>In this context, to define the input graphs of the models, the graphical lasso method was used (<xref ref-type="bibr" rid="B37">Huang et al., 2010</xref>). Briefly, this method estimates the inverse covariance matrix by applying <italic>L</italic>1 regularization, weighted by &#x003BB; (<xref ref-type="bibr" rid="B24">Friedman et al., 2008</xref>). To fit the model, FDG data from all CN and MCI participants were used. As in Section 2.5.3, brain metabolism data were aggregated into ROIs using the AAL atlas, and the mean SUVR values were used as model input. The estimation resulted in an undirected unweighted connectivity graph with 116 nodes, exhibiting a specific pattern of connections regulated by &#x003BB;. In the experiments, &#x003BB; was one of the hyperparameters explored.</p>
<p>Accordingly, the mapping function &#x003D5; was defined to map the three-dimensional image information onto a graph domain. Specifically, the &#x003D5;<sup>mri</sup> function used GM and WM values as node features, while the &#x003D5;<sup>{fdg, amy}</sup> functions incorporated the mean and standard deviation of the SUVR values.</p>
<p>The models <inline-formula><mml:math id="M30"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M31"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> were implemented according to the formulation of <xref ref-type="disp-formula" rid="EQ7">Equation 7</xref>. In the <inline-formula><mml:math id="M32"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> model, features from different neuroimaging modalities were concatenated into a single vector and treated as node features. At the architectural level, different types of aggregation and update layers were explored, including Graph Convolutional Networks (GCNs) with symmetric normalization (<xref ref-type="bibr" rid="B44">Kipf and Welling, 2016</xref>); Graph Attention Networks (GATs) (<xref ref-type="bibr" rid="B75">Veli&#x0010D;kovi&#x00107; et al., 2017</xref>), which included self-loops and followed the attention mechanism of <xref ref-type="bibr" rid="B7">Bahdanau et al. (2014)</xref>; and Graph Isomorphism Networks (GINs) (<xref ref-type="bibr" rid="B78">Xu et al., 2018a</xref>).</p>
<p>The explored model hyperparameters included the use of jumping knowledge connections (<xref ref-type="bibr" rid="B79">Xu et al., 2018b</xref>), mean and sum graph pooling operators, the number of message-passing layers, and the size of the embedding representations. Furthermore, all models incorporated graph normalization (<xref ref-type="bibr" rid="B9">Cai et al., 2021</xref>), applied a dropout rate of 0.1, and used ELU as the activation function. For more details on the implementation of the models, see <xref ref-type="supplementary-material" rid="SM1">Appendix 4</xref>.</p>
</sec>
</sec>
<sec>
<label>2.6</label>
<title>Experimental setup</title>
<sec>
<label>2.6.1</label>
<title>Model evaluation</title>
<p>To evaluate the models, the data were split into training (60%), validation (10%), and test sets (30%). The training set was used for model fitting, the validation set for hyperparameter tuning and model selection, and the test set for final evaluation. For model pre-training, the data were divided into training (85%) and validation (15%), ensuring that no test data were used in this step.</p>
<p>To evaluate the results and select the models, the Matthews Correlation Coefficient (MCC) (<xref ref-type="bibr" rid="B10">Chicco and Jurman, 2023</xref>) was used as the primary metric for classification tasks, and the explained variance (EV) for regression tasks. For the final evaluation, additional metrics were also considered, including the F1-score (weighted-macro F1-score for multiclass problems) and the area under the curve (AUC) for classification, and Pearson&#x00027;s correlation coefficient for regression.</p>
</sec>
<sec>
<label>2.6.2</label>
<title>Optimization and implementation details</title>
<p>Section 2.5 described the different parameterizations used for <inline-formula><mml:math id="M33"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">emb</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> and <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula>, which were subjected to hyperparameter optimization as outlined in <xref ref-type="supplementary-material" rid="SM1">Appendix 4</xref>. Implementation details of the components <inline-formula><mml:math id="M35"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, <inline-formula><mml:math id="M36"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, <inline-formula><mml:math id="M37"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">pre-train</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula> and MHA (<xref ref-type="disp-formula" rid="EQ3">Equations 3</xref>, <xref ref-type="disp-formula" rid="EQ6">6</xref>), along with aspects related to model fitting, are provided below:</p>
<p><inline-formula><mml:math id="M38"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>: this component of the model consisted of a four-layer FFN with hidden layer dimensions of 64, 48, 32, and 16, respectively. SiLU activation functions were used (<xref ref-type="bibr" rid="B20">Elfwing et al., 2018</xref>), and the activation function of the output layer was adapted based on the specific task. Batch normalization was applied to the first four layers, and a dropout rate of 0.1 was used in the first layer.</p>
<p><inline-formula><mml:math id="M39"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">aux</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>: this module was implemented as a two-layer FFN with a hidden dimension of 32, using SiLU activation functions. A dropout rate of 0.1 was applied to the first layer, and batch normalization was incorporated in both layers. Additionally, a linear projection was included to align the output dimensionality with that required by the embeddings in <xref ref-type="disp-formula" rid="EQ3">Equations 3</xref>, <xref ref-type="disp-formula" rid="EQ6">6</xref>.</p>
<p><inline-formula><mml:math id="M40"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">pre-train</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>: this component was designed as a multitask adapter to support the pre-training stage of <inline-formula><mml:math id="M41"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> (<xref ref-type="disp-formula" rid="EQ4">Equation 4</xref>). Its architecture was similar to <inline-formula><mml:math id="M42"><mml:msubsup><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x003B8;</mml:mi></mml:mrow><mml:mrow><mml:mstyle class="text"><mml:mtext class="textrm" mathvariant="normal">fusion</mml:mtext></mml:mstyle></mml:mrow></mml:msubsup></mml:math></inline-formula>, with the exception that the final two layers, with dimensions 32 and 16, were replicated <italic>d</italic> times to serve as task-specific heads.</p>
<p><italic>Multi-head attention</italic> (MHA): the attention component of <xref ref-type="disp-formula" rid="EQ3">Equations 3</xref>, <xref ref-type="disp-formula" rid="EQ6">6</xref> consisted of an 8-head MHA layer with an embedding dimension of 16. A dropout rate of 0.1 was applied, and the resulting attention-weighted embeddings were aggregated via mean pooling.</p>
<p>All models were trained using backpropagation and optimized with the Adam algorithm (<xref ref-type="bibr" rid="B43">Kingma and Ba, 2014</xref>). The learning rate, batch size, and number of training epochs were treated as tunable hyperparameters. A learning rate scheduler was applied to all models, reducing the learning rate by a factor of 0.5 every 20% of the total number of epochs. Early stopping was also employed, terminating training when the validation loss failed to improve for a number of consecutive epochs corresponding to 10% of the total training duration.</p>
<p>For simplicity, in <xref ref-type="disp-formula" rid="EQ5">Equation 5</xref>, all task weights were set to 1, and the Huber loss function was used with a fixed &#x003B4; value of 2.5. During pre-training, the predicted targets included the current diagnosis, the diagnosis at two and four years, and quantitative measures of cognitive decline.</p>
<p>Moreover, for modeling quantitative cognitive decline, the target variables were standardized to z-scores using the mean and standard deviation of the training set. Prior to standardization, outliers were trimmed based on the 5th and 95th percentiles. Similarly, standardization (without trimming) was applied to the input data for all models. In the models that included neuropsychological and demographic information, the tests considered as input were those described in Section 2.3, and the demographic variables were age, sex, and educational level.</p>
<p>All developed models were compared with RF baselines trained with (i) neuropsychological and demographic data only, (ii) neuroimaging data only, and (iii) a combination of neuropsychological, demographic, and neuroimaging data.</p>
<p>Experiments were performed using an NVIDIA RTX A5000 (CUDA 12.6). The models were developed in Python (v3.11.5), using the PyTorch (v2.5.1) and PyTorch Geometric (v2.5.1) libraries. The code used for data preparation and model implementation is provided on <ext-link ext-link-type="uri" xlink:href="https://github.com/FernandoGaGu/Domain-cognitive-decline-prediction">GitHub</ext-link>. For more details on the hyperparameters explored for each type of model, see <xref ref-type="supplementary-material" rid="SM1">Appendix 4</xref>.</p>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<label>3</label>
<title>Results</title>
<p>The following sections present the results of future diagnosis and cognitive decline prediction. Models were labeled as <italic>ni</italic> (non-imaging) when neuroimaging data were excluded, <italic>nd</italic> when neuropsychological and demographic information were included, and <italic>pt</italic> when pre-training was incorporated. By default, if none of these labels were specified, it was assumed that the models had been trained using only neuroimaging data.</p>
<sec>
<label>3.1</label>
<title>Diagnosis prediction at 2 years</title>
<p>In the two-year diagnostic prediction task, the best performance was achieved by models that incorporated both neuropsychological and demographic information, as well as a pre-training phase (<xref ref-type="fig" rid="F2">Figure 2a</xref>). Notably, CNN- and GNN-based architectures outperformed all other models. The CNN-pt-nd model yielded the highest performance (MCC 0.580, F1 0.745), followed closely by the GNN-pt-nd model (MCC 0.557, F1 0.736) (<xref ref-type="fig" rid="F2">Figure 2c</xref>).</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>Comparison of the Matthews Correlation Coefficient (MCC) of the best models <bold>(a, b)</bold> and confusion matrices associated with the best model <bold>(c, d)</bold> for predicting future diagnoses at <bold>(a, c)</bold> 2 years and <bold>(b, d)</bold> 4 years. Higher MCC values indicate better predictive performance. The figure presents results from the best-performing models obtained by excluding neuroimaging data (no-imaging), incorporating neuropsychological and demographic information (np &#x00026; demo), and including a pre-training step (pre-trained).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1731062-g0002.tif">
<alt-text content-type="machine-generated">Four line graphs show correlation (Corr) versus sample proportion for four cognitive domains: (a) Memory, (b) Language, (c) Visuospatial, and (d) Executive. Each graph depicts a blue line with error bars, indicating correlation increases with sample proportion. Red stars represent CNN-nd-pt values: 0.572 for Memory, 0.596 for Language, 0.527 for Visuospatial, and 0.601 for Executive.</alt-text>
</graphic>
</fig>
<p>In contrast, the baseline RF model trained exclusively on neuroimaging data performed the worst (MCC 0.101, F1 0.460). However, performance improved markedly when neuropsychological and demographic features were included, either alone or in combination with neuroimaging data (MCC 0.478, F1 0.677).</p>
<p>Similarly, FFNs, CNNs, and GNNs trained solely on neuroimaging data without pre-training achieved modest results (e.g., GNN: MCC 0.222, F1 0.544), although they still outperformed the neuroimaging-only RF baseline. When neuropsychological and demographic features were added to these models, performance improved substantially (e.g., CNN-nd: MCC 0.439, F1 0.665), reaching levels comparable to the RF-nd model.</p>
</sec>
<sec>
<label>3.2</label>
<title>Diagnosis prediction at 4 years</title>
<p>In the four-year diagnostic prediction task, models consistently exhibited greater predictive capacity compared to the 2-year results (<xref ref-type="fig" rid="F2">Figure 2b</xref>).</p>
<p>Within the baseline models, the best performance was achieved by the RF-ni model (MCC 0.429, F1 0.653), while those relying solely on neuroimaging data without pre-training showed the weakest performance (MCC &#x0003C; 0.4). In contrast, CNN and GNN models incorporating neuropsychological and demographic information outperformed all RF-based baselines, with the CNN-nd model achieving an MCC of 0.468 and F1 of 0.672.</p>
<p>Across all configurations, pre-trained models consistently outperformed models without pre-training. Notably, CNNs demonstrated the highest predictive accuracy, with the CNN-pt-nd model achieving the highest performance (MCC 0.646, F1 0.779) (<xref ref-type="fig" rid="F2">Figure 2d</xref>).</p>
</sec>
<sec>
<label>3.3</label>
<title>Detection of diagnosis transitions</title>
<p>Predictive models for dementia conversion are widely explored in the literature (<xref ref-type="bibr" rid="B5">Ansart et al., 2021</xref>; <xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>; <xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al., 2024b</xref>; <xref ref-type="bibr" rid="B47">Lee et al., 2024</xref>), typically aiming to distinguish individuals with MCI who progress to dementia (pMCI) from those who remain stable (sMCI).</p>
<p>In contrast, this study directly modeled future clinical diagnoses without explicitly targeting conversion events. Nevertheless, the model outputs enabled retrospective assessment of the models&#x00027; ability to identify dementia conversion and other diagnostic transitions.</p>
<p>In this context, the best-performing model (CNN-pt-nd) achieved an AUC of 0.772 and an F1 of 0.629 for detecting sMCI/pMCI at two years, and an AUC of 0.909 with an F1 of 0.800 at 4 years.</p>
<p>Similarly, for identifying diagnostic transitions over a maximum period of 4 years&#x02014;such as transitions from CN to MCI or dementia, and from MCI to dementia&#x02014;the best model reached an AUC of 0.795 and an F1 of 0.645.</p>
<p><xref ref-type="supplementary-material" rid="SM1">Appendix 6</xref> shows the results obtained for the different models explored in the study.</p>
</sec>
<sec>
<label>3.4</label>
<title>Quantitative prediction of cognitive decline</title>
<p><xref ref-type="table" rid="T2">Table 2</xref> summarizes the performance of the evaluated models in predicting continuous cognitive decline (see Section 2.4). Additional metrics are detailed in <xref ref-type="supplementary-material" rid="SM1">Appendix 6</xref>.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Explained variance of regression models predicting continuous measures of cognitive decline.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Memory</bold></th>
<th valign="top" align="center"><bold>Language</bold></th>
<th valign="top" align="center"><bold>Visuospatial</bold></th>
<th valign="top" align="center"><bold>Executive</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">RF-ni</td>
<td valign="top" align="center">0.06 [&#x02013;0.04, 0.14]</td>
<td valign="top" align="center">0.20 [0.08, 0.30]</td>
<td valign="top" align="center">0.26 [0.16, 0.35]</td>
<td valign="top" align="center">0.12 [&#x02013;0.05, 0.26]</td>
</tr>
<tr>
<td valign="top" align="left">RF</td>
<td valign="top" align="center">0.17 [0.08, 0.26]</td>
<td valign="top" align="center">0.11 [&#x02013;0.00, 0.21]</td>
<td valign="top" align="center">0.10 [&#x02013;0.00, 0.19]</td>
<td valign="top" align="center">0.20 [0.08, 0.29]</td>
</tr>
<tr>
<td valign="top" align="left">FFN</td>
<td valign="top" align="center">0.07 [&#x02013;0.06, 0.19]</td>
<td valign="top" align="center">0.18 [0.02, 0.32]</td>
<td valign="top" align="center">0.06 [&#x02013;0.08, 0.17]</td>
<td valign="top" align="center">0.21 [0.09, 0.33]</td>
</tr>
<tr>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">0.11 [0.03, 0.18]</td>
<td valign="top" align="center">0.09 [0.01, 0.17]</td>
<td valign="top" align="center">0.07 [&#x02013;0.03, 0.16]</td>
<td valign="top" align="center">0.12 [&#x02013;0.02, 0.22]</td>
</tr>
<tr>
<td valign="top" align="left">GNN</td>
<td valign="top" align="center">0.19 [0.09, 0.29]</td>
<td valign="top" align="center">0.11 [0.01, 0.21]</td>
<td valign="top" align="center">0.12 [0.01, 0.22]</td>
<td valign="top" align="center">0.21 [0.09, 0.33]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-pt</td>
<td valign="top" align="center">0.19 [&#x02013;0.01, 0.36]</td>
<td valign="top" align="center">0.21 [0.08, 0.34]</td>
<td valign="top" align="center">0.11 [&#x02013;0.00, 0.19]</td>
<td valign="top" align="center">0.23 [0.11, 0.34]</td>
</tr>
<tr>
<td valign="top" align="left">CNN-pt</td>
<td valign="top" align="center">0.28 [0.12, 0.41]</td>
<td valign="top" align="center">0.34 [0.18, 0.47]</td>
<td valign="top" align="center">0.26 [0.13, 0.37]</td>
<td valign="top" align="center">0.34 [0.21, 0.45]</td>
</tr>
<tr>
<td valign="top" align="left">GNN-pt</td>
<td valign="top" align="center">0.28 [0.15, 0.39]</td>
<td valign="top" align="center">0.31 [0.13, 0.46]</td>
<td valign="top" align="center">0.23 [0.10, 0.35]</td>
<td valign="top" align="center">0.35 [0.16, 0.51]</td>
</tr>
<tr>
<td valign="top" align="left">RF-nd</td>
<td valign="top" align="center">0.19 [0.08, 0.29]</td>
<td valign="top" align="center">0.16 [0.05, 0.26]</td>
<td valign="top" align="center">0.26 [0.17, 0.34]</td>
<td valign="top" align="center">0.23 [0.09, 0.34]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-nd</td>
<td valign="top" align="center">0.15 [0.02, 0.26]</td>
<td valign="top" align="center">0.14 [&#x02013;0.01, 0.26]</td>
<td valign="top" align="center">0.11 [&#x02013;0.02, 0.24]</td>
<td valign="top" align="center">0.14 [&#x02013;0.02, 0.27]</td>
</tr>
<tr>
<td valign="top" align="left">CNN-nd</td>
<td valign="top" align="center">0.13 [&#x02013;0.01, 0.25]</td>
<td valign="top" align="center">0.19 [0.02, 0.31]</td>
<td valign="top" align="center">0.24 [0.10, 0.36]</td>
<td valign="top" align="center">0.16 [&#x02013;0.01, 0.29]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>GNN-nd</bold></td>
<td valign="top" align="center">0.17 [0.03, 0.28]</td>
<td valign="top" align="center">0.19 [0.04, 0.33]</td>
<td valign="top" align="center"><bold>0.31 [0.18, 0.43]</bold></td>
<td valign="top" align="center">0.29 [0.14, 0.41]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-pt-nd</td>
<td valign="top" align="center">0.21 [0.07, 0.35]</td>
<td valign="top" align="center">0.25 [0.12, 0.37]</td>
<td valign="top" align="center">0.16 [0.03, 0.28]</td>
<td valign="top" align="center">0.26 [0.14, 0.37]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>CNN-pt-nd</bold></td>
<td valign="top" align="center"><bold>0.29 [0.15, 0.42]</bold></td>
<td valign="top" align="center"><bold>0.35 [0.20, 0.47]</bold></td>
<td valign="top" align="center">0.28 [0.15, 0.40]</td>
<td valign="top" align="center"><bold>0.36 [0.21, 0.49]</bold></td>
</tr>
<tr>
<td valign="top" align="left">GNN-pt-nd</td>
<td valign="top" align="center">0.28 [0.14, 0.40]</td>
<td valign="top" align="center">0.31 [0.17, 0.43]</td>
<td valign="top" align="center">0.24 [0.10, 0.35]</td>
<td valign="top" align="center">0.35 [0.20, 0.47]</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The 95% confidence intervals, presented in square brackets, were estimated from 1,000 bootstrap iterations on the test set. The best models for each task are highlighted in bold. <italic>ni</italic>, no-imaging; <italic>pt</italic>, models with pre-training; <italic>nd</italic>, use of neuropsychology and demographic information.</p>
</table-wrap-foot>
</table-wrap>
<p>Across all cognitive domains, with the exception of visuospatial, models that incorporated the pre-training step generally outperformed those without it. Moreover, CNN-based models consistently yielded the highest predictive performance, followed by GNN-based models, while tabular models (FFN and RF) exhibited the weakest results.</p>
<p>In the memory domain, the CNN-pt-nd achieved the best results, with an EV of 0.294 and a correlation of 0.572 between predicted and actual scores (<xref ref-type="fig" rid="F3">Figure 3a</xref>). Other models with strong performance&#x02014;including CNN-pt, GNN-pt-nd, and GNN-pt&#x02014;also exceeded an EV of 0.25. These models clearly outperformed the RF baseline using only neuropsychological and demographic data (EV 0.062), with the remaining models ranging between EVs of 0.073 and 0.212.</p>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>Predicted values (y-axis) vs. actual values (x-axis) from the best-performing regression models for predicting cognitive decline across the evaluated cognitive domains: <bold>(a)</bold> memory, <bold>(b)</bold> language, <bold>(c)</bold> visuospatial abilities, and <bold>(d)</bold> executive functions. Cognitive decline ratios are expressed as z-scores. The plot includes the regression line along with 95% confidence intervals. &#x003C1;, Pearson correlation coefficient; <italic>R</italic><sup>2</sup>, coefficient of determination.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1731062-g0003.tif">
<alt-text content-type="machine-generated">Scatter plots showing the relationship between Y true and Y pred for four cognitive domains: (a) Memory, (b) Language, (c) Visuospatial, (d) Executive. Each plot includes a red trend line with shaded confidence interval, demonstrating positive correlation. Correlation coefficients and R-squared values are: (a) &#x003C1; = 0.572, R&#x000B2; = 0.281; (b) &#x003C1; = 0.596, R&#x000B2; = 0.338; (c) &#x003C1; = 0.559, R&#x000B2; = 0.296; (d) &#x003C1; = 0.601, R&#x000B2; = 0.358.</alt-text>
</graphic>
</fig>
<p>In the language domain, the CNN-pt-nd again led in performance (EV 0.346, correlation 0.596; <xref ref-type="fig" rid="F3">Figure 3b</xref>). As in the memory domain, CNN and GNN models achieved EVs above 0.3, while FFN models with pre-training showed moderate performance (EV &#x0003E;0.2). The rest of the models performed poorly, with EVs below 0.2.</p>
<p>In the visuospatial domain, models also achieved EVs above 0.3 (<xref ref-type="fig" rid="F3">Figure 3c</xref>), although the advantage of pre-training was less evident. Interestingly, the RF model excluding neuroimaging data performed competitively (EV 0.259), ranking among the top models. CNN variants&#x02014;CNN-pt-nd, CNN-pt, and CNN-nd&#x02014;followed closely (EVs 0.277, 0.255, and 0.239, respectively). In contrast, models trained solely on neuroimaging features performed poorly (EV &#x0003C; 0.124).</p>
<p>Finally, the executive functions domain yielded the highest EV values across all tasks (<xref ref-type="fig" rid="F3">Figure 3d</xref>). The CNN-pt-nd again achieved the best performance (EV 0.360, correlation 0.601), followed by the GNN-pt, which attained the highest correlation (0.611) and an EV of 0.354. Models leveraging pre-training or neuropsychological and demographic information (pt, nd, pt-nd) consistently outperformed the remaining configurations (EV &#x0003E;0.25). Conversely, models without pre-training or based on tabular inputs achieved the worst results. The RF-ni model, which excluded neuroimaging, showed the lowest predictive accuracy.</p>
</sec>
<sec>
<label>3.5</label>
<title>Qualitative prediction of cognitive decline</title>
<p>The results and comparisons of the top-performing models for qualitative classification of cognitive decline are presented in <xref ref-type="table" rid="T3">Table 3</xref>. Further performance metrics are presented in <xref ref-type="supplementary-material" rid="SM1">Appendix 6</xref>.</p>
<table-wrap position="float" id="T3">
<label>Table 3</label>
<caption><p>Matthews correlation coefficient of classification models predicting binary measures of cognitive decline.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Memory</bold></th>
<th valign="top" align="center"><bold>Language</bold></th>
<th valign="top" align="center"><bold>Visuospatial</bold></th>
<th valign="top" align="center"><bold>Executive</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">RF-ni</td>
<td valign="top" align="center">0.31 [0.18, 0.44]</td>
<td valign="top" align="center">0.30 [0.16, 0.43]</td>
<td valign="top" align="center">0.32 [0.19, 0.45]</td>
<td valign="top" align="center">0.31 [0.18, 0.44]</td>
</tr>
<tr>
<td valign="top" align="left">RF</td>
<td valign="top" align="center">0.38 [0.23, 0.52]</td>
<td valign="top" align="center">0.38 [0.23, 0.52]</td>
<td valign="top" align="center">0.23 [0.10, 0.37]</td>
<td valign="top" align="center">0.34 [0.20, 0.45]</td>
</tr>
<tr>
<td valign="top" align="left">FFN</td>
<td valign="top" align="center">0.32 [0.16, 0.45]</td>
<td valign="top" align="center">0.39 [0.25, 0.52]</td>
<td valign="top" align="center">0.28 [0.15, 0.41]</td>
<td valign="top" align="center">0.36 [0.21, 0.51]</td>
</tr>
<tr>
<td valign="top" align="left">CNN</td>
<td valign="top" align="center">0.36 [0.22, 0.51]</td>
<td valign="top" align="center">0.35 [0.22, 0.48]</td>
<td valign="top" align="center">0.17 [0.02, 0.30]</td>
<td valign="top" align="center">0.37 [0.24, 0.49]</td>
</tr>
<tr>
<td valign="top" align="left">GNN</td>
<td valign="top" align="center">0.34 [0.22, 0.45]</td>
<td valign="top" align="center">0.39 [0.24, 0.52]</td>
<td valign="top" align="center">0.29 [0.15, 0.41]</td>
<td valign="top" align="center">0.33 [0.20, 0.46]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-pt</td>
<td valign="top" align="center">0.38 [0.24, 0.50]</td>
<td valign="top" align="center">0.47 [0.33, 0.59]</td>
<td valign="top" align="center">0.40 [0.28, 0.54]</td>
<td valign="top" align="center">0.41 [0.29, 0.53]</td>
</tr>
<tr>
<td valign="top" align="left">CNN-pt</td>
<td valign="top" align="center">0.54 [0.42, 0.65]</td>
<td valign="top" align="center">0.57 [0.44, 0.67]</td>
<td valign="top" align="center">0.35 [0.22, 0.49]</td>
<td valign="top" align="center">0.46 [0.33, 0.57]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>GNN-pt</bold></td>
<td valign="top" align="center">0.52 [0.37, 0.64]</td>
<td valign="top" align="center">0.51 [0.37, 0.63]</td>
<td valign="top" align="center">0.32 [0.18, 0.44]</td>
<td valign="top" align="center"><bold>0.49 [0.34, 0.62]</bold></td>
</tr>
<tr>
<td valign="top" align="left">RF-nd</td>
<td valign="top" align="center">0.39 [0.25, 0.52]</td>
<td valign="top" align="center">0.41 [0.27, 0.55]</td>
<td valign="top" align="center">0.23 [0.11, 0.36]</td>
<td valign="top" align="center">0.39 [0.25, 0.52]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-nd</td>
<td valign="top" align="center">0.35 [0.19, 0.49]</td>
<td valign="top" align="center">0.40 [0.26, 0.53]</td>
<td valign="top" align="center">0.21 [0.08, 0.34]</td>
<td valign="top" align="center">0.41 [0.27, 0.53]</td>
</tr>
<tr>
<td valign="top" align="left">CNN-nd</td>
<td valign="top" align="center">0.31 [0.17, 0.44]</td>
<td valign="top" align="center">0.36 [0.22, 0.49]</td>
<td valign="top" align="center">0.29 [0.17, 0.42]</td>
<td valign="top" align="center">0.33 [0.20, 0.46]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>GNN-nd</bold></td>
<td valign="top" align="center">0.36 [0.21, 0.50]</td>
<td valign="top" align="center">0.37 [0.23, 0.51]</td>
<td valign="top" align="center"><bold>0.42 [0.28, 0.55]</bold></td>
<td valign="top" align="center">0.36 [0.21, 0.48]</td>
</tr>
<tr>
<td valign="top" align="left">FFN-pt-nd</td>
<td valign="top" align="center">0.43 [0.28, 0.57]</td>
<td valign="top" align="center">0.50 [0.35, 0.63]</td>
<td valign="top" align="center">0.32 [0.19, 0.45]</td>
<td valign="top" align="center">0.44 [0.29, 0.56]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>CNN-pt-nd</bold></td>
<td valign="top" align="center">0.48 [0.36, 0.61]</td>
<td valign="top" align="center"><bold>0.61 [0.48, 0.72]</bold></td>
<td valign="top" align="center">0.37 [0.24, 0.50]</td>
<td valign="top" align="center">0.45 [0.33, 0.57]</td>
</tr>
<tr>
<td valign="top" align="left"><bold>GNN-pt-nd</bold></td>
<td valign="top" align="center"><bold>0.56 [0.43, 0.68]</bold></td>
<td valign="top" align="center">0.56 [0.43, 0.68]</td>
<td valign="top" align="center">0.38 [0.24, 0.51]</td>
<td valign="top" align="center">0.48 [0.34, 0.60]</td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>The 95% confidence intervals, presented in square brackets, were estimated from 1,000 bootstrap iterations on the test set. The best models for each task are highlighted in bold. <italic>ni</italic>, no-imaging; <italic>pt</italic>, models with pre-training; <italic>nd</italic>, use of neuropsychology and demographic information.</p>
</table-wrap-foot>
</table-wrap>
<p>Across the cognitive domains of memory, language, and executive functions, models that incorporated the pre-training strategy consistently outperformed those that did not. In particular, GNNs and CNNs demonstrated superior predictive performance compared to models based solely on tabular data.</p>
<p>The highest predictive performance was observed in the language domain, where the pre-trained GNN (GNN-pt) achieved an AUC of 0.882, with a sensitivity and specificity of 86.62% and 77.55% respectively (<xref ref-type="fig" rid="F4">Figure 4b</xref>). For the memory domain, the best-performing model was the GNN-pt-nd, which reached an AUC of 0.840, with a sensitivity of 86.11% and a specificity of 72.34% (<xref ref-type="fig" rid="F4">Figure 4a</xref>). In the executive functions domain, the highest performance was obtained by the GNN-pt model, achieving an AUC of 0.814, with sensitivity and specificity values of 83.57% and 66.67% (<xref ref-type="fig" rid="F4">Figure 4d</xref>).</p>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>Confusion matrices obtained on the test set for the identification of decliners and stable subjects. Results are shown for the best-performing model in each cognitive domain: <bold>(a)</bold> memory, <bold>(b)</bold> language, <bold>(c)</bold> visuospatial abilities, and <bold>(d)</bold> executive functions.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1731062-g0004.tif">
<alt-text content-type="machine-generated">(Image a) Bar chart showing MCC values for different models in 2-year diagnosis, with CNN-pt-nd performing best. (Image b) Similar chart for 4-year diagnosis. (Image c) Confusion matrix for 2-year diagnosis with 62 true negatives for CN. (Image d) Confusion matrix for 4-year diagnosis showing data for CN, MCI, and Dementia predictions. Color codes indicate model types: no-imaging, np &#x00026; demo, and pre-trained.</alt-text>
</graphic>
</fig>
<p>In contrast, the visuospatial abilities domain yielded the lowest predictive performance. Here, the best result was obtained using the GNN without pre-training (GNN-nd), with an AUC of 0.691, sensitivity of 89.72%, and specificity of 47.62% (<xref ref-type="fig" rid="F4">Figure 4c</xref>).</p>
</sec>
<sec>
<label>3.6</label>
<title>Architectural analysis of the best models</title>
<p>This section analyzes the architectures derived from hyperparameter optimization for the models that demonstrated superior performance, specifically those based on GNNs and CNNs with pre-training.</p>
<p>For the CNN-based models, the highest-performing configurations consistently followed the architecture proposed in <xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al. (2024b)</xref>. This architecture comprises eight convolutional layers, incorporating normalization, max pooling, and dropout. Notably, instance normalization significantly outperformed batch normalization across experiments. The top-performing CNN models for both the MRI and FDG imaging modalities adopted this architecture. Each of these CNN models contained approximately 150K trainable parameters.</p>
<p>In contrast, the best-performing pre-trained CNN model for the AV45 modality was based on the DenseNet architecture. This configuration included three bottleneck blocks with 4, 6, and 8 bottleneck layers, respectively, interleaved with transition layers. A growth rate of 4 and a compression factor of 1.0 were used, alongside instance normalization. The total number of trainable parameters in this model was approximately 215K.</p>
<p>For the GNN-based models, hyperparameter optimization favored shallow architectures, typically comprising six message-passing layers and a high channel dimensionality (128 channels). Moreover, the use of jumping knowledge connections with concatenation, combined with average pooling over node embeddings, demonstrated superior performance. Among the evaluated models, the GAT architecture yielded the best results for the MRI and FDG modalities, whereas GIN proved most effective for the AV45 modality.</p>
<p>Regarding &#x003BB;, no consistent trend was observed. For MRI, the best-performing configuration employed a &#x003BB; value of 0.4 (sparsest graph), while the optimal values for FDG and AV45 were 0.3 and 0.2, respectively (densest graph). The approximate number of trainable parameters was 315K for GAT-based models and 250K for GIN-based models.</p>
</sec>
<sec>
<label>3.7</label>
<title>Pre-training evaluation</title>
<p>The results presented in Sections 3.1&#x02013;3.5 indicate that the proposed pre-training strategy consistently yielded substantial performance improvements across all evaluated tasks.</p>
<p>To further examine the influence of pre-training dataset size on model performance&#x02014;and, by extension, the scalability of the strategy with increased data availability&#x02014;learning curves of the CNN-pt-nd model were analyzed.</p>
<p>Specifically, 100 models were pre-trained for each modality using varying proportions of the original pre-training dataset, followed by fine-tuning as described in Section 2.6. In this analysis, no hyperparameter optimization was performed; instead, the configuration that demonstrated the most consistent performance across tasks was selected. Moreover, learning curves were generated exclusively for the quantitative prediction of cognitive decline across different cognitive domains. The resulting curves, assessed in terms of mean correlation, are presented in <xref ref-type="fig" rid="F5">Figure 5</xref>.</p>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>Learning curves for the cognitive domains of <bold>(a)</bold> memory, <bold>(b)</bold> language, <bold>(c)</bold> visuospatial abilities, and <bold>(d)</bold> executive functions. Each curve shows the mean correlation and standard deviation across 100 runs of the CNN-pt-nd model on the test set, obtained using models pre-trained with varying proportions of the original pre-training dataset. The reported result of the best model trained on all data is highlighted in red.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-08-1731062-g0005.tif">
<alt-text content-type="machine-generated">Diagram illustrating two approaches to multimodal training. (a) Standard multimodal training involves processing complete 3D images from MRI, FDG-PET, and AV45-PET. Images undergo various representations and are fed into a joint model producing modality and auxiliary embeddings that are fused for outputs. (b) Multitask pre-training uses data from all subjects, forming separate modality embeddings and outputs. Pre-trained models are then applied to subjects with complete data, using frozen weights for the embeddings, which are stacked and fused for final outputs. Both approaches incorporate neuropsychology and demographic information.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>This study addresses the pressing need for domain-specific prediction of cognitive decline in AD, moving beyond conventional global cognitive assessments. Despite the inherent challenges associated with neuroimaging-based AI applications, such as high data dimensionality and limited sample sizes, our framework demonstrates a strong capacity for modeling individualized cognitive trajectories across distinct cognitive domains. By incorporating a dedicated pre-training step, we significantly enhanced predictive performance, surpassing conventional baselines in capturing the heterogeneous patterns of decline in memory, language, executive function, and visuospatial abilities.</p>
<p>We first evaluated the classification capabilities of our models to assess the reliability of the proposed framework. To this end, we compared our approach with benchmark neuroimaging models reported in recent studies on clinical diagnosis prediction (<xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>; <xref ref-type="bibr" rid="B42">Kaur and Sachdeva, 2025</xref>). The resulting weighted macro&#x02013;F1 scores, 0.745 for the 2-year and 0.779 for the 4-year forecasts, place our models within the range of high-performing neuroimaging classifiers. Most misclassifications occurred between CN and MCI groups (<xref ref-type="fig" rid="F2">Figures 2c</xref>&#x02013;<xref ref-type="fig" rid="F2">d</xref>), which reflects the well-documented clinical ambiguity of these transitional stages (<xref ref-type="bibr" rid="B62">Petersen, 2004</xref>).</p>
<p>Furthermore, although the framework was not specifically optimized for conversion prediction, it achieved robust performance in identifying converters within four years (AUC = 0.909), consistent with current state-of-the-art results (<xref ref-type="bibr" rid="B5">Ansart et al., 2021</xref>; <xref ref-type="bibr" rid="B70">Sharma et al., 2023</xref>; <xref ref-type="bibr" rid="B47">Lee et al., 2024</xref>; <xref ref-type="bibr" rid="B80">Yang et al., 2025</xref>). Performance at two years (AUC = 0.772) was comparatively lower, likely reflecting diagnostic uncertainty and the limited number of conversion events in brief follow-up intervals (<xref ref-type="bibr" rid="B22">Farias et al., 2009</xref>; <xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al., 2024b</xref>). These findings underscore the robustness of our multimodal modeling approach and align with previously reported AUCs ranging from 0.78 to 0.96 in longitudinal AD studies (<xref ref-type="bibr" rid="B3">Abrol et al., 2020</xref>; <xref ref-type="bibr" rid="B76">Wang et al., 2024</xref>; <xref ref-type="bibr" rid="B15">Dao et al., 2025</xref>).</p>
<p>In these diagnostic classification problems, the superior performance observed at the four-year horizon likely stems from intrinsic features of AD progression. Longer follow-up periods mitigate diagnostic noise and class imbalance, allowing disease-related patterns to manifest more clearly, whereas short-term variability is often influenced by individual factors such as cognitive reserve and comorbidities (<xref ref-type="bibr" rid="B22">Farias et al., 2009</xref>; <xref ref-type="bibr" rid="B18">Duara and Barker, 2022</xref>; <xref ref-type="bibr" rid="B2">Abdelnour et al., 2022</xref>). Consequently, extended observation windows provide more stable, pathology-driven predictions.</p>
<p>While these diagnostic results support the validity of our framework, the primary contribution of this study lies in advancing domain-specific modeling of cognitive decline rather than establishing new diagnostic benchmarks. Recent evidence suggests that conversion-prediction models may have reached a performance plateau, indicating that future progress toward clinical translation will depend more on expanding sample sizes, improving cross-cohort generalization, and enhancing model reliability (<xref ref-type="bibr" rid="B21">El-Sappagh et al., 2023</xref>). Within this context, domain-specific modeling of cognitive decline remains comparatively underexplored, despite its critical relevance for early intervention and clinical decision-making (<xref ref-type="bibr" rid="B2">Abdelnour et al., 2022</xref>; <xref ref-type="bibr" rid="B32">Hampel et al., 2022</xref>; <xref ref-type="bibr" rid="B16">Devanarayan et al., 2025</xref>).</p>
<p>Accordingly, our findings demonstrate that patterns of cognitive decline can be inferred from neuroimaging data, particularly through spatial features extracted from brain images. Although the models explained a moderate proportion of the variance in cognitive decline (29.4%&#x02013;36.0%), the predicted trajectories were well correlated with the observed outcomes (correlation &#x0003E; 0.55). In the binary classification setting, the models also exhibited strong performance in identifying individuals whose decline deviated from normative aging trajectories. Notably, AUC values exceeded 0.83 in the memory, language, and executive function domains.</p>
<p>Visuospatial abilities constituted the main exception to these trends. In this domain, performance dropped markedly when cognitive decline was binarized, despite regression models achieving errors comparable to the other domains. This discrepancy is likely driven by the high sensitivity of the binary label to the normative cut-off, because annual change rates cluster near the threshold and small score fluctuations can flip individuals between <italic>stable</italic> and <italic>decliner</italic> categories, adding label noise. In addition, the visuospatial composite relies on tests with limited score range and pronounced ceiling and floor effects, particularly among CN and MCI populations (<xref ref-type="bibr" rid="B45">Kueper et al., 2018</xref>). This further increases the risk that subtle variations are misclassified as categorical change rather than genuine trajectory shifts. Future work should incorporate more fine-grained visuospatial measures and larger normative samples to reduce threshold sensitivity and stabilize qualitative predictions.</p>
<p>To date, few studies based on multimodal neuroimaging have modeled cognitive decline. Moreover, most existing works rely on tabular data and are typically limited to neuropsychological assessments of global cognition, such as the MMSE (<xref ref-type="bibr" rid="B82">Zhang et al., 2012</xref>; <xref ref-type="bibr" rid="B36">Huang et al., 2016</xref>; <xref ref-type="bibr" rid="B23">Franzmeier et al., 2020</xref>; <xref ref-type="bibr" rid="B49">Lei et al., 2020</xref>; <xref ref-type="bibr" rid="B14">Dansson et al., 2021</xref>; <xref ref-type="bibr" rid="B48">Lei et al., 2022</xref>; <xref ref-type="bibr" rid="B51">Maheux et al., 2023</xref>; <xref ref-type="bibr" rid="B17">Devanarayan et al., 2024</xref>). For example, in <xref ref-type="bibr" rid="B23">Franzmeier et al. (2020)</xref>, the authors used MRI, FDG, and amyloid-PET data to predict cognitive decline rates&#x02014;similar to those defined in this study&#x02014;explaining approximately 25% of the variance in global cognition and memory. Similarly, in studies such as <xref ref-type="bibr" rid="B14">Dansson et al. (2021)</xref> and <xref ref-type="bibr" rid="B17">Devanarayan et al. (2024)</xref>, cognitive decline was modeled in terms of MMSE score changes at two and four years, achieving R<sup>2</sup> values of 0.325 and 0.228, respectively. In <xref ref-type="bibr" rid="B82">Zhang et al. (2012)</xref>, correlation coefficients of 0.697 and 0.739 were reported for MMSE and ADAS-Cog. Notably, one of the studies most closely related to our work modeled cognitive decline across the domains of memory, language, executive function, and visuospatial abilities, yielding correlation values between 0.42 and 0.50, lower than those obtained in the present study (<xref ref-type="bibr" rid="B54">Moradi et al., 2025</xref>).</p>
<p>In addition, several studies have focused on modeling future cognitive scores&#x02014;i.e., point estimates of the expected value at a single time point&#x02014;for tests such as the ADAS-Cog, CDR, or MMSE, usually reporting R<sup>2</sup> values above 0.65 (<xref ref-type="bibr" rid="B36">Huang et al., 2016</xref>; <xref ref-type="bibr" rid="B49">Lei et al., 2020</xref>, <xref ref-type="bibr" rid="B48">2022</xref>; <xref ref-type="bibr" rid="B51">Maheux et al., 2023</xref>). However, this point-estimate approach presents limitations when assessing the true predictive capacity of the models, as baseline scores already account for a substantial proportion of the variance. As a result, the reported explained variance may be misleading. For instance, in groups where cognitive scores remain relatively stable over time, a naive strategy that simply reproduces baseline scores can yield high correlations with future outcomes. This highlights one of the key strengths of the approach presented here, which focuses instead on modeling general trends in cognitive trajectories rather than static future scores.</p>
<p>Our analysis also yielded several technical insights. First, models based on CNNs and GNNs consistently outperformed alternative approaches. Incorporating spatial features was particularly advantageous for complex tasks such as cognitive decline prediction and surpassed the performance of the tabular AI models that we implemented. These gains validate the extra implementation effort required to ingest volumetric or surface-based inputs and suggest that future neuro-AI pipelines should treat spatial encoding as a core design criterion rather than an optional add-on.</p>
<p>Second, although CNNs achieved the highest overall performance, their advantage over GNNs was often marginal, indicating that the increased memory demands of CNNs may not always be justified. In scenarios where hospitals deploy models on embedded hardware, or where cloud inference is billed by runtime and memory usage, the difference in resource usage between GNNs and CNNs may outweigh the slight accuracy advantage of CNNs, making GNNs the more pragmatic choice.</p>
<p>Finally, the proposed pre-training strategy consistently enhanced performance across tasks. Moreover, the learning curves suggest that larger pre-training datasets could lead to further improvements, which offer promising opportunities for model development in settings with limited data.</p>
<p>Within this context, the present study represents a substantial advancement over previous work by modeling cognitive decline across specific domains rather than relying solely on global cognition scores. This design enables a more granular characterization of neurodegenerative trajectories, an aspect of considerable clinical relevance.</p>
<p>Methodologically, our work extends previous multimodal fusion approaches (<xref ref-type="bibr" rid="B83">Zhang et al., 2023</xref>; <xref ref-type="bibr" rid="B81">Yu et al., 2024</xref>; <xref ref-type="bibr" rid="B1">Abdelaziz et al., 2025</xref>) by (i) shifting the focus from purely diagnostic endpoints to domain-specific cognitive trajectories, (ii) integrating several imaging modalities together with rich neuropsychological profiles within a unified comparative framework across FFN, CNN, and GNN backbones, and (iii) introducing a modality-specific multi-task pre-training stage that produces trajectory-aware embeddings for subsequent late fusion.</p>
<p>From an applied standpoint, accurately predicting cognitive decline in AD yields several clinically significant benefits. First, it enables genuinely personalized, dynamic therapy, since clinicians could adjust pharmacological or behavioral interventions in real time as the model detects accelerating progression (<xref ref-type="bibr" rid="B40">Jutten et al., 2021</xref>). Second, reliable forecasts support advance-care planning, giving patients and families evidence-based timelines for financial, legal, and caregiving decisions. Third, at the operational level, trajectory prediction helps healthcare systems triage high-risk individuals for closer surveillance and allocate scarce resources, such as specialist memory-clinic appointments, more efficiently (<xref ref-type="bibr" rid="B4">Alzheimer&#x00027;s Association, 2024</xref>).</p>
<p>In the context of drug development, prognostic models could facilitate tighter enrichment of clinical-trial cohorts and more informed endpoint selection. For example, endpoints could be tailored to the specific cognitive domains most likely to decline within a given cohort, thereby lowering required sample sizes and increasing power to detect disease-modifying effects (<xref ref-type="bibr" rid="B40">Jutten et al., 2021</xref>; <xref ref-type="bibr" rid="B18">Duara and Barker, 2022</xref>). Ultimately, domain-level prediction of cognitive decline may be particularly valuable for informing personalized cognitive therapies that target the most at-risk cognitive functions. Collectively, these capabilities have the potential to advance both everyday clinical care and translational AD research.</p>
<p>Despite these promising results, the study has several limitations. Most notably, the predictive performance for cognitive decline remains moderate, which reflects the complexity of neurodegenerative processes. These processes are shaped by multifactorial and stochastic interactions among genetic, environmental, and lifestyle factors, resulting in substantial interindividual variability that is challenging to model with limited sample sizes (<xref ref-type="bibr" rid="B4">Alzheimer&#x00027;s Association, 2024</xref>). To improve model accuracy, increasing cohort sizes and incorporating complementary data sources, such as genetic profiles or plasma biomarkers, will be essential. In this context, federated learning offers a scalable and privacy-preserving approach to integrate multi-site data and enhance model generalizability (<xref ref-type="bibr" rid="B65">Rieke et al., 2020</xref>).</p>
<p>From a methodological perspective, our comparative non-deep learning baseline was an RF model, and we did not include other non-deep multimodal baselines or hybrid ensembles. Given that the available literature on this task is limited and predominantly deep-learning oriented (<xref ref-type="bibr" rid="B42">Kaur and Sachdeva, 2025</xref>), a more comprehensive benchmarking analysis should be undertaken in future work to better contextualize the state of the art in domain-level cognitive-decline prediction.</p>
<p>Another widely recognized limitation that applies to this study, and to much of the current literature, is the difficulty of generalizing models across cohorts. Our analyses were restricted to the ADNI dataset, limiting the findings to this specific population (<xref ref-type="bibr" rid="B68">Samper-Gonz&#x000E1;lez et al., 2017</xref>). This constraint may reduce the clinical applicability of the results, as research cohorts such as ADNI often differ from real-world clinical populations in demographic composition, disease spectrum, and imaging protocols. This issue is particularly relevant given the known variability in the demographics and clinical presentation of AD. Therefore, external validation on independent, multi-site cohorts (e.g., AIBL, OASIS) and, ideally, prospective clinical studies conducted in real-world settings will be essential to establish robustness, generalizability, and clinical utility. Future work should prioritize evaluating the proposed framework in such diverse cohorts as a necessary step toward translational deployment.</p>
<p>Future studies should also investigate whether incorporating longitudinal information, such as serial neuroimaging scans or repeated cognitive assessments, provides meaningful improvements in predictive performance. However, acquiring such data is logistically challenging, because repeated MRI or PET sessions are costly, time-consuming, and consequently rare in routine clinical practice. Rigorous analyses will therefore be required to determine whether the incremental predictive gains offered by longitudinal data justify the substantial logistical and financial burden associated with their acquisition.</p>
<p>Finally, future research should assess the biological validity of model predictions using explainable AI (XAI) techniques tailored to neuroimaging. Demonstrating that <italic>post hoc</italic> attribution methods (e.g., saliency maps, attention weights, or counterfactual lesioning) assign importance to neuroanatomically plausible regions would substantially strengthen the claim that the model leverages disease-relevant signals (<xref ref-type="bibr" rid="B6">Arrieta et al., 2020</xref>; <xref ref-type="bibr" rid="B8">Bloch et al., 2024</xref>). While such interpretability analyses were beyond the scope of the present study, which primarily focused on establishing predictive performance, they constitute a critical avenue for future work. Supporting the feasibility of this approach, previous work from our group has shown that XAI pipelines in related neuroimaging contexts can produce stable attributions in canonical disease-relevant circuits (<xref ref-type="bibr" rid="B27">Garc&#x00301;&#x00131;a-Guti&#x000E9;rrez et al., 2024b</xref>). Conducting these analyses is essential, as they can help clinicians understand the rationale behind model predictions and thereby foster clinical trust, improve patient safety, and facilitate eventual regulatory evaluation.</p>
</sec>
<sec sec-type="conclusions" id="s5">
<label>5</label>
<title>Conclusion</title>
<p>In conclusion, we developed AI-based models using multimodal neuroimaging data and systematically compared alternative information-representation strategies to predict domain-specific cognitive decline in Alzheimer&#x00027;s disease. Theoretically, our findings show that spatially encoded multimodal features can capture individualized trajectories across memory, language, executive, and visuospatial domains, moving beyond conventional global cognition scores and point-estimate forecasting. Practically, our approach provides a scalable tool for stratifying patients according to their risk of accelerated decline in specific cognitive domains, thereby supporting personalized intervention planning, optimized resource allocation, and the enrichment and endpoint selection of clinical trials. By maximizing the use of all available neuroimaging information to generate domain-level trajectory predictions, this study offers both a conceptual and methodological advance in multimodal modeling and contributes to the development of clinically meaningful prognostic tools for neurodegenerative diseases.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The datasets generated and/or analyzed during the current study are available from the ADNI database (<ext-link ext-link-type="uri" xlink:href="https://adni.loni.usc.edu/">https://adni.loni.usc.edu/</ext-link>). The code used for this research is available at: <ext-link ext-link-type="uri" xlink:href="https://github.com/FernandoGaGu/Domain-cognitive-decline-prediction">https://github.com/FernandoGaGu/Domain-cognitive-decline-prediction</ext-link>. Further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="ethics-statement" id="s7">
<title>Ethics statement</title>
<p>This study used only publicly available, de-identified data obtained from the ADNI database. All ADNI participants provided written informed consent, and the ADNI study protocols were approved by the Institutional Review Boards of all participating institutions. No additional ethical approval was required for the secondary analysis of these data.</p>
</sec>
<sec sec-type="author-contributions" id="s8">
<title>Author contributions</title>
<p>FG-G: Writing &#x02013; review &#x00026; editing, Validation, Investigation, Conceptualization, Methodology, Writing &#x02013; original draft, Formal analysis, Visualization, Data curation. JM-G: Supervision, Writing &#x02013; review &#x00026; editing, Resources, Project administration, Conceptualization, Validation, Funding acquisition. JA: Funding acquisition, Resources, Conceptualization, Validation, Investigation, Writing &#x02013; review &#x00026; editing, Project administration, Formal analysis, Supervision.</p>
</sec>
<ack>
<p>Data collection and sharing for this project was funded by the Alzheimer&#x00027;s Disease Neuroimaging Initiative (ADNI) (National Institutes of Health Grant U19 AG024904) and DOD ADNI (Department of Defense award number W81XWH-12-2-0012). ADNI by the National Institute on Aging, the National Institute of Biomedical is funded Imaging and Bioengineering, and through generous contributions from the following: AbbVie, Alzheimer&#x00027;s Association; Alzheimer&#x00027;s Drug Discovery Foundation; Araclon Biotech; BioClinica, Biogen; Bristol-Myers Squibb Company; CereSpir, Inc.; Cogstate; Eisai Inc.; Inc.; Elan Pharmaceuticals, Inc.; Eli Lilly and Company; EuroImmun; F. Hoffmann-La Roche Ltd and its affiliated company Genentech, Inc.; Fujirebio; GE Healthcare; IXICO Ltd.; Janssen Alzheimer Immunotherapy Research &#x00026; Development, LLC.; Johnson &#x00026; Johnson Pharmaceutical Research &#x00026; Development LLC.; Lumosity; Lundbeck; Merck &#x00026; Co., Inc.; Meso Scale Diagnostics, LLC.; NeuroRx Research; Neurotrack Technologies; Novartis Pharmaceuticals Corporation; Pfizer Inc.; Piramal Imaging; Servier; Takeda Pharmaceutical Company; and Transition Therapeutics. The Canadian Institutes of Health Research is providing funds to support ADNI clinical sites in Canada. Private sector contributions are facilitated by the Foundation for the National Institutes of Health (<ext-link ext-link-type="uri" xlink:href="http://www.fnih.org">www.fnih.org</ext-link>). The grantee organization is the Northern California Institute for Research and Education, and the study is coordinated by the Alzheimer&#x00027;s Therapeutic Research Institute at the University of Southern California. ADNI data are disseminated by the Laboratory for Neuro Imaging at the University of Southern California.</p>
</ack>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The author(s) declare that no Gen AI was used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s12">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frai.2025.1731062/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frai.2025.1731062/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/></sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdelaziz</surname> <given-names>M.</given-names></name> <name><surname>Wang</surname> <given-names>T.</given-names></name> <name><surname>Anwaar</surname> <given-names>W.</given-names></name> <name><surname>Elazab</surname> <given-names>A.</given-names></name></person-group> (<year>2025</year>). <article-title>Multi-scale multimodal deep learning framework for Alzheimer&#x00027;s disease diagnosis</article-title>. <source>Comput. Biol. Med</source>. <volume>184</volume>:<fpage>109438</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.109438</pub-id><pub-id pub-id-type="pmid">39579666</pub-id></mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdelnour</surname> <given-names>C.</given-names></name> <name><surname>Agosta</surname> <given-names>F.</given-names></name> <name><surname>Bozzali</surname> <given-names>M.</given-names></name> <name><surname>Foug&#x000E9;re</surname> <given-names>B.</given-names></name> <name><surname>Iwata</surname> <given-names>A.</given-names></name> <name><surname>Nilforooshan</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Perspectives and challenges in patient stratification in Alzheimer&#x00027;s disease</article-title>. <source>Alzheimer&#x00027;s Res. Ther</source>. <volume>14</volume>:<fpage>112</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13195-022-01055-y</pub-id><pub-id pub-id-type="pmid">35964143</pub-id></mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abrol</surname> <given-names>A.</given-names></name> <name><surname>Bhattarai</surname> <given-names>M.</given-names></name> <name><surname>Fedorov</surname> <given-names>A.</given-names></name> <name><surname>Du</surname> <given-names>Y.</given-names></name> <name><surname>Plis</surname> <given-names>S.</given-names></name> <name><surname>Calhoun</surname> <given-names>V.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Deep residual learning for neuroimaging: an application to predict progression to Alzheimer&#x00027;s disease</article-title>. <source>J. Neurosci. Methods</source> <volume>339</volume>:<fpage>108701</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jneumeth.2020.108701</pub-id><pub-id pub-id-type="pmid">32275915</pub-id></mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><collab>Alzheimer&#x00027;s Association</collab></person-group> (<year>2024</year>). <article-title>2024 Alzheimer&#x00027;s disease facts and figures</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>20</volume>, <fpage>3708</fpage>&#x02013;<lpage>3821</lpage>. doi: <pub-id pub-id-type="doi">10.1002/alz.13809</pub-id></mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ansart</surname> <given-names>M.</given-names></name> <name><surname>Epelbaum</surname> <given-names>S.</given-names></name> <name><surname>Bassignana</surname> <given-names>G.</given-names></name> <name><surname>B&#x000F4;ne</surname> <given-names>A.</given-names></name> <name><surname>Bottani</surname> <given-names>S.</given-names></name> <name><surname>Cattai</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Predicting the progression of mild cognitive impairment using machine learning: a systematic, quantitative and critical review</article-title>. <source>Med. Image Anal</source>. <volume>67</volume>:<fpage>101848</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2020.101848</pub-id><pub-id pub-id-type="pmid">33091740</pub-id></mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arrieta</surname> <given-names>A. B.</given-names></name> <name><surname>D&#x000ED;az-Rodr&#x000ED;guez</surname> <given-names>N.</given-names></name> <name><surname>Del Ser</surname> <given-names>J.</given-names></name> <name><surname>Bennetot</surname> <given-names>A.</given-names></name> <name><surname>Tabik</surname> <given-names>S.</given-names></name> <name><surname>Barbado</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Explainable artificial intelligence (xai): concepts, taxonomies, opportunities and challenges toward responsible AI</article-title>. <source>Inform. Fusion</source> <volume>58</volume>, <fpage>82</fpage>&#x02013;<lpage>115</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.inffus.2019.12.012</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bahdanau</surname> <given-names>D.</given-names></name> <name><surname>Cho</surname> <given-names>K.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>2014</year>). <article-title>Neural machine translation by jointly learning to align and translate</article-title>. <source>arXiv preprint arXiv:1409.0473</source>.</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bloch</surname> <given-names>L.</given-names></name> <name><surname>Friedrich</surname> <given-names>C. M.</given-names></name> Alzheimer&#x00027;s Disease Neuroimaging Initiative</person-group> (<year>2024</year>). <article-title>Systematic comparison of 3d deep learning and classical machine learning explanations for Alzheimer&#x00027;s disease detection</article-title>. <source>Comput. Biol. Med</source>. <volume>170</volume>:<fpage>108029</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108029</pub-id><pub-id pub-id-type="pmid">38308870</pub-id></mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Cai</surname> <given-names>T.</given-names></name> <name><surname>Luo</surname> <given-names>S.</given-names></name> <name><surname>Xu</surname> <given-names>K.</given-names></name> <name><surname>He</surname> <given-names>D.</given-names></name> <name><surname>Liu</surname> <given-names>T.-Y.</given-names></name> <name><surname>Wang</surname> <given-names>L.</given-names></name></person-group> (<year>2021</year>). <article-title>&#x0201C;Graphnorm: a principled approach to accelerating graph neural network training,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>PMLR</publisher-loc>), <fpage>1204</fpage>&#x02013;<lpage>1215</lpage>.</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chicco</surname> <given-names>D.</given-names></name> <name><surname>Jurman</surname> <given-names>G.</given-names></name></person-group> (<year>2023</year>). <article-title>The Matthews correlation coefficient (MCC) should replace the roc AUC as the standard metric for assessing binary classification</article-title>. <source>BioData Min</source>. <volume>16</volume>:<fpage>4</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13040-023-00322-4</pub-id><pub-id pub-id-type="pmid">36800973</pub-id></mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Choi</surname> <given-names>S.-E.</given-names></name> <name><surname>Mukherjee</surname> <given-names>S.</given-names></name> <name><surname>Gibbons</surname> <given-names>L. E.</given-names></name> <name><surname>Sanders</surname> <given-names>R. E.</given-names></name> <name><surname>Jones</surname> <given-names>R. N.</given-names></name> <name><surname>Tommet</surname> <given-names>D.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Development and validation of language and visuospatial composite scores in adni</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>6</volume>:<fpage>e12072</fpage>. <pub-id pub-id-type="pmid">33313380</pub-id></mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Corso</surname> <given-names>G.</given-names></name> <name><surname>Stark</surname> <given-names>H.</given-names></name> <name><surname>Jegelka</surname> <given-names>S.</given-names></name> <name><surname>Jaakkola</surname> <given-names>T.</given-names></name> <name><surname>Barzilay</surname> <given-names>R.</given-names></name></person-group> (<year>2024</year>). <article-title>Graph neural networks</article-title>. <source>Nat. Rev. Methods Prim</source>. <volume>4</volume>:<fpage>17</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s43586-024-00294-7</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Crane</surname> <given-names>P. K.</given-names></name> <name><surname>Carle</surname> <given-names>A.</given-names></name> <name><surname>Gibbons</surname> <given-names>L. E.</given-names></name> <name><surname>Insel</surname> <given-names>P.</given-names></name> <name><surname>Mackin</surname> <given-names>R. S.</given-names></name> <name><surname>Gross</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Development and assessment of a composite score for memory in the Alzheimer&#x00027;s disease neuroimaging initiative (adni)</article-title>. <source>Brain Imaging Behav</source>. <volume>6</volume>, <fpage>502</fpage>&#x02013;<lpage>516</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11682-012-9186-z</pub-id><pub-id pub-id-type="pmid">22782295</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dansson</surname> <given-names>H. V.</given-names></name> <name><surname>Stempfle</surname> <given-names>L.</given-names></name> <name><surname>Egilsd&#x000F3;ttir</surname> <given-names>H.</given-names></name> <name><surname>Schliep</surname> <given-names>A.</given-names></name> <name><surname>Portelius</surname> <given-names>E.</given-names></name> <name><surname>Blennow</surname> <given-names>K.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Predicting progression and cognitive decline in amyloid-positive patients with Alzheimer&#x00027;s disease</article-title>. <source>Alzheimer&#x00027;s Res. Ther</source>. <volume>13</volume>, <fpage>1</fpage>&#x02013;<lpage>16</lpage>. doi: <pub-id pub-id-type="doi">10.1186/s13195-021-00886-5</pub-id><pub-id pub-id-type="pmid">34488882</pub-id></mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dao</surname> <given-names>D.-P.</given-names></name> <name><surname>Yang</surname> <given-names>H.-J.</given-names></name> <name><surname>Kim</surname> <given-names>J.</given-names></name> <name><surname>Ho</surname> <given-names>N.-H.</given-names></name> <collab>for the Alzheimer&#x00027;s Disease Neuroimaging Initiative</collab></person-group> (<year>2025</year>). <article-title>Longitudinal Alzheimer&#x00027;s disease progression prediction with modality uncertainty and optimization of information flow</article-title>. <source>IEEE J. Biomed. Health Inf</source>. <volume>29</volume>, <fpage>259</fpage>&#x02013;<lpage>272</lpage>. doi: <pub-id pub-id-type="doi">10.1109/JBHI.2024.3472462</pub-id><pub-id pub-id-type="pmid">39356605</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Devanarayan</surname> <given-names>V.</given-names></name> <name><surname>Donohue</surname> <given-names>M. C.</given-names></name> <name><surname>Sperling</surname> <given-names>R. A.</given-names></name> <name><surname>Johnson</surname> <given-names>K. A.</given-names></name> <name><surname>Ye</surname> <given-names>Y.</given-names></name> <name><surname>Charil</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Multimodal prognostic modeling of individual cognitive trajectories to enhance trial efficiency in preclinical Alzheimer&#x00027;s disease</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>21</volume>:<fpage>e70702</fpage>. doi: <pub-id pub-id-type="doi">10.1002/alz.70702</pub-id><pub-id pub-id-type="pmid">40990131</pub-id></mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Devanarayan</surname> <given-names>V.</given-names></name> <name><surname>Ye</surname> <given-names>Y.</given-names></name> <name><surname>Charil</surname> <given-names>A.</given-names></name> <name><surname>Andreozzi</surname> <given-names>E.</given-names></name> <name><surname>Sachdev</surname> <given-names>P.</given-names></name> <name><surname>Llano</surname> <given-names>D. A.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Predicting clinical progression trajectories of early Alzheimer&#x00027;s disease patients</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>20</volume>, <fpage>1725</fpage>&#x02013;<lpage>1738</lpage>. doi: <pub-id pub-id-type="doi">10.1002/alz.13565</pub-id><pub-id pub-id-type="pmid">38087949</pub-id></mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Duara</surname> <given-names>R.</given-names></name> <name><surname>Barker</surname> <given-names>W.</given-names></name></person-group> (<year>2022</year>). <article-title>Heterogeneity in Alzheimer&#x00027;s disease diagnosis and progression rates: implications for therapeutic trials</article-title>. <source>Neurotherapeutics</source> <volume>19</volume>, <fpage>8</fpage>&#x02013;<lpage>25</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s13311-022-01185-z</pub-id><pub-id pub-id-type="pmid">35084721</pub-id></mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dukart</surname> <given-names>J.</given-names></name> <name><surname>Mueller</surname> <given-names>K.</given-names></name> <name><surname>Horstmann</surname> <given-names>A.</given-names></name> <name><surname>Vogt</surname> <given-names>B.</given-names></name> <name><surname>Frisch</surname> <given-names>S.</given-names></name> <name><surname>Barthel</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>Differential effects of global and cerebellar normalization on detection and differentiation of dementia in FDG-pet studies</article-title>. <source>Neuroimage</source> <volume>49</volume>, <fpage>1490</fpage>&#x02013;<lpage>1495</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2009.09.017</pub-id><pub-id pub-id-type="pmid">19770055</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Elfwing</surname> <given-names>S.</given-names></name> <name><surname>Uchibe</surname> <given-names>E.</given-names></name> <name><surname>Doya</surname> <given-names>K.</given-names></name></person-group> (<year>2018</year>). <article-title>Sigmoid-weighted linear units for neural network function approximation in reinforcement learning</article-title>. <source>Neural Netw</source>. <volume>107</volume>, <fpage>3</fpage>&#x02013;<lpage>11</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neunet.2017.12.012</pub-id><pub-id pub-id-type="pmid">29395652</pub-id></mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>El-Sappagh</surname> <given-names>S.</given-names></name> <name><surname>Alonso-Moral</surname> <given-names>J. M.</given-names></name> <name><surname>Abuhmed</surname> <given-names>T.</given-names></name> <name><surname>Ali</surname> <given-names>F.</given-names></name> <name><surname>Bugar&#x000ED;n-Diz</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Trustworthy artificial intelligence in Alzheimer&#x00027;s disease: state of the art, opportunities, and challenges</article-title>. <source>Artif. Intell. Rev</source>. <volume>56</volume>, <fpage>11149</fpage>&#x02013;<lpage>11296</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10462-023-10415-5</pub-id></mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Farias</surname> <given-names>S. T.</given-names></name> <name><surname>Mungas</surname> <given-names>D.</given-names></name> <name><surname>Reed</surname> <given-names>B. R.</given-names></name> <name><surname>Harvey</surname> <given-names>D.</given-names></name> <name><surname>DeCarli</surname> <given-names>C.</given-names></name></person-group> (<year>2009</year>). <article-title>Progression of mild cognitive impairment to dementia in clinic-vs community-based cohorts</article-title>. <source>Arch. Neurol</source>. <volume>66</volume>, <fpage>1151</fpage>&#x02013;<lpage>1157</lpage>. doi: <pub-id pub-id-type="doi">10.1001/archneurol.2009.106</pub-id></mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Franzmeier</surname> <given-names>N.</given-names></name> <name><surname>Koutsouleris</surname> <given-names>N.</given-names></name> <name><surname>Benzinger</surname> <given-names>T.</given-names></name> <name><surname>Goate</surname> <given-names>A.</given-names></name> <name><surname>Karch</surname> <given-names>C. M.</given-names></name> <name><surname>Fagan</surname> <given-names>A. M.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Predicting sporadic Alzheimer&#x00027;s disease progression via inherited Alzheimer&#x00027;s disease-informed machine-learning</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>16</volume>, <fpage>501</fpage>&#x02013;<lpage>511</lpage>. doi: <pub-id pub-id-type="doi">10.1002/alz.12032</pub-id><pub-id pub-id-type="pmid">32043733</pub-id></mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Friedman</surname> <given-names>J.</given-names></name> <name><surname>Hastie</surname> <given-names>T.</given-names></name> <name><surname>Tibshirani</surname> <given-names>R.</given-names></name></person-group> (<year>2008</year>). <article-title>Sparse inverse covariance estimation with the graphical lasso</article-title>. <source>Biostatistics</source> <volume>9</volume>, <fpage>432</fpage>&#x02013;<lpage>441</lpage>. doi: <pub-id pub-id-type="doi">10.1093/biostatistics/kxm045</pub-id><pub-id pub-id-type="pmid">18079126</pub-id></mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a-Guti&#x000E9;rrez</surname> <given-names>F.</given-names></name> <name><surname>Alegret</surname> <given-names>M.</given-names></name> <name><surname>Marqui&#x000E9;</surname> <given-names>M.</given-names></name> <name><surname>Mu&#x000F1;oz</surname> <given-names>N.</given-names></name> <name><surname>Ortega</surname> <given-names>G.</given-names></name> <name><surname>Cano</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2024a</year>). <article-title>Unveiling the sound of the cognitive status: machine learning-based speech analysis in the Alzheimer&#x00027;s disease spectrum</article-title>. <source>Alzheimer&#x00027;s Res. Ther</source>. <volume>16</volume>:<fpage>26</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13195-024-01394-y</pub-id><pub-id pub-id-type="pmid">38308366</pub-id></mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a-Guti&#x000E9;rrez</surname> <given-names>F.</given-names></name> <name><surname>Delgado-Alvarez</surname> <given-names>A.</given-names></name> <name><surname>Delgado-Alonso</surname> <given-names>C.</given-names></name> <name><surname>D&#x000ED;az-&#x000C1;lvarez</surname> <given-names>J.</given-names></name> <name><surname>Pytel</surname> <given-names>V.</given-names></name> <name><surname>Valles-Salgado</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Diagnosis of Alzheimer&#x00027;s disease and behavioural variant frontotemporal dementia with machine learning-aided neuropsychological assessment using feature engineering and genetic algorithms</article-title>. <source>Int. J. Geriatr. Psychiatry</source> <volume>37</volume>:<fpage>e5667</fpage>. doi: <pub-id pub-id-type="doi">10.1002/gps.5667</pub-id><pub-id pub-id-type="pmid">34894410</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Garc&#x000ED;a-Guti&#x000E9;rrez</surname> <given-names>F.</given-names></name> <name><surname>Hern&#x000E1;ndez-Lorenzo</surname> <given-names>L.</given-names></name> <name><surname>Cabrera-Mart&#x000ED;n</surname> <given-names>M. N.</given-names></name> <name><surname>Matias-Guiu</surname> <given-names>J. A.</given-names></name> <name><surname>Ayala</surname> <given-names>J. L.</given-names></name> <name><surname>Initiative</surname> <given-names>A. D. N.</given-names></name> <etal/></person-group>. (<year>2024b</year>). <article-title>Predicting changes in brain metabolism and progression from mild cognitive impairment to dementia using multitask deep learning models and explainable AI</article-title>. <source>Neuroimage</source> <volume>297</volume>:<fpage>120695</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2024.120695</pub-id><pub-id pub-id-type="pmid">38942101</pub-id></mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gibbons</surname> <given-names>L. E.</given-names></name> <name><surname>Carle</surname> <given-names>A. C.</given-names></name> <name><surname>Mackin</surname> <given-names>R. S.</given-names></name> <name><surname>Harvey</surname> <given-names>D.</given-names></name> <name><surname>Mukherjee</surname> <given-names>S.</given-names></name> <name><surname>Insel</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>A composite score for executive functioning, validated in Alzheimer&#x00027;s disease neuroimaging initiative (ADNI) participants with baseline mild cognitive impairment</article-title>. <source>Brain Imaging Behav</source>. <volume>6</volume>, <fpage>517</fpage>&#x02013;<lpage>527</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11682-012-9176-1</pub-id><pub-id pub-id-type="pmid">22644789</pub-id></mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Goodglass</surname> <given-names>H.</given-names></name> <name><surname>Kaplan</surname> <given-names>E.</given-names></name> <name><surname>Barresi</surname> <given-names>B.</given-names></name></person-group> (<year>2001</year>). <source>The Assessment of Aphasia and Related Disorders</source>. <publisher-loc>Philadelphia</publisher-loc>: <publisher-name>Lippincott Williams &#x00026;Wilkins</publisher-name>.</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gravina</surname> <given-names>M.</given-names></name> <name><surname>Garc&#x000ED;a-Pedrero</surname> <given-names>A.</given-names></name> <name><surname>Gonzalo-Mart&#x000ED;n</surname> <given-names>C.</given-names></name> <name><surname>Sansone</surname> <given-names>C.</given-names></name> <name><surname>Soda</surname> <given-names>P.</given-names></name></person-group> (<year>2024</year>). <article-title>Multi input-multi output 3D CNN for dementia severity assessment with incomplete multimodal data</article-title>. <source>Artif. Intell. Med</source>. <volume>149</volume>:<fpage>102774</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.artmed.2024.102774</pub-id><pub-id pub-id-type="pmid">38462278</pub-id></mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Hair Jr</surname> <given-names>J. F.</given-names></name> <name><surname>Hult</surname> <given-names>G. T. M.</given-names></name> <name><surname>Ringle</surname> <given-names>C. M.</given-names></name> <name><surname>Sarstedt</surname> <given-names>M.</given-names></name> <name><surname>Danks</surname> <given-names>N. P.</given-names></name> <name><surname>Ray</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>&#x0201C;An introduction to structural equation modeling,&#x0201D;</article-title> in <source>Partial Least Squares Structural Equation Modeling (PLS-SEM) Using R: a Workbook</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>29</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-3-030-80519-7_1</pub-id><pub-id pub-id-type="pmid">26152937</pub-id></mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hampel</surname> <given-names>H.</given-names></name> <name><surname>Au</surname> <given-names>R.</given-names></name> <name><surname>Mattke</surname> <given-names>S.</given-names></name> <name><surname>van der Flier</surname> <given-names>W. M.</given-names></name> <name><surname>Aisen</surname> <given-names>P.</given-names></name> <name><surname>Apostolova</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Designing the next-generation clinical care pathway for Alzheimer&#x00027;s disease</article-title>. <source>Nature Aging</source> <volume>2</volume>, <fpage>692</fpage>&#x02013;<lpage>703</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s43587-022-00269-x</pub-id><pub-id pub-id-type="pmid">37118137</pub-id></mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hersi</surname> <given-names>M.</given-names></name> <name><surname>Irvine</surname> <given-names>B.</given-names></name> <name><surname>Gupta</surname> <given-names>P.</given-names></name> <name><surname>Gomes</surname> <given-names>J.</given-names></name> <name><surname>Birkett</surname> <given-names>N.</given-names></name> <name><surname>Krewski</surname> <given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>Risk factors associated with the onset and progression of Alzheimer&#x00027;s disease: a systematic review of the evidence</article-title>. <source>Neurotoxicology</source> <volume>61</volume>, <fpage>143</fpage>&#x02013;<lpage>187</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuro.2017.03.006</pub-id><pub-id pub-id-type="pmid">28363508</pub-id></mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hoopes</surname> <given-names>A.</given-names></name> <name><surname>Mora</surname> <given-names>J. S.</given-names></name> <name><surname>Dalca</surname> <given-names>A. V.</given-names></name> <name><surname>Fischl</surname> <given-names>B.</given-names></name> <name><surname>Hoffmann</surname> <given-names>M.</given-names></name></person-group> (<year>2022</year>). <article-title>Synthstrip: skull-stripping for any brain image</article-title>. <source>Neuroimage</source> <volume>260</volume>:<fpage>119474</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2022.119474</pub-id><pub-id pub-id-type="pmid">35842095</pub-id></mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>G.</given-names></name> <name><surname>Liu</surname> <given-names>Z.</given-names></name> <name><surname>Van Der Maaten</surname> <given-names>L.</given-names></name> <name><surname>Weinberger</surname> <given-names>K. Q.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Densely connected convolutional networks,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>, 4700&#x02013;4708. doi: <pub-id pub-id-type="doi">10.1109/CVPR.2017.243</pub-id></mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>L.</given-names></name> <name><surname>Jin</surname> <given-names>Y.</given-names></name> <name><surname>Gao</surname> <given-names>Y.</given-names></name> <name><surname>Thung</surname> <given-names>K.-H.</given-names></name> <name><surname>Shen</surname> <given-names>D.</given-names></name> Alzheimer&#x00027;s Disease Neuroimaging Initiative</person-group>. (<year>2016</year>). <article-title>Longitudinal clinical score prediction in Alzheimer&#x00027;s disease with soft-split sparse regression based random forest</article-title>. <source>Neurobiol. Aging</source> <volume>46</volume>, <fpage>180</fpage>&#x02013;<lpage>191</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neurobiolaging.2016.07.005</pub-id><pub-id pub-id-type="pmid">27500865</pub-id></mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>S.</given-names></name> <name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Sun</surname> <given-names>L.</given-names></name> <name><surname>Ye</surname> <given-names>J.</given-names></name> <name><surname>Fleisher</surname> <given-names>A.</given-names></name> <name><surname>Wu</surname> <given-names>T.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>Learning brain connectivity of Alzheimer&#x00027;s disease by sparse inverse covariance estimation</article-title>. <source>Neuroimage</source> <volume>50</volume>, <fpage>935</fpage>&#x02013;<lpage>949</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2009.12.120</pub-id><pub-id pub-id-type="pmid">20079441</pub-id></mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jagust</surname> <given-names>W. J.</given-names></name> <name><surname>Landau</surname> <given-names>S. M.</given-names></name> <name><surname>Koeppe</surname> <given-names>R. A.</given-names></name> <name><surname>Reiman</surname> <given-names>E. M.</given-names></name> <name><surname>Chen</surname> <given-names>K.</given-names></name> <name><surname>Mathis</surname> <given-names>C. A.</given-names></name> <etal/></person-group>. (<year>2015</year>). <article-title>The Alzheimer&#x00027;s disease neuroimaging initiative 2 pet core: 2015</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>11</volume>, <fpage>757</fpage>&#x02013;<lpage>771</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jalz.2015.05.001</pub-id><pub-id pub-id-type="pmid">26194311</pub-id></mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Johnson</surname> <given-names>K. A.</given-names></name> <name><surname>Sperling</surname> <given-names>R. A.</given-names></name> <name><surname>Gidicsin</surname> <given-names>C. M.</given-names></name> <name><surname>Carmasin</surname> <given-names>J. S.</given-names></name> <name><surname>Maye</surname> <given-names>J. E.</given-names></name> <name><surname>Coleman</surname> <given-names>R. E.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Florbetapir (f18-av-45) pet to assess amyloid burden in Alzheimer&#x00027;s disease dementia, mild cognitive impairment, and normal aging</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>9</volume>, <fpage>S72</fpage>&#x02013;<lpage>83</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jalz.2012.10.007</pub-id><pub-id pub-id-type="pmid">23375563</pub-id></mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jutten</surname> <given-names>R. J.</given-names></name> <name><surname>Sikkes</surname> <given-names>S. A.</given-names></name> <name><surname>Van der Flier</surname> <given-names>W. M.</given-names></name> <name><surname>Scheltens</surname> <given-names>P.</given-names></name> <name><surname>Visser</surname> <given-names>P. J.</given-names></name> <name><surname>Tijms</surname> <given-names>B. M.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Finding treatment effects in Alzheimer trials in the face of disease progression heterogeneity</article-title>. <source>Neurology</source> <volume>96</volume>, <fpage>e2673</fpage>&#x02013;<lpage>e2684</lpage>. doi: <pub-id pub-id-type="doi">10.1212/WNL.0000000000012022</pub-id><pub-id pub-id-type="pmid">34550903</pub-id></mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kaplan</surname> <given-names>E.</given-names></name> <name><surname>Goodglass</surname> <given-names>H.</given-names></name> <name><surname>Weintraub</surname> <given-names>S.</given-names></name></person-group> (<year>2001</year>). <article-title>&#x0201C;Boston naming test,&#x0201D;</article-title> in <source>The Clinical Neuropsychologist</source>.</mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kaur</surname> <given-names>I.</given-names></name> <name><surname>Sachdeva</surname> <given-names>R.</given-names></name></person-group> (<year>2025</year>). <article-title>Prediction models for early detection of Alzheimer: recent trends and future prospects</article-title>. <source>Arch. Comput. Methods Eng</source>. <volume>32</volume>, <fpage>3565</fpage>&#x02013;<lpage>3592</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11831-025-10246-3</pub-id></mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kingma</surname> <given-names>D. P.</given-names></name> <name><surname>Ba</surname> <given-names>J.</given-names></name></person-group> (<year>2014</year>). <article-title>Adam: a method for stochastic optimization</article-title>. <source>arXiv preprint arXiv:1412.6980</source>.</mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kipf</surname> <given-names>T. N.</given-names></name> <name><surname>Welling</surname> <given-names>M.</given-names></name></person-group> (<year>2016</year>). <article-title>Semi-supervised classification with graph convolutional networks</article-title>. <source>arXiv preprint arXiv:1609.02907</source>.</mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kueper</surname> <given-names>J. K.</given-names></name> <name><surname>Speechley</surname> <given-names>M.</given-names></name> <name><surname>Montero-Odasso</surname> <given-names>M.</given-names></name></person-group> (<year>2018</year>). <article-title>The Alzheimer&#x00027;s disease assessment scale-cognitive subscale (ADAS-cog): modifications and responsiveness in pre-dementia populations. A narrative review</article-title>. <source>J. Alzheimer&#x00027;s Dis</source>. <volume>63</volume>, <fpage>423</fpage>&#x02013;<lpage>444</lpage>. doi: <pub-id pub-id-type="doi">10.3233/JAD-170991</pub-id><pub-id pub-id-type="pmid">29660938</pub-id></mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lam</surname> <given-names>B.</given-names></name> <name><surname>Masellis</surname> <given-names>M.</given-names></name> <name><surname>Freedman</surname> <given-names>M.</given-names></name> <name><surname>Stuss</surname> <given-names>D. T.</given-names></name> <name><surname>Black</surname> <given-names>S. E.</given-names></name></person-group> (<year>2013</year>). <article-title>Clinical, imaging, and pathological heterogeneity of the Alzheimer&#x00027;s disease syndrome</article-title>. <source>Alzheimer&#x00027;s Res. Ther</source>. <volume>5</volume>, <fpage>1</fpage>&#x02013;<lpage>14</lpage>. doi: <pub-id pub-id-type="doi">10.1186/alzrt155</pub-id><pub-id pub-id-type="pmid">23302773</pub-id></mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lee</surname> <given-names>M.-W.</given-names></name> <name><surname>Kim</surname> <given-names>H. W.</given-names></name> <name><surname>Choe</surname> <given-names>Y. S.</given-names></name> <name><surname>Yang</surname> <given-names>H. S.</given-names></name> <name><surname>Lee</surname> <given-names>J.</given-names></name> <name><surname>Lee</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>A multimodal machine learning model for predicting dementia conversion in Alzheimer&#x00027;s disease</article-title>. <source>Sci. Rep</source>. <volume>14</volume>:<fpage>12276</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-60134-2</pub-id><pub-id pub-id-type="pmid">38806509</pub-id></mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lei</surname> <given-names>B.</given-names></name> <name><surname>Liang</surname> <given-names>E.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Yang</surname> <given-names>P.</given-names></name> <name><surname>Zhou</surname> <given-names>F.</given-names></name> <name><surname>Tan</surname> <given-names>E.-L.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Predicting clinical scores for Alzheimer&#x00027;s disease based on joint and deep learning</article-title>. <source>Expert Syst. Appl</source>. <volume>187</volume>:<fpage>115966</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.eswa.2021.115966</pub-id></mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lei</surname> <given-names>B.</given-names></name> <name><surname>Yang</surname> <given-names>M.</given-names></name> <name><surname>Yang</surname> <given-names>P.</given-names></name> <name><surname>Zhou</surname> <given-names>F.</given-names></name> <name><surname>Hou</surname> <given-names>W.</given-names></name> <name><surname>Zou</surname> <given-names>W.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Deep and joint learning of longitudinal data for Alzheimer&#x00027;s disease prediction</article-title>. <source>Pattern Recognit</source>. <volume>102</volume>:<fpage>107247</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.patcog.2020.107247</pub-id></mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X.</given-names></name> <name><surname>Zhou</surname> <given-names>Y.</given-names></name> <name><surname>Dvornek</surname> <given-names>N.</given-names></name> <name><surname>Zhang</surname> <given-names>M.</given-names></name> <name><surname>Gao</surname> <given-names>S.</given-names></name> <name><surname>Zhuang</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Braingnn: interpretable brain graph neural network for fMRI analysis</article-title>. <source>Med. Image Anal</source>. <volume>74</volume>:<fpage>102233</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2021.102233</pub-id><pub-id pub-id-type="pmid">34655865</pub-id></mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Maheux</surname> <given-names>E.</given-names></name> <name><surname>Koval</surname> <given-names>I.</given-names></name> <name><surname>Ortholand</surname> <given-names>J.</given-names></name> <name><surname>Birkenbihl</surname> <given-names>C.</given-names></name> <name><surname>Archetti</surname> <given-names>D.</given-names></name> <name><surname>Bouteloup</surname> <given-names>V.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Forecasting individual progression trajectories in Alzheimer&#x00027;s disease</article-title>. <source>Nat. Commun</source>. <volume>14</volume>:<fpage>761</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41467-022-35712-5</pub-id><pub-id pub-id-type="pmid">36765056</pub-id></mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Marinescu</surname> <given-names>R. V.</given-names></name> <name><surname>Oxtoby</surname> <given-names>N. P.</given-names></name> <name><surname>Young</surname> <given-names>A. L.</given-names></name> <name><surname>Bron</surname> <given-names>E. E.</given-names></name> <name><surname>Toga</surname> <given-names>A. W.</given-names></name> <name><surname>Weiner</surname> <given-names>M. W.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>&#x0201C;Tadpole challenge: Accurate Alzheimer&#x00027;s disease prediction through crowdsourced forecasting of future data,&#x0201D;</article-title> in <source>Predictive Intelligence in Medicine: Second International Workshop, PRIME 2019, Held in Conjunction with MICCAI 2019, Shenzhen, China, October 13, 2019, Proceedings 2</source> (<publisher-loc>Springer</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-3-030-32281-6_1</pub-id><pub-id pub-id-type="pmid">32587957</pub-id></mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mohs</surname> <given-names>R. C.</given-names></name> <name><surname>Knopman</surname> <given-names>D.</given-names></name> <name><surname>Petersen</surname> <given-names>R. C.</given-names></name> <name><surname>Ferris</surname> <given-names>S. H.</given-names></name> <name><surname>Ernesto</surname> <given-names>C.</given-names></name> <name><surname>Grundman</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>1997</year>). <article-title>Development of cognitive instruments for use in clinical trials of antidementia drugs: additions to the Alzheimer&#x00027;s disease assessment scale that broaden its scope</article-title>. <source>Alzheimer Dis. Assoc. Disor</source>. <volume>11</volume>, <fpage>13</fpage>&#x02013;<lpage>21</lpage>. doi: <pub-id pub-id-type="doi">10.1097/00002093-199700112-00003</pub-id></mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moradi</surname> <given-names>E.</given-names></name> <name><surname>Dahnke</surname> <given-names>R.</given-names></name> <name><surname>Imani</surname> <given-names>V.</given-names></name> <name><surname>Gaser</surname> <given-names>C.</given-names></name> <name><surname>Solomon</surname> <given-names>A.</given-names></name> <name><surname>Tohka</surname> <given-names>J.</given-names></name></person-group> (<year>2025</year>). <article-title>Integrating plasma, mri, and cognitive biomarkers for personalized prediction of decline across cognitive domains</article-title>. <source>Neurobiol. Aging</source> <volume>155</volume>, <fpage>53</fpage>&#x02013;<lpage>65</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neurobiolaging.2025.06.010</pub-id><pub-id pub-id-type="pmid">40712372</pub-id></mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Morris</surname> <given-names>J. C.</given-names></name> <name><surname>Heyman</surname> <given-names>A.</given-names></name> <name><surname>Mohs</surname> <given-names>R. C.</given-names></name> <name><surname>Hughes</surname> <given-names>J. P.</given-names></name> <name><surname>van Belle</surname> <given-names>G.</given-names></name> <name><surname>Fillenbaum</surname> <given-names>G.</given-names></name> <etal/></person-group>. (<year>1989</year>). <article-title>The consortium to establish a registry for Alzheimer&#x00027;s disease (cerad). Part i. clinical and neuropsychological assessment of Alzheimer&#x00027;s disease</article-title>. <source>Neurology</source> <volume>39</volume>, <fpage>1159</fpage>&#x02013;<lpage>1165</lpage>. doi: <pub-id pub-id-type="doi">10.1212/WNL.39.9.1159</pub-id><pub-id pub-id-type="pmid">2771064</pub-id></mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nguyen</surname> <given-names>M.</given-names></name> <name><surname>He</surname> <given-names>T.</given-names></name> <name><surname>An</surname> <given-names>L.</given-names></name> <name><surname>Alexander</surname> <given-names>D. C.</given-names></name> <name><surname>Feng</surname> <given-names>J.</given-names></name> <name><surname>Yeo</surname> <given-names>B. T.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Predicting Alzheimer&#x00027;s disease progression using deep recurrent neural networks</article-title>. <source>Neuroimage</source> <volume>222</volume>:<fpage>117203</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2020.117203</pub-id><pub-id pub-id-type="pmid">32763427</pub-id></mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nordberg</surname> <given-names>A.</given-names></name> <name><surname>Rinne</surname> <given-names>J. O.</given-names></name> <name><surname>Kadir</surname> <given-names>A.</given-names></name> <name><surname>L&#x000E5;ngstr&#x000F6;m</surname> <given-names>B.</given-names></name></person-group> (<year>2010</year>). <article-title>The use of pet in Alzheimer disease</article-title>. <source>Nat. Rev. Neurol</source>. <volume>6</volume>, <fpage>78</fpage>&#x02013;<lpage>87</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nrneurol.2009.217</pub-id></mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>L. Q.</given-names></name> <name><surname>Gross</surname> <given-names>A. L.</given-names></name> <name><surname>McLaren</surname> <given-names>D. G.</given-names></name> <name><surname>Pa</surname> <given-names>J.</given-names></name> <name><surname>Johnson</surname> <given-names>J. K.</given-names></name> <name><surname>Mitchell</surname> <given-names>M.</given-names></name> <etal/></person-group>. (<year>2012</year>). <article-title>Confirmatory factor analysis of the adni neuropsychological battery</article-title>. <source>Brain Imaging Behav</source>. <volume>6</volume>, <fpage>528</fpage>&#x02013;<lpage>539</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11682-012-9190-3</pub-id><pub-id pub-id-type="pmid">22777078</pub-id></mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pe na-Casanova</surname> <given-names>J.</given-names></name> <name><surname>S&#x000E1;nchez-Benavides</surname> <given-names>G.</given-names></name> <name><surname>de Sola</surname> <given-names>S.</given-names></name> <name><surname>Manero-Borr&#x000E1;s</surname> <given-names>R. M.</given-names></name> <name><surname>Casals-Coll</surname> <given-names>M.</given-names></name></person-group> (<year>2012</year>). <article-title>Neuropsychology of Alzheimer&#x00027;s disease</article-title>. <source>Arch. Med. Res</source>. <volume>43</volume>, <fpage>686</fpage>&#x02013;<lpage>693</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.arcmed.2012.08.015</pub-id></mixed-citation>
</ref>
<ref id="B60">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pedregosa</surname> <given-names>F.</given-names></name> <name><surname>Varoquaux</surname> <given-names>G.</given-names></name> <name><surname>Gramfort</surname> <given-names>A.</given-names></name> <name><surname>Michel</surname> <given-names>V.</given-names></name> <name><surname>Thirion</surname> <given-names>B.</given-names></name> <name><surname>Grisel</surname> <given-names>O.</given-names></name> <etal/></person-group>. (<year>2011</year>). <article-title>Scikit-learn: machine learning in python</article-title>. <source>J. Mach. Learn. Res</source>. <volume>12</volume>, <fpage>2825</fpage>&#x02013;<lpage>2830</lpage>.</mixed-citation>
</ref>
<ref id="B61">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peng</surname> <given-names>H.</given-names></name> <name><surname>Gong</surname> <given-names>W.</given-names></name> <name><surname>Beckmann</surname> <given-names>C. F.</given-names></name> <name><surname>Vedaldi</surname> <given-names>A.</given-names></name> <name><surname>Smith</surname> <given-names>S. M.</given-names></name></person-group> (<year>2021</year>). <article-title>Accurate brain age prediction with lightweight deep neural networks</article-title>. <source>Med. Image Anal</source>. <volume>68</volume>:<fpage>101871</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2020.101871</pub-id><pub-id pub-id-type="pmid">33197716</pub-id></mixed-citation>
</ref>
<ref id="B62">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Petersen</surname> <given-names>R. C.</given-names></name></person-group> (<year>2004</year>). <article-title>Mild cognitive impairment as a diagnostic entity</article-title>. <source>J. Intern. Med</source>. <volume>256</volume>, <fpage>183</fpage>&#x02013;<lpage>194</lpage>. doi: <pub-id pub-id-type="doi">10.1111/j.1365-2796.2004.01388.x</pub-id><pub-id pub-id-type="pmid">15324362</pub-id></mixed-citation>
</ref>
<ref id="B63">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Reitan</surname> <given-names>R. M.</given-names></name> <name><surname>Wolfson</surname> <given-names>D.</given-names></name></person-group> (<year>2014</year>). <article-title>&#x0201C;The halstead-reitan neuropsychological test battery and aging,&#x0201D;</article-title> in <source>Clinical Gerontology</source>, 39&#x02013;61. doi: <pub-id pub-id-type="doi">10.1300/J018v05n01_03</pub-id></mixed-citation>
</ref>
<ref id="B64">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Rey</surname> <given-names>A.</given-names></name></person-group> (<year>1958</year>). <source>L&#x00027;examen clinique en psychologie</source>. <publisher-loc>Paris</publisher-loc>: <publisher-name>Presses Universitaires de France</publisher-name>.</mixed-citation>
</ref>
<ref id="B65">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rieke</surname> <given-names>N.</given-names></name> <name><surname>Hancox</surname> <given-names>J.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Milletari</surname> <given-names>F.</given-names></name> <name><surname>Roth</surname> <given-names>H. R.</given-names></name> <name><surname>Albarqouni</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>The future of digital health with federated learning</article-title>. <source>NPJ Dig. Med</source>. <volume>3</volume>:<fpage>119</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41746-020-00323-1</pub-id><pub-id pub-id-type="pmid">33015372</pub-id></mixed-citation>
</ref>
<ref id="B66">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rolls</surname> <given-names>E. T.</given-names></name> <name><surname>Huang</surname> <given-names>C.-C.</given-names></name> <name><surname>Lin</surname> <given-names>C.-P.</given-names></name> <name><surname>Feng</surname> <given-names>J.</given-names></name> <name><surname>Joliot</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Automated anatomical labelling atlas 3</article-title>. <source>Neuroimage</source> <volume>206</volume>:<fpage>116189</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2019.116189</pub-id><pub-id pub-id-type="pmid">31521825</pub-id></mixed-citation>
</ref>
<ref id="B67">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rosseel</surname> <given-names>Y.</given-names></name></person-group> (<year>2012</year>). <article-title>LAVAAN: an r package for structural equation modeling</article-title>. <source>J. Stat. Softw</source>. <volume>48</volume>, <fpage>1</fpage>&#x02013;<lpage>36</lpage>. doi: <pub-id pub-id-type="doi">10.18637/jss.v048.i02</pub-id></mixed-citation>
</ref>
<ref id="B68">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Samper-Gonz&#x000E1;lez</surname> <given-names>J.</given-names></name> <name><surname>Burgos</surname> <given-names>N.</given-names></name> <name><surname>Fontanella</surname> <given-names>S.</given-names></name> <name><surname>Bertin</surname> <given-names>H.</given-names></name> <name><surname>Habert</surname> <given-names>M.-O.</given-names></name> <name><surname>Durrleman</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>&#x0201C;Yet another adni machine learning paper? Paving the way towards fully-reproducible research on classification of Alzheimer&#x00027;s disease,&#x0201D;</article-title> in <source>International Workshop on Machine Learning in Medical Imaging</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>), <fpage>53</fpage>&#x02013;<lpage>60</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-3-319-67389-9_7</pub-id></mixed-citation>
</ref>
<ref id="B69">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Seo</surname> <given-names>H.</given-names></name> <name><surname>Brand</surname> <given-names>L.</given-names></name> <name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Initiative</surname> <given-names>A. D. N.</given-names></name></person-group> (<year>2024</year>). <article-title>Learning semi-supervised enrichment of longitudinal imaging-genetic data for improved prediction of cognitive decline</article-title>. <source>BMC Med. Inform. Decis. Mak</source>. <volume>24</volume>:<fpage>61</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12911-024-02455-w</pub-id><pub-id pub-id-type="pmid">38807132</pub-id></mixed-citation>
</ref>
<ref id="B70">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sharma</surname> <given-names>R.</given-names></name> <name><surname>Goel</surname> <given-names>T.</given-names></name> <name><surname>Tanveer</surname> <given-names>M.</given-names></name> <name><surname>Lin</surname> <given-names>C.</given-names></name> <name><surname>Murugan</surname> <given-names>R.</given-names></name></person-group> (<year>2023</year>). <article-title>Deep learning based diagnosis and prognosis of Alzheimer&#x00027;s disease: a comprehensive review</article-title>. <source>IEEE Trans. Cogn. Dev. Syst</source>. <volume>15</volume>, <fpage>1123</fpage>&#x02013;<lpage>1138</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TCDS.2023.3254209</pub-id></mixed-citation>
</ref>
<ref id="B71">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shekari</surname> <given-names>M.</given-names></name> <name><surname>V&#x000E1;llez Garc&#x000ED;a</surname> <given-names>D.</given-names></name> <name><surname>Collij</surname> <given-names>L. E.</given-names></name> <name><surname>Altomare</surname> <given-names>D.</given-names></name> <name><surname>Heeman</surname> <given-names>F.</given-names></name> <name><surname>Pemberton</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Stress testing the centiloid: precision and variability of pet quantification of amyloid pathology</article-title>. <source>Alzheimer&#x00027;s Dement</source>. <volume>20</volume>, <fpage>5102</fpage>&#x02013;<lpage>5113</lpage>. doi: <pub-id pub-id-type="doi">10.1002/alz.13883</pub-id><pub-id pub-id-type="pmid">38961808</pub-id></mixed-citation>
</ref>
<ref id="B72">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Spasov</surname> <given-names>S.</given-names></name> <name><surname>Passamonti</surname> <given-names>L.</given-names></name> <name><surname>Duggento</surname> <given-names>A.</given-names></name> <name><surname>Lio</surname> <given-names>P.</given-names></name> <name><surname>Toschi</surname> <given-names>N.</given-names></name> <name><surname>Initiative</surname> <given-names>A. D. N.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>A parameter-efficient deep learning approach to predict conversion from mild cognitive impairment to Alzheimer&#x00027;s disease</article-title>. <source>Neuroimage</source> <volume>189</volume>, <fpage>276</fpage>&#x02013;<lpage>287</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2019.01.031</pub-id><pub-id pub-id-type="pmid">30654174</pub-id></mixed-citation>
</ref>
<ref id="B73">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tustison</surname> <given-names>N. J.</given-names></name> <name><surname>Avants</surname> <given-names>B. B.</given-names></name> <name><surname>Cook</surname> <given-names>P. A.</given-names></name> <name><surname>Zheng</surname> <given-names>Y.</given-names></name> <name><surname>Egan</surname> <given-names>A.</given-names></name> <name><surname>Yushkevich</surname> <given-names>P. A.</given-names></name> <etal/></person-group>. (<year>2010</year>). <article-title>N4itk: improved n3 bias correction</article-title>. <source>IEEE Trans. Med. Imaging</source> <volume>29</volume>, <fpage>1310</fpage>&#x02013;<lpage>1320</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TMI.2010.2046908</pub-id><pub-id pub-id-type="pmid">20378467</pub-id></mixed-citation>
</ref>
<ref id="B74">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A.</given-names></name> <name><surname>Shazeer</surname> <given-names>N.</given-names></name> <name><surname>Parmar</surname> <given-names>N.</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J.</given-names></name> <name><surname>Jones</surname> <given-names>L.</given-names></name> <name><surname>Gomez</surname> <given-names>A. N.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>&#x0201C;Attention is all you need,&#x0201D;</article-title> in <source>Advances in Neural Information Processing Systems</source> (<publisher-loc>Curran Associates, Inc.</publisher-loc>).</mixed-citation>
</ref>
<ref id="B75">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Veli&#x0010D;kovi&#x00107;</surname> <given-names>P.</given-names></name> <name><surname>Cucurull</surname> <given-names>G.</given-names></name> <name><surname>Casanova</surname> <given-names>A.</given-names></name> <name><surname>Romero</surname> <given-names>A.</given-names></name> <name><surname>Lio</surname> <given-names>P.</given-names></name> <name><surname>Bengio</surname> <given-names>Y.</given-names></name></person-group> (<year>2017</year>). <article-title>Graph attention networks</article-title>. <source>arXiv preprint arXiv:1710.10903</source>.</mixed-citation>
</ref>
<ref id="B76">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Gao</surname> <given-names>R.</given-names></name> <name><surname>Wei</surname> <given-names>T.</given-names></name> <name><surname>Johnston</surname> <given-names>L.</given-names></name> <name><surname>Yuan</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Predicting long-term progression of Alzheimer&#x00027;s disease using a multimodal deep learning model incorporating interaction effects</article-title>. <source>J. Transl. Med</source>. <volume>22</volume>:<fpage>265</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12967-024-05025-w</pub-id><pub-id pub-id-type="pmid">38468358</pub-id></mixed-citation>
</ref>
<ref id="B77">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>White</surname> <given-names>I. R.</given-names></name> <name><surname>Royston</surname> <given-names>P.</given-names></name> <name><surname>Wood</surname> <given-names>A. M.</given-names></name></person-group> (<year>2011</year>). <article-title>Multiple imputation using chained equations: issues and guidance for practice</article-title>. <source>Stat. Med</source>. <volume>30</volume>, <fpage>377</fpage>&#x02013;<lpage>399</lpage>. doi: <pub-id pub-id-type="doi">10.1002/sim.4067</pub-id><pub-id pub-id-type="pmid">21225900</pub-id></mixed-citation>
</ref>
<ref id="B78">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>K.</given-names></name> <name><surname>Hu</surname> <given-names>W.</given-names></name> <name><surname>Leskovec</surname> <given-names>J.</given-names></name> <name><surname>Jegelka</surname> <given-names>S.</given-names></name></person-group> (<year>2018a</year>). How powerful are graph neural networks? <italic>arXiv preprint arXiv:1810.00826</italic>.</mixed-citation>
</ref>
<ref id="B79">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>K.</given-names></name> <name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Tian</surname> <given-names>Y.</given-names></name> <name><surname>Sonobe</surname> <given-names>T.</given-names></name> <name><surname>Kawarabayashi</surname> <given-names>K.-I.</given-names></name> <name><surname>Jegelka</surname> <given-names>S.</given-names></name></person-group> (<year>2018b</year>). <article-title>&#x0201C;Representation learning on graphs with jumping knowledge networks,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>PMLR</publisher-loc>), <fpage>5453</fpage>&#x02013;<lpage>5462</lpage>.</mixed-citation>
</ref>
<ref id="B80">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>S.</given-names></name> <name><surname>Zhang</surname> <given-names>X.</given-names></name> <name><surname>Du</surname> <given-names>X.</given-names></name> <name><surname>Yan</surname> <given-names>P.</given-names></name> <name><surname>Zhang</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>W.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Prediction of cognitive conversion within the Alzheimer&#x00027;s disease continuum using deep learning</article-title>. <source>Alzheimer&#x00027;s Res. Ther</source>. <volume>17</volume>:<fpage>41</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13195-025-01686-x</pub-id><pub-id pub-id-type="pmid">39948600</pub-id></mixed-citation>
</ref>
<ref id="B81">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>Q.</given-names></name> <name><surname>Ma</surname> <given-names>Q.</given-names></name> <name><surname>Da</surname> <given-names>L.</given-names></name> <name><surname>Li</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>A transformer-based unified multimodal framework for Alzheimer&#x00027;s disease assessment</article-title>. <source>Comput. Biol. Med</source>. <volume>180</volume>:<fpage>108979</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.108979</pub-id><pub-id pub-id-type="pmid">39098237</pub-id></mixed-citation>
</ref>
<ref id="B82">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Shen</surname> <given-names>D.</given-names></name> Alzheimer&#x00027;s Disease Neuroimaging Initiative</person-group> (<year>2012</year>). <article-title>Multi-modal multi-task learning for joint prediction of multiple regression and classification variables in Alzheimer&#x00027;s disease</article-title>. <source>Neuroimage</source> <volume>59</volume>, <fpage>895</fpage>&#x02013;<lpage>907</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.neuroimage.2011.09.069</pub-id><pub-id pub-id-type="pmid">21992749</pub-id></mixed-citation>
</ref>
<ref id="B83">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y.</given-names></name> <name><surname>He</surname> <given-names>X.</given-names></name> <name><surname>Chan</surname> <given-names>Y. H.</given-names></name> <name><surname>Teng</surname> <given-names>Q.</given-names></name> <name><surname>Rajapakse</surname> <given-names>J. C.</given-names></name></person-group> (<year>2023</year>). <article-title>Multi-modal graph neural network for early diagnosis of Alzheimer&#x00027;s disease from sMRI and pet scans</article-title>. <source>Comput. Biol. Med</source>. <volume>164</volume>:<fpage>107328</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.107328</pub-id><pub-id pub-id-type="pmid">37573721</pub-id></mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2537780/overview">Moolchand Sharma</ext-link>, Maharaja Agrasen Institute of Technology, India</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1849955/overview">Vikash Yadav</ext-link>, Government Polytechnic Bighapur, Unnao, India</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2956520/overview">Monica Bhutani</ext-link>, Bharati Vidyapeeth&#x00027;s College of Engineering, India</p>
</fn>
</fn-group>
</back>
</article>