<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xml:lang="EN" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Neurosci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Neuroscience</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Neurosci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1662-453X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fnins.2026.1750045</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Toward automated neonatal EEG analysis: multi-center validation of a reliable deep learning pipeline</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Hermans</surname> <given-names>Tim</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2789070/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Dereymaeker</surname> <given-names>Anneleen</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/2905761/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Lemmens</surname> <given-names>Katrien</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/3373777/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Jansen</surname> <given-names>Katrien</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/679049/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Usman</surname> <given-names>Fatima</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Robinson</surname> <given-names>Shellie</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Naulaers</surname> <given-names>Gunnar</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/298178/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>De Vos</surname> <given-names>Maarten</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/1539398/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Hartley</surname> <given-names>Caroline</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<uri xlink:href="http://loop.frontiersin.org/people/691594/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Electrical Engineering (ESAT), STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, KU Leuven</institution>, <city>Leuven</city>, <country country="be">Belgium</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Development and Regeneration, KU Leuven</institution>, <city>Leuven</city>, <country country="be">Belgium</country></aff>
<aff id="aff3"><label>3</label><institution>Neonatal Intensive Care Unit, University Hospitals Leuven</institution>, <city>Leuven</city>, <country country="be">Belgium</country></aff>
<aff id="aff4"><label>4</label><institution>Child Neurology, University Hospitals Leuven</institution>, <city>Leuven</city>, <country country="be">Belgium</country></aff>
<aff id="aff5"><label>5</label><institution>Department of Paediatrics, University of Oxford</institution>, <city>Oxford</city>, <country country="gb">United Kingdom</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Tim Hermans, <email xlink:href="mailto:tim.hermans@esat.kuleuven.be">tim.hermans@esat.kuleuven.be</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-27">
<day>27</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>20</volume>
<elocation-id>1750045</elocation-id>
<history>
<date date-type="received">
<day>19</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>29</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>05</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Hermans, Dereymaeker, Lemmens, Jansen, Usman, Robinson, Naulaers, De Vos and Hartley.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Hermans, Dereymaeker, Lemmens, Jansen, Usman, Robinson, Naulaers, De Vos and Hartley</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-27">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Objectives</title>
<p>To evaluate the reliability and generalization of NeoNaid, a fully automated software tool for neonatal EEG analysis, based on functional brain age (FBA) estimation and sleep staging.</p>
</sec>
<sec>
<title>Methods</title>
<p>NeoNaid combines a multi-task deep learning model with proposed quality control routines detecting artifacts, out-of-distribution inputs, and uncertain predictions. Based on a raw EEG input, it outputs one global FBA estimate and a continuous 2-state hypnogram. We validated performance on two independent hospital settings: an internal dataset (33 EEGs, 17 infants, median 900 min/recording) and an external dataset (38 EEGs, 24 infants, median 124 min/recording).</p>
</sec>
<sec>
<title>Results</title>
<p>Quality control rejected a comparable number of segments in the internal and external datasets, reducing extreme errors in FBA estimation, and modestly improving sleep staging accuracy. Across the internal and external data, NeoNaid achieved median absolute FBA errors of 0.50 and 0.55 weeks and Cohen&#x2019;s Kappa values of 0.89 and 0.87 for quiet sleep detection, respectively.</p>
</sec>
<sec>
<title>Discussion</title>
<p>NeoNaid demonstrated improved reliability through integrated quality control and maintained performance across two independent datasets. By focusing on validation and trustworthiness, this work takes an essential step toward clinical adoption of automated neonatal EEG analysis and supports its utility for both NICU practice and large-scale research.</p>
</sec>
</abstract>
<kwd-group>
<kwd>automated analysis</kwd>
<kwd>clinical validation</kwd>
<kwd>functional brain age</kwd>
<kwd>neonatal EEG</kwd>
<kwd>quality control</kwd>
<kwd>sleep staging</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This research received funding from the European Union as part of the Cost action &#x201C;Maximizing impact of multidisciplinary research in early diagnosis of neonatal brain injury&#x201D; (AI-4-NICU) CA20124, Industrieel Onderzoeksfonds KU Leuven (IOF): C3 &#x201C;NEONAID - Neonatal Online Neuro-Analysis for Interpretation and Decision-support - Bringing automated neonatal monitoring to the bedside,&#x201D; C3/24/023, the Horizon program: HORIZON-HLTH-2022-IND-13: &#x201C;Privacy compliant health data as a service for AI development (PHASE IV AI),&#x201D; funded by the European Union, under Grant Agreement #101095384, the Flemish Government (AI Research Program), MDV and TH are affiliated to Leuven.AI - KU Leuven institute for AI, B-3000, Leuven, Belgium, and CH and SR were funded by the Wellcome Trust/Royal Society through a Sir Henry Dale Fellowship awarded to CH (grant reference: 213486/Z/18/Z). FU was funded by the Commonwealth Scholarship Commission.</funding-statement>
</funding-group>
<counts>
<fig-count count="6"/>
<table-count count="0"/>
<equation-count count="0"/>
<ref-count count="22"/>
<page-count count="10"/>
<word-count count="6377"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Brain Imaging Methods</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="S1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Neonatal electroencephalography (EEG) is the gold standard for assessing brain function in newborns and has proven to be a valuable monitoring tool in the neonatal intensive care unit (NICU) (<xref ref-type="bibr" rid="B11">McCoy and Hahn, 2013</xref>; <xref ref-type="bibr" rid="B9">Karamian and Wusthoff, 2021</xref>). With its high temporal resolution and multi-channel recordings, EEG provides rich information on brain maturation, sleep stages, and pathological activity such as seizures. However, neonatal EEG is challenging to interpret due to the complexity of the signal, and recordings often span many hours. Therefore, visual analysis requires expertise and is time-consuming. These factors limit the routine clinical use of EEG, despite its potential to provide valuable insights into neonatal brain health.</p>
<p>Over the past decade, a range of data-driven and artificial intelligence (AI) methods have been developed to assist with neonatal EEG interpretation. These include automated approaches for seizure detection (<xref ref-type="bibr" rid="B21">Temko et al., 2015</xref>; <xref ref-type="bibr" rid="B1">Ansari et al., 2019</xref>; <xref ref-type="bibr" rid="B4">Borovac et al., 2022</xref>; <xref ref-type="bibr" rid="B17">Raeisi et al., 2022</xref>), background grading (<xref ref-type="bibr" rid="B5">Dereymaeker et al., 2019</xref>; <xref ref-type="bibr" rid="B12">Moghadam et al., 2021</xref>; <xref ref-type="bibr" rid="B18">Raurale et al., 2021</xref>), sleep staging (<xref ref-type="bibr" rid="B16">Piryatinska et al., 2009</xref>; <xref ref-type="bibr" rid="B10">Koolen et al., 2017</xref>; <xref ref-type="bibr" rid="B2">Ansari et al., 2020</xref>), and functional brain age (FBA) estimation (<xref ref-type="bibr" rid="B15">Pillay et al., 2020</xref>; <xref ref-type="bibr" rid="B20">Stevenson et al., 2020</xref>; <xref ref-type="bibr" rid="B3">Ansari et al., 2024</xref>). Other advancements include the deployment of a cloud-based service, Babacloud, that implements an automated analysis pipeline computing neonatal EEG summary metrics (including measures of brain state and artifact detection) from uploaded EEG recordings (<xref ref-type="bibr" rid="B14">Montazeri et al., 2024</xref>). Furthermore, to improve robustness and generalization in EEG deep learning, recent work has also shown the power of self-supervised and multi-task training strategies (<xref ref-type="bibr" rid="B7">Hermans et al., 2023</xref>; <xref ref-type="bibr" rid="B13">Mohammadi Foumani et al., 2024</xref>; <xref ref-type="bibr" rid="B8">Hojjati et al., 2025</xref>). Such strategies allow models to be trained using multiple datasets from different cohorts, which may be unlabeled, or labeled for different tasks. By not restricting training to data collected for a single specific task, this approach enables the integration of heterogeneous datasets and learning from larger and more diverse data collections, thereby improving generalizability. Collectively, these developments highlight the potential of AI to scale EEG analysis beyond the limits of manual review and to provide decision-support tools for the clinical environment.</p>
<p>Of the various AI approaches to neonatal EEG, sleep staging and FBA estimation are especially informative for assessing neurodevelopment, and are the focus of this study. Sleep organization is an important marker of neurological development (<xref ref-type="bibr" rid="B6">Dereymaeker et al., 2017</xref>; <xref ref-type="bibr" rid="B19">Shellhaas et al., 2017</xref>). Automated sleep staging can provide continuous, objective measurements that would otherwise be impractical for clinicians to obtain. Similarly, FBA estimation offers a quantitative measure of brain maturation by comparing EEG-derived estimates of age with the infant&#x2019;s postmenstrual age (PMA). Deviations between the FBA and PMA can indicate atypical development and have prognostic value (<xref ref-type="bibr" rid="B20">Stevenson et al., 2020</xref>; <xref ref-type="bibr" rid="B3">Ansari et al., 2024</xref>). Together, these applications can support both clinical decision-making and long-term research on neonatal neurodevelopment.</p>
<p>Despite this potential, significant barriers remain to clinical adoption. Most published models are validated only on internal test data, raising concerns about their robustness to data from different hospitals, recording systems, or electrode montages. Long NICU recordings also unavoidably contain artifacts caused by movement, poor electrode contact, or physiological interference. Models trained primarily on clean data may fail when applied to such segments. For clinical usefulness, automated EEG tools must not only be accurate but also usable in practice and reliable across diverse datasets. Another important aspect for clinical deployment of automated EEG algorithms is their robustness to differences in recording setups. In neonatal care, EEG systems can vary in electrode montages, hardware characteristics, and acquisition protocols across centers. To be trusted in clinical workflows, algorithms must maintain performance despite these variations, including when applied to data from a hospital other than the one in which they were developed. Finally, to be adopted into clinical practice, tools must be easy to use with software that can aid interpretation.</p>
<p>To meet this need, we developed NeoNaid, a software tool that integrates a multi-task deep learning model for neonatal EEG analysis into a user-friendly graphical interface. Its design has been refined in discussion with clinicians to ensure interpretability and usability. This tool automatically processes long EEG recordings and provides robust, clinically relevant interpretations of the EEG, including sleep staging and FBA estimates. The underlying AI model builds on our previously published work and was trained on a large in-house dataset of neonatal EEG. Beyond this, NeoNaid implements quality control routines designed to improve the reliability and trustworthiness of the tool when used in clinical practice. These routines flag EEG segments likely to yield unreliable predictions by detecting artifacts, out-of-distribution inputs, or high model uncertainty.</p>
<p>In this paper, we focus on validating NeoNaid as a tool for neonatal EEG analysis. Unlike prior studies that primarily introduce new model architectures, our emphasis is on quality control and external validation. Specifically, we evaluate how NeoNaid performs on two independent datasets: an internal cohort from Leuven and an external cohort from Oxford. This cross-center validation is crucial for assessing generalizability and for building trust in real-world clinical use.</p>
<p>The aims of this study are 2-fold: (i) to assess the contribution of quality control to improving the reliability of automated FBA and sleep analysis, and (ii) to validate NeoNaid&#x2019;s performance for FBA estimation and sleep staging across both internal and external datasets. Together, these analyses address the key requirements for clinical usefulness: trustworthiness and generalizability.</p>
</sec>
<sec id="S2" sec-type="materials|methods">
<label>2</label>
<title>Materials and methods</title>
<sec id="S2.SS1">
<label>2.1</label>
<title>Automated EEG analysis using NeoNaid</title>
<p>NeoNaid is an in-house developed software tool for automated neonatal EEG analysis. It integrates preprocessing, deep learning&#x2013;based predictions, and quality control routines within a graphical user interface designed for clinical use (<xref ref-type="fig" rid="F1">Figure 1</xref>). A key feature of NeoNaid is its ability to handle variable input montages. The underlying model uses a channel-agnostic architecture with shared weights to process each channel independently, followed by an attention-based mechanism to aggregate per-channel predictions.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption><p>Screenshot of the NeoNaid GUI. The graphs show the predicted sleep hypnogram and the brain age as a function of time during a 16 h 45 min recording. Besides these visual outputs, the GUI produces result files which were used to analyze the results for this paper.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g001.tif">
<alt-text content-type="machine-generated">NeoNaid software interface displays patient and file information with sensitive data redacted. Two main panels show results: the sleep panel presents hypnogram data with wake and sleep stages over a sixteen-hour period, while the brain age panel displays a green line graph of functional brain age in weeks, with a median value of forty point six and an interquartile range of one point one. Movement and artefact annotations appear as colored marks beneath both graphs.</alt-text>
</graphic>
</fig>
<p>The input to NeoNaid is a raw multi-channel EEG recording. Predictions are generated for non-overlapping 30-s segments. For each channel in the input EEG, NeoNaid produces four main outputs: (i) an artifact mask indicating which samples in the input are likely to be noise; (ii) a prediction of the sleep stage: quiet sleep (QS), active sleep (AS), or wake; (iii) an estimate of functional brain age (FBA); (iv) an attention weight for downstream aggregation. After computation of these per-channel outputs, the attention weights are used to aggregate the per-channel sleep and FBA predictions into a single global output by means of a weighted average. This mechanism allows NeoNaid to prioritize informative signals and mitigate the impact of noisy signals.</p>
<p>The core of the software is a multi-task deep learning model, developed by converging our previously published models (<xref ref-type="bibr" rid="B1">Ansari et al., 2019</xref>, <xref ref-type="bibr" rid="B2">2020</xref>, <xref ref-type="bibr" rid="B3">2024</xref>; <xref ref-type="bibr" rid="B5">Dereymaeker et al., 2019</xref>; <xref ref-type="bibr" rid="B7">Hermans et al., 2023</xref>) into one common methodology. Its architecture builds on our previously published convolutional neural network (<xref ref-type="bibr" rid="B7">Hermans et al., 2023</xref>) and consists of a shared encoder connected to multiple output heads, one for each task. Unlike prior single-task models, NeoNaid was trained in a multi-task setting, enabling simultaneous supervised learning from different neonatal EEG datasets labeled for different tasks. The training data included approximately 1,326 h of EEG from 124 recordings with age labels for FBA estimation, 565 h from 132 recordings with sleep annotations, and 44 h from 73 recordings with artifact annotations.</p>
</sec>
<sec id="S2.SS2">
<label>2.2</label>
<title>Dataset</title>
<p>We evaluated NeoNaid using two independent neonatal EEG datasets: one internal dataset collected at University Hospitals Leuven, Belgium (Dataset A) and one external dataset recorded at the John Radcliffe Hospital in Oxford, UK (Dataset B). These datasets differ in acquisition hardware, electrode configuration, and local recording protocols, allowing us to assess generalization of the algorithm to different recording conditions. The internal Dataset A originated from the same hospital as the data used for the development of the NeoNaid algorithms. However, all recordings Dataset A were independent of the development data and thus represent previously unseen cases.</p>
<sec id="S2.SS2.SSS1">
<label>2.2.1</label>
<title>Dataset A (internal)</title>
<p>Dataset A consists of 33 EEG recordings from 17 neonates monitored at the NICU of University Hospitals Leuven, Belgium. The cohort includes both term and preterm infants with postmenstrual ages (PMA) at the time of recording ranging from 27.3 to 47 weeks (median 37.7). EEGs were acquired using the BrainRT EEG system (Onafhankelijke Software Groep (OSG), Kontich, Belgium) with a sampling rate of 250 or 256 Hz. Electrodes were positioned according to a modified 10&#x2013;20 system, including the following channels: Fp1, Fp2, C3, C4, T3, T4, O1, and O2, with Cz as the reference electrode.</p>
<p>The EEG recordings had a median duration of 900 (IQR: 90&#x2013;980) minutes and were collected as part of routine clinical care. All recordings came from neonates without major developmental abnormalities, allowing PMA at recording to serve as a reference measure for FBA. Two-class sleep annotations (QS versus AS or wake) were available for a subset of 28 recordings, scored by trained experts. All data were anonymized prior to analysis. The study was approved by the local ethics committee of University Hospitals Leuven, in accordance with the Declaration of Helsinki.</p>
</sec>
<sec id="S2.SS2.SSS2">
<label>2.2.2</label>
<title>Dataset B (external)</title>
<p>Dataset B comprises 38 EEG recordings from 24 neonates recorded at the Newborn Care Unit of the John Radcliffe Hospital, Oxford University Hospitals NHS Foundation Trust, Oxford, UK. The PMA at the time of recording ranged from 29.4 to 41.4 (median 34.5) weeks. Compared to Dataset A, Dataset B used a different EEG system and montage. EEGs were recorded using the SynAmps RT 64-channel headbox and amplifiers and CURRYscan7 neuroimaging suite (Compumedics Neuroscan) with a sampling rate of 2,000 Hz. The electrode configuration consisted of FCz, C3, C4, Cz, CPz, T3, T4, and Oz, referenced to Fz.</p>
<p>Recordings had a median duration of 124 (IQR: 91&#x2013;143) minutes. Similar to dataset A, all recordings were from patients without any neurological abnormalities (infants were excluded from studies if they had a grade III or IV intraventricular hemorrhage, hypoxic ischemic encephalopathy, or major congenital malformations), making the PMA at time of recording a suitable reference for functional brain age. Of the 38 EEGs, 18 were randomly selected for sleep staging, ensuring good coverage across PMA and recording durations, and were subsequently labeled for sleep by a trained expert in Leuven.</p>
<p>The EEG recordings were collected for research purposes, as part of an independent study (<xref ref-type="bibr" rid="B22">Usman et al., 2025</xref>). Eligible families were given verbal and written information about the study, and written parental consent was signed before inclusion in the study. This dataset was fully anonymized and provided under a data sharing agreement between the University of Oxford and KU Leuven. Ethical approval was obtained through the relevant UK regulatory bodies (National Research Ethics Service reference: 12/SC/0447).</p>
</sec>
<sec id="S2.SS2.SSS3">
<label>2.2.3</label>
<title>Pre-processing</title>
<p>To ensure consistency across datasets and to reduce the influence of differing reference electrodes, EEG recordings were transformed into a common bipolar montage using electrode pairs present in both datasets (C3&#x2013;C4, C3&#x2013;Cz, C3&#x2013;T4, C4&#x2013;Cz, C4&#x2013;T3). This was achieved by re-referencing channel signals accordingly using electrode pairs present in both datasets. No other channel interpolation was applied. Following this, signals were bandpass-filtered (0.25&#x2013;30 Hz) and downsampled to 64 Hz. The recordings were then segmented into 30-s non-overlapping epochs. Within each recording, channel amplitudes were normalized by the median standard deviation across all segments. These pre-processing steps ensured consistency across acquisition systems and were automatically executed by the NeoNaid software.</p>
</sec>
</sec>
<sec id="S2.SS3">
<label>2.3</label>
<title>Quality control</title>
<p>Typically, the majority of training data is from a clean and labeled dataset, thus making existing trained models unreliable for artifact-containing data segments. Additionally, deep learning models perform well on data that is similar to the training data, but their performance becomes unreliable when applied to data that has significantly different characteristics compared to the training data (out-of-distribution data).</p>
<p>A central feature of NeoNaid is its quality control algorithm, which evaluates the reliability of each 30-s EEG segment before downstream interpretation. This process involves evaluating three independent reliability criteria: artifact content, novelty detection (to detect out-of-distribution inputs), and (un)certainty level.</p>
<sec id="S2.SS3.SSS1">
<label>2.3.1</label>
<title>Artifacts</title>
<p>Each segment is assigned an artifact score based on the percentage of samples identified as noise by the dedicated detection head of the model. Segments with more than 50% artifact content are flagged as unreliable, therefore preventing the model from producing predictions on segments where brain activity is largely obscured. This threshold was informed by an analysis of the relationship between performance and artifact content during NeoNaid&#x2019;s development using independent datasets. As most segments exhibited either very low or very high artifact content, the precise threshold was not critical and the 50% cutoff was adopted as a pragmatic criterion. Segments that do not pass the artifact check are flagged as unreliable and either excluded (for FBA) or explicitly marked (for sleep staging).</p>
</sec>
<sec id="S2.SS3.SSS2">
<label>2.3.2</label>
<title>Novelties</title>
<p>Out-of-distribution inputs (novelties) are automatically identified by applying a novelty detection model to each channel in a segment. The novelty detection model in NeoNaid is an isolation forest that uses a set of nine spectral features as input (<xref ref-type="bibr" rid="B16">Piryatinska et al., 2009</xref>), predicting for each channel in every segment whether it is an inlier or a novelty (with respect to the NeoNaid training data). A separate novelty detection model was fitted per task, using the data that was used for training that part of the multi-task model. Channels that was labeled as a novelty by NeoNaid were excluded from the model&#x2019;s channel aggregation, reducing their impact on the global predictions. Furthermore, if more than half of the channels in a segment were labeled as novelties, the entire segment was flagged.</p>
</sec>
<sec id="S2.SS3.SSS3">
<label>2.3.3</label>
<title>Uncertainty</title>
<p>One of the outputs of NeoNaid are channel-wise attention weights. After normalization, these weights are used to compute a weighted average of channel-wise predictions. Prior to normalization, these channel-specific attention weights reflect the model&#x2019;s certainty regarding each channel&#x2019;s input. To identify segments where all channels are deemed uncertain, the maximum unnormalized attention weight across channels was computed for each segment. Segments where this maximum attention weight falls below a predefined threshold are flagged as unreliable. This threshold was defined and fixed during NeoNaid&#x2019;s development as the 1<italic><sup>st</sup></italic> percentile of maximum attention weights observed in an independent calibration dataset. In addition to attention-based flagging, for the sleep staging outputs, segments with QS probabilities near 0.5 were also marked as uncertain.</p>
<p>Together, these three criteria provide a conservative safeguard against unreliable predictions. NeoNaid then aggregates the segment-wise outputs and quality flags into clinically interpretable results. For FBA, a single robust estimate is obtained as the median across reliable segments, while for sleep staging, a continuous hypnogram is constructed by smoothing probabilities and interpolating over short, unreliable intervals using simple heuristic rules.</p>
</sec>
</sec>
<sec id="S2.SS4">
<label>2.4</label>
<title>Performance metrics</title>
<sec id="S2.SS4.SSS1">
<label>2.4.1</label>
<title>Functional brain age</title>
<p>For each recording, the global FBA estimate was defined as the median of all segment-wise predictions that passed quality control. The performance was quantified in terms of the absolute error, which was defined as the absolute difference between the global FBA estimate and the infant&#x2019;s PMA at the time of recording. Lower errors indicate better performance. The interquartile range (IQR) of the retained segment-wise estimates was reported as a measure of prediction confidence, with wider IQRs indicating a lower prediction certainty.</p>
</sec>
<sec id="S2.SS4.SSS2">
<label>2.4.2</label>
<title>Sleep staging</title>
<p>For sleep analysis, we evaluated NeoNaid&#x2019;s ability to detect QS. To this end, the model&#x2019;s AS and wake predictions were combined into a single category, representing the non-quiet sleep class. Predictions on segments flagged as unreliable were excluded. The performance was measured in terms of Cohen&#x2019;s kappa score, where a higher score indicates better agreement between predicted sleep stages and expert annotations.</p>
</sec>
</sec>
<sec id="S2.SS5">
<label>2.5</label>
<title>Analysis</title>
<p>Our analyses were designed to evaluate both the impact of quality control and the generalizability of NeoNaid across datasets. We applied the full processing pipeline to both datasets: Dataset A (internal) and Dataset B (external). We compared two approaches: a na&#x00EF;ve approach that included all segment-wise predictions, and a robust approach that excluded segments flagged by quality control.</p>
<sec id="S2.SS5.SSS1">
<label>2.5.1</label>
<title>Effect of quality control</title>
<p>To check for differences in data quality and characteristics between the two datasets, we first quantified the occurrence of quality control flags within each dataset. More specifically, we computed the proportion of EEG segments flagged for artifacts and novelty detection; these were independently assessed for the FBA and sleep tasks.</p>
<p>For FBA estimation, performance was evaluated as a function of EEG recording duration, as the effect of the robust method is most clear for EEGs of shorter durations. To simulate varying EEG durations, we extracted sub-epochs ranging from 30 s to 1 h from each recording. For each duration, 1,000 sub-epochs were randomly selected per recording. The median FBA and corresponding performance metrics were calculated for each duration, allowing assessment of how recording length and the inclusion of quality control routines affect prediction errors.</p>
<p>For sleep staging, we computed Cohen&#x2019;s kappa score for quiet sleep detection on the full recordings using both the naive and robust approaches and compared the performance.</p>
</sec>
<sec id="S2.SS5.SSS2">
<label>2.5.2</label>
<title>Cross-center validation</title>
<p>Finally, we validated the robust methodology (i.e., including quality control) on the complete EEG recordings. Per-recording results were analyzed and visualized in two ways. Firstly, FBA and sleep performance metrics are reported for each channel separately, i.e., using the per-channel predictions prior to channel aggregation, as well as for the global result obtained after aggregation of the per-channel predictions. Secondly, we showed the performance as a function of PMA to investigate whether prediction accuracy is systematically affected by the age of the neonates.</p>
</sec>
</sec>
</sec>
<sec id="S3" sec-type="results">
<label>3</label>
<title>Results</title>
<sec id="S3.SS1">
<label>3.1</label>
<title>NeoNaid quality control reduces errors</title>
<p>We investigated how the quality control routine affects the automated analysis. For FBA, the median rejection rate was 21.5% in the internal dataset and 16.1% in the external dataset (<xref ref-type="fig" rid="F2">Figure 2</xref>). These were lower for sleep staging, with medians of 2.0% (internal) and 2.4% (external), mainly due to the heuristic postprocessing, which interpolates short unreliable intervals, and retains segments with high-amplitude movement artifacts when they occur during predicted wake cycle.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption><p>Rejection rates due to quality control in the two datasets. Rejected %: the total percentage of segments that were rejected in each recording.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g002.tif">
<alt-text content-type="machine-generated">Box plot comparing percentages of rejected segments for FBA and QS methods with internal and external categories. FBA shows higher medians and greater variability than QS, with several outliers above 40 percent for FBA.</alt-text>
</graphic>
</fig>
<p>Analyzing the rejection rates according to the three quality control criteria (which are not mutually exclusive), segments were most frequently flagged as artifacts. In the FBA data, 22.9% of segments in the internal dataset and 19.6% in the external dataset were marked as artifactual; similar rates were observed for sleep data (22.3% and 19.3%, respectively). Novelty detection contributed less, flagging 1.1% and 2.4% of FBA segments, and 4.1% and 3.4% of sleep segments in the internal and external data, respectively. The contribution of segments flagged as uncertain for FBA was 3.3% (internal) and 7.2% (external), and for sleep 4.6% (internal) and 2.2% (external). For sleep staging, these values are observed prior to heuristic postprocessing, thus accounting for the generally lower rejection percentages than the proportion of initially flagged segments. Overall, data quality was comparable between centers, and the external data did not appear out-of-distribution despite differences in acquisition systems and protocols.</p>
<p>Next, we investigated the effect of the quality control on FBA performance in both datasets. Overall, the robust method (which applies segment rejection) and the naive method (which does not) yielded similar median FBA error values (<xref ref-type="fig" rid="F3">Figure 3</xref>). This is expected, as the use of the median as an aggregation metric reduces the influence of outlier segments. Nevertheless, the robust method consistently showed a lower likelihood of producing extreme outliers, particularly in shorter recordings, demonstrating its value in minimizing risk. The median IQR of the FBA estimate initially increased with data length and then plateaued after approximately 20 min of usable data. This suggests that IQR values in recordings with less than 20 min of non-rejected EEG should be interpreted with care, as limited data availability may underestimate the IQR of the actual underlying distribution.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption><p>FBA performance for different simulated data lengths (1,000 repetitions per window, per recording). The x-axis represents the duration of data used to compute the estimates (excluding the rejected segments in the robust case), of which median + min&#x2013;max is shown. Robust method includes quality control and subsequent segment rejection, na&#x00EF;ve method does not. Left: internal Dataset A, right: external Dataset B.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g003.tif">
<alt-text content-type="machine-generated">Four line graphs compare FBA absolute error and interquartile range in weeks across used data lengths in minutes, separated by Internal and External data. Na&#x00EF;ve method (green) shows higher error and variance than Robust method (red), with differences decreasing as data length increases.</alt-text>
</graphic>
</fig>
<p>Quiet sleep detection performance using the naive and robust methods (<xref ref-type="fig" rid="F4">Figure 4</xref>) showed modest performance improvement following quality control on the already relatively cleaned and labeled sleep data. Unusable segments were not annotated by the experts and, therefore, were not included in the performance evaluation. While the benefit of quality control is less pronounced in this evaluation setup, its primary value lies in preventing unreliable predictions when poor-quality data are encountered in practice.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption><p>Sleep performance with (robust) and without (naive) artifact rejection and heuristic postprocessing.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g004.tif">
<alt-text content-type="machine-generated">Boxplot compares Cohen&#x2019;s kappa values for internal and external tests under Naive and Robust conditions. Both conditions show higher kappa values internally. Robust condition slightly increases minimum kappa for external tests. Outliers visible below each distribution.</alt-text>
</graphic>
</fig>
</sec>
<sec id="S3.SS2">
<label>3.2</label>
<title>Cross-center validation of NeoNaid</title>
<p>We investigated the results of the robust method (with quality control) of both datasets to validate the cross-center generalization of NeoNaid. Using the robust methodology on full recordings, the mean absolute FBA error across the datasets is 0.60 (median: 0.50, IQR: 0.21&#x2013;0.78) weeks for Dataset A, and 0.69 (median: 0.55, IQR: 0.23&#x2013;1.02) weeks for Dataset B. The percentage of recordings within 1 week error is 79% (100% within 2 weeks) and 74% (97% within 2 weeks) for Dataset A and Dataset B, respectively. Moreover, the true PMA for Dataset A fell within the IQR of the per-segment predictions in 70% of recordings, compared to 58% for Dataset B.</p>
<p>For QS detection, performance was high in both datasets. In Dataset A, per-recording kappa scores averaged 0.86 (median: 0.89, IQR: 0.80&#x2013;0.96), while in Dataset B they averaged 0.79 (median: 0.87, IQR: 0.70&#x2013;0.94). Comparable results were obtained when restricting the analysis to single channels, particularly for bipolar derivations around C3, Cz, and C4 (<xref ref-type="fig" rid="F5">Figure 5</xref>). When pooling all recordings, the overall kappa was 0.874 (95% CI: 0.865&#x2013;0.884) in Dataset A and 0.831 (95% CI: 0.811&#x2013;0.851) in Dataset B.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption><p>Per-recording performance per channel. &#x201C;All&#x201D; refers to attention-weighted average of single-channel predictions. Top: brain age error; bottom: quiet sleep detection score. Left: internal Dataset A, right: external Dataset B.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g005.tif">
<alt-text content-type="machine-generated">Four-panel boxplot graphic comparing FBA absolute error in weeks (top row) and Cohen&#x2019;s kappa (bottom row) for internal (left, blue) versus external (right, tan) datasets across six EEG channel configurations. Separate distributions are shown for each configuration and aggregated results, for both metrics.</alt-text>
</graphic>
</fig>
<p>The performance of FBA and quiet sleep detection was relatively consistent across channels (<xref ref-type="fig" rid="F5">Figure 5</xref>), suggesting that the software can deliver reliable results even with limited or single-channel input. However, while combining all channels did not always outperform the best individual channel in terms of median performance, it generally helped reduce errors in outlier recordings, offering improved robustness.</p>
<p>Finally, we assessed the model performance compared with the age of the infants. For the FBA model, there is no significant linear correlation between the FBA error and PMA (<xref ref-type="fig" rid="F6">Figure 6</xref>), indicating that the model performed equally at all ages. In contrast, for sleep staging, there was a trend that QS performance improves with increasing age (<xref ref-type="fig" rid="F6">Figure 6</xref>). In one case in Dataset B, the model failed to detect any quiet sleep, whereas the expert annotation indicated a 15-min QS bout, resulting in a Kappa score of zero. The 15-min QS epoch included multiple high-amplitude artifacts, which resulted in the model misclassifying it as wake.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption><p>The relation between the performance and postmenstrual age (PMA). Left: internal Dataset A, right: external Dataset B.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fnins-20-1750045-g006.tif">
<alt-text content-type="machine-generated">Four scatterplots compare internal (left, blue) and external (right, orange) performance; top row shows FBA absolute error versus PMA with weak, non-significant correlations, while bottom row shows sleep performance versus PMA with a strong, significant correlation internally and a weaker, non-significant correlation externally.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="S4" sec-type="discussion">
<label>4</label>
<title>Discussion</title>
<p>The aim of this study was 2-fold: first, to assess the contribution of quality control routines in improving the reliability of automated neonatal EEG analysis, and second, to validate the performance of NeoNaid across independent datasets from two hospital settings.</p>
<p>We found that quality control reduced extreme errors and provided transparent confidence measures, particularly for functional brain age estimation in shorter or noisier recordings. NeoNaid&#x2019;s integrated quality control framework assesses the reliability of each EEG segment through artifact detection, novelty detection, and attention-based certainty scoring. Our results show that this quality control framework improves the reliability of automated neonatal EEG analysis by reducing the likelihood of extreme errors while preserving valid information. While average performance metrics changed only modestly, the safeguards proved valuable in minimizing risk, particularly in shorter or lower-quality recordings (<xref ref-type="fig" rid="F3">Figures 3</xref>, <xref ref-type="fig" rid="F4">4</xref>). Therefore, these routines increase the clinical trust in the outputs generated by NeoNaid.</p>
<p>NeoNaid maintained performance when validated on an independent, external dataset, despite differences in acquisition hardware, montages, and protocols. Additionally, the external data were not identified as out-of-distribution by the quality control routines. This cross-center validation provides preliminary evidence that the approach can generalize across clinical environments, addressing an important barrier to the adoption of automated neonatal EEG analysis.</p>
<p>A key feature of NeoNaid is that it is not restricted to a specific montage or number of channels. As shown in <xref ref-type="fig" rid="F5">Figure 5</xref>, the model produces comparable results when operating on a single bipolar derivation compared to the full five-channel setup. This is particularly relevant for clinical scenarios with limited channel availability, such as amplitude-integrated EEG (aEEG) monitoring, where only one or two channels may be accessible. This flexibility further enhances NeoNaid&#x2019;s clinical applicability across different monitoring environments.</p>
<p>Performance did not vary significantly with PMA for FBA, although there was a tendency toward slightly larger errors in recordings below 30 or above 40 weeks PMA. This pattern likely reflects the limited amount of training and validation data available in these extreme age ranges rather than a systematic bias. In contrast, sleep staging performance clearly improved with age. As shown in <xref ref-type="fig" rid="F6">Figure 6</xref>, older neonates achieved higher kappa scores for quiet sleep detection, whereas younger infants showed slightly lower performance. This is likely due to the less distinct differentiation between active and quiet sleep at earlier developmental stages.</p>
<p>Several prior studies have developed deep learning models for brain age estimation (<xref ref-type="bibr" rid="B15">Pillay et al., 2020</xref>; <xref ref-type="bibr" rid="B20">Stevenson et al., 2020</xref>; <xref ref-type="bibr" rid="B3">Ansari et al., 2024</xref>) and sleep staging. While direct comparison of performance metrics across studies should be done cautiously due to differences in test datasets, our findings show that the FBA performance of NeoNaid (mean absolute errors of 0.60 and 0.69 on internal and external datasets, respectively) is comparable, if not superior, to these earlier models, which typically report mean absolute errors between 0.7 and 1.0 weeks. Similarly, NeoNaid&#x2019;s performance in QS detection aligns with prior models that have reported kappa values up to 0.77.</p>
<p>NeoNaid offers value for both clinical and research applications. In clinical practice, automatic sleep staging and FBA estimation can aid in monitoring brain development, particularly in preterm infants. The IQR accompanying the FBA estimate provides a practical measure of confidence, helping clinicians interpret results more effectively. The built-in quality control indicators can alert users to unreliable segments, reducing the risk of misinterpretation due to artifacts or signal degradation. For researchers, NeoNaid provides a scalable solution for annotating large EEG datasets in a standardized way. It is especially useful in studies investigating neurodevelopmental trajectories, sleep&#x2013;wake organization, and responses to therapeutic interventions.</p>
<p>This study has several limitations. The number of recordings in the external dataset was relatively small, limiting the statistical power of generalization claims. Sleep annotations were made by different raters at each site, without a formal assessment of inter-rater reliability, which could introduce bias. Moreover, the external recordings were shorter in duration and lacked accompanying physiological or video data, which made them more challenging to annotate. Finally, external validation was limited to a single center. Future work will focus on expanding collaborations to include additional external datasets and broader populations to further validate and refine the NeoNaid platform.</p>
<p>To conclude, we have shown that NeoNaid is a robust tool for automated neonatal EEG analysis, maintaining performance across two datasets with different recording setups. Its integrated quality control routines reduce extreme errors and improve trustworthiness, addressing a critical requirement for clinical adoption. This represents an important step toward the broader use of AI tools in neonatal EEG, where differences in acquisition setups are common and difficult to standardize. Ultimately, these result support the potential of NeoNaid for both NICU practice and large-scale research on neonatal brain monitoring.</p>
</sec>
</body>
<back>
<sec id="S5" sec-type="data-availability">
<title>Data availability statement</title>
<p>The data analyzed in this study is subject to the following licenses/restrictions: the datasets presented in this article are not readily available because they contain sensitive medical data and cannot be publicly shared for privacy and ethical reasons. Requests to access the Leuven data should be directed to TH, <email xlink:href="mailto:tim.hermans@esat.kuleuven.be">tim.hermans@esat.kuleuven.be</email>, and requests to access the Oxford data should be directed to CH, <email xlink:href="mailto:caroline.hartley@paediatrics.ox.ac.uk">caroline.hartley@paediatrics.ox.ac.uk</email>.</p>
</sec>
<sec id="S6" sec-type="ethics-statement">
<title>Ethics statement</title>
<p>The studies involving humans were approved by Medical Ethics Committee of University Hospitals Leuven (UZ Leuven), Belgium, and NHS Health Research Authority &#x2014; National Research Ethics Service (NRES), United Kingdom. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants&#x2019; legal guardians/next of kin.</p>
</sec>
<sec id="S7" sec-type="author-contributions">
<title>Author contributions</title>
<p>TH: Conceptualization, Methodology, Software, Validation, Writing &#x2013; original draft. AD: Conceptualization, Investigation, Methodology, Writing &#x2013; review &#x0026; editing. KL: Writing &#x2013; review &#x0026; editing, Investigation. KJ: Conceptualization, Investigation, Writing &#x2013; review &#x0026; editing. FU: Investigation, Writing &#x2013; review &#x0026; editing. SR: Investigation, Writing &#x2013; review &#x0026; editing. GN: Conceptualization, Investigation, Writing &#x2013; review &#x0026; editing. MDV: Conceptualization, Methodology, Resources, Supervision, Writing &#x2013; review &#x0026; editing. CH: Conceptualization, Methodology, Resources, Supervision, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<sec id="S9" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
<p>The author GN declared that they were an editorial board member of Frontiers at the time of submission. This had no impact on the peer review process and the final decision.</p>
</sec>
<sec id="S10" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was used in the creation of this manuscript. During the preparation of this work, the authors used ChatGPT in order to improve readability and language in portions of the manuscript. After using this tool/service, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="S11" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ansari</surname> <given-names>A.</given-names></name> <name><surname>Cherian</surname> <given-names>P. J.</given-names></name> <name><surname>Caicedo</surname> <given-names>A.</given-names></name> <name><surname>Naulaers</surname> <given-names>G.</given-names></name> <name><surname>De Vos</surname> <given-names>M.</given-names></name> <name><surname>Van Huffel</surname> <given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Neonatal seizure detection using deep convolutional neural networks.</article-title> <source><italic>Int. J. Neural Syst.</italic></source> <volume>29</volume>:<fpage>1850011</fpage>. <pub-id pub-id-type="doi">10.1142/s0129065718500119</pub-id> <pub-id pub-id-type="pmid">29747532</pub-id></mixed-citation></ref>
<ref id="B2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ansari</surname> <given-names>A.</given-names></name> <name><surname>De Wel</surname> <given-names>O.</given-names></name> <name><surname>Pillay</surname> <given-names>K.</given-names></name> <name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Jansen</surname> <given-names>K.</given-names></name> <name><surname>Van Huffel</surname> <given-names>S.</given-names></name><etal/></person-group> (<year>2020</year>). <article-title>A convolutional neural network outperforming state-of-the-art sleep staging algorithms for both preterm and term infants.</article-title> <source><italic>J. Neural Eng.</italic></source> <volume>17</volume>:<fpage>16028</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/ab5469</pub-id> <pub-id pub-id-type="pmid">31689694</pub-id></mixed-citation></ref>
<ref id="B3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ansari</surname> <given-names>A.</given-names></name> <name><surname>Pillay</surname> <given-names>K.</given-names></name> <name><surname>Arasteh</surname> <given-names>E.</given-names></name> <name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Mellado</surname> <given-names>G. S.</given-names></name> <name><surname>Jansen</surname> <given-names>K.</given-names></name><etal/></person-group> (<year>2024</year>). <article-title>Resting state electroencephalographic brain activity in neonates can predict age and is indicative of neurodevelopmental outcome.</article-title> <source><italic>Clin. Neurophysiol.</italic></source> <volume>163</volume> <fpage>226</fpage>&#x2013;<lpage>235</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2024.05.002</pub-id> <pub-id pub-id-type="pmid">38797002</pub-id></mixed-citation></ref>
<ref id="B4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Borovac</surname> <given-names>A.</given-names></name> <name><surname>Gudmundsson</surname> <given-names>S.</given-names></name> <name><surname>Thorvardsson</surname> <given-names>G.</given-names></name> <name><surname>Moghadam</surname> <given-names>S. M.</given-names></name> <name><surname>Nevalainen</surname> <given-names>P.</given-names></name> <name><surname>Stevenson</surname> <given-names>N.</given-names></name><etal/></person-group> (<year>2022</year>). <article-title>Ensemble Learning Using Individual Neonatal Data for Seizure Detection.</article-title> <source><italic>IEEE J. Transl. Eng. Health Med.</italic></source> <volume>10</volume>:<fpage>4901111</fpage>. <pub-id pub-id-type="doi">10.1109/jtehm.2022.3201167</pub-id> <pub-id pub-id-type="pmid">36147876</pub-id></mixed-citation></ref>
<ref id="B5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Matic</surname> <given-names>V.</given-names></name> <name><surname>Vervisch</surname> <given-names>J.</given-names></name> <name><surname>Cherian</surname> <given-names>P. J.</given-names></name> <name><surname>Ansari</surname> <given-names>A.</given-names></name> <name><surname>De Wel</surname> <given-names>O.</given-names></name><etal/></person-group> (<year>2019</year>). <article-title>Automated EEG background analysis to identify neonates with hypoxic-ischemic encephalopathy treated with hypothermia at risk for adverse outcome: A pilot study.</article-title> <source><italic>Pediatr. Neonatol.</italic></source> <volume>60</volume> <fpage>50</fpage>&#x2013;<lpage>58</lpage>. <pub-id pub-id-type="doi">10.1016/j.pedneo.2018.03.010</pub-id> <pub-id pub-id-type="pmid">29705390</pub-id></mixed-citation></ref>
<ref id="B6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Pillay</surname> <given-names>K.</given-names></name> <name><surname>Vervisch</surname> <given-names>J.</given-names></name> <name><surname>De Vos</surname> <given-names>M.</given-names></name> <name><surname>Van Huffel</surname> <given-names>S.</given-names></name> <name><surname>Jansen</surname> <given-names>K.</given-names></name><etal/></person-group> (<year>2017</year>). <article-title>Review of sleep-EEG in preterm and term neonates.</article-title> <source><italic>Early Hum. Dev.</italic></source> <volume>113</volume> <fpage>87</fpage>&#x2013;<lpage>103</lpage>. <pub-id pub-id-type="doi">10.1016/j.earlhumdev.2017.07.003</pub-id> <pub-id pub-id-type="pmid">28711233</pub-id></mixed-citation></ref>
<ref id="B7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hermans</surname> <given-names>T.</given-names></name> <name><surname>Smets</surname> <given-names>L.</given-names></name> <name><surname>Lemmens</surname> <given-names>K.</given-names></name> <name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Jansen</surname> <given-names>K.</given-names></name> <name><surname>Naulaers</surname> <given-names>G.</given-names></name><etal/></person-group> (<year>2023</year>). <article-title>A multi-task and multi-channel convolutional neural network for semi-supervised neonatal artefact detection.</article-title> <source><italic>J. Neural Eng.</italic></source> <volume>20</volume>:<fpage>26013</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/acbc4b</pub-id> <pub-id pub-id-type="pmid">36791462</pub-id></mixed-citation></ref>
<ref id="B8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hojjati</surname> <given-names>A.</given-names></name> <name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Hameed</surname> <given-names>I.</given-names></name> <name><surname>Yazidi</surname> <given-names>A.</given-names></name> <name><surname>Lind</surname> <given-names>P. G.</given-names></name> <name><surname>Khadka</surname> <given-names>R.</given-names></name></person-group> (<year>2025</year>). <article-title>From video to EEG: Adapting joint embedding predictive architecture to uncover visual concepts in brain signal analysis.</article-title> <source><italic>arXiv</italic></source> [<comment>Preprint</comment>] <pub-id pub-id-type="doi">10.48550/arXiv.2507.03633</pub-id></mixed-citation></ref>
<ref id="B9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Karamian</surname> <given-names>A. G. S.</given-names></name> <name><surname>Wusthoff</surname> <given-names>C. J.</given-names></name></person-group> (<year>2021</year>). <article-title>Current and future uses of continuous EEG in the NICU.</article-title> <source><italic>Front. Pediatr.</italic></source> <volume>9</volume>:<fpage>768670</fpage>. <pub-id pub-id-type="doi">10.3389/fped.2021.768670</pub-id> <pub-id pub-id-type="pmid">34805053</pub-id></mixed-citation></ref>
<ref id="B10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Koolen</surname> <given-names>N.</given-names></name> <name><surname>Oberdorfer</surname> <given-names>L.</given-names></name> <name><surname>Rona</surname> <given-names>Z.</given-names></name> <name><surname>Giordano</surname> <given-names>V.</given-names></name> <name><surname>Werther</surname> <given-names>T.</given-names></name> <name><surname>Klebermass-Schrehof</surname> <given-names>K.</given-names></name><etal/></person-group> (<year>2017</year>). <article-title>Automated classification of neonatal sleep states using EEG.</article-title> <source><italic>Clin. Neurophysiol.</italic></source> <volume>128</volume> <fpage>1100</fpage>&#x2013;<lpage>1108</lpage>. <pub-id pub-id-type="doi">10.1016/j.clinph.2017.02.025</pub-id> <pub-id pub-id-type="pmid">28359652</pub-id></mixed-citation></ref>
<ref id="B11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>McCoy</surname> <given-names>B.</given-names></name> <name><surname>Hahn</surname> <given-names>C. D.</given-names></name></person-group> (<year>2013</year>). <article-title>Continuous EEG monitoring in the neonatal intensive care unit.</article-title> <source><italic>J. Clin. Neurophysiol.</italic></source> <volume>30</volume> <fpage>106</fpage>&#x2013;<lpage>114</lpage>. <pub-id pub-id-type="doi">10.1097/WNP.0b013e3182872919</pub-id> <pub-id pub-id-type="pmid">23545760</pub-id></mixed-citation></ref>
<ref id="B12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moghadam</surname> <given-names>S. M.</given-names></name> <name><surname>Pinchefsky</surname> <given-names>E.</given-names></name> <name><surname>Tse</surname> <given-names>I.</given-names></name> <name><surname>Marchi</surname> <given-names>V.</given-names></name> <name><surname>Kohonen</surname> <given-names>J.</given-names></name> <name><surname>Kauppila</surname> <given-names>M.</given-names></name><etal/></person-group> (<year>2021</year>). <article-title>Building an open source classifier for the neonatal EEG background: A systematic feature-based approach from expert scoring to clinical visualization.</article-title> <source><italic>Front. Hum. Neurosci.</italic></source> <volume>15</volume>:<fpage>675154</fpage>. <pub-id pub-id-type="doi">10.3389/fnhum.2021.675154</pub-id> <pub-id pub-id-type="pmid">34135744</pub-id></mixed-citation></ref>
<ref id="B13"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Mohammadi Foumani</surname> <given-names>N.</given-names></name> <name><surname>Mackellar</surname> <given-names>G.</given-names></name> <name><surname>Ghane</surname> <given-names>S.</given-names></name> <name><surname>Irtza</surname> <given-names>S.</given-names></name> <name><surname>Nguyen</surname> <given-names>N.</given-names></name> <name><surname>Salehi</surname> <given-names>M.</given-names></name></person-group> (<year>2024</year>). &#x201C;<article-title>EEG2Rep: Enhancing self-supervised EEG representation through informative masked inputs</article-title>,&#x201D; in <source><italic>Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</italic></source>, (<publisher-loc>New York, NY</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>5544</fpage>&#x2013;<lpage>5555</lpage>. <pub-id pub-id-type="doi">10.1145/3637528.3671600</pub-id></mixed-citation></ref>
<ref id="B14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Montazeri</surname> <given-names>S.</given-names></name> <name><surname>Nevalainen</surname> <given-names>P.</given-names></name> <name><surname>Mets&#x00E4;ranta</surname> <given-names>M.</given-names></name> <name><surname>Stevenson</surname> <given-names>N. J.</given-names></name> <name><surname>Vanhatalo</surname> <given-names>S.</given-names></name></person-group> (<year>2024</year>). <article-title>Clinical outcome prediction with an automated EEG trend, Brain State of the Newborn, after perinatal asphyxia.</article-title> <source><italic>Clin. Neurophysiol.</italic></source> <volume>162</volume> <fpage>68</fpage>&#x2013;<lpage>76</lpage>. <pub-id pub-id-type="doi">10.1016/J.CLINPH.2024.03.007</pub-id> <pub-id pub-id-type="pmid">38583406</pub-id></mixed-citation></ref>
<ref id="B15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pillay</surname> <given-names>K.</given-names></name> <name><surname>Dereymaeker</surname> <given-names>A.</given-names></name> <name><surname>Jansen</surname> <given-names>K.</given-names></name> <name><surname>Naulaers</surname> <given-names>G.</given-names></name> <name><surname>De Vos</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Applying a data-driven approach to quantify EEG maturational deviations in preterms with normal and abnormal neurodevelopmental outcomes.</article-title> <source><italic>Sci. Rep.</italic></source> <volume>10</volume>:<fpage>7288</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-020-64211-0</pub-id> <pub-id pub-id-type="pmid">32350387</pub-id></mixed-citation></ref>
<ref id="B16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Piryatinska</surname> <given-names>A.</given-names></name> <name><surname>Terdik</surname> <given-names>G.</given-names></name> <name><surname>Woyczynski</surname> <given-names>W. A.</given-names></name> <name><surname>Loparo</surname> <given-names>K. A.</given-names></name> <name><surname>Scher</surname> <given-names>M. S.</given-names></name> <name><surname>Zlotnik</surname> <given-names>A.</given-names></name></person-group> (<year>2009</year>). <article-title>Automated detection of neonate EEG sleep stages.</article-title> <source><italic>Comput. Methods Programs Biomed.</italic></source> <volume>95</volume> <fpage>31</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2009.01.006</pub-id> <pub-id pub-id-type="pmid">19233504</pub-id></mixed-citation></ref>
<ref id="B17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raeisi</surname> <given-names>K.</given-names></name> <name><surname>Khazaei</surname> <given-names>M.</given-names></name> <name><surname>Croce</surname> <given-names>P.</given-names></name> <name><surname>Tamburro</surname> <given-names>G.</given-names></name> <name><surname>Comani</surname> <given-names>S.</given-names></name> <name><surname>Zappasodi</surname> <given-names>F.</given-names></name></person-group> (<year>2022</year>). <article-title>A graph convolutional neural network for the automated detection of seizures in the neonatal EEG.</article-title> <source><italic>Comput. Methods Programs Biomed.</italic></source> <volume>222</volume>:<fpage>106950</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2022.106950</pub-id> <pub-id pub-id-type="pmid">35717740</pub-id></mixed-citation></ref>
<ref id="B18"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raurale</surname> <given-names>S. A.</given-names></name> <name><surname>Boylan</surname> <given-names>G. B.</given-names></name> <name><surname>Mathieson</surname> <given-names>S. R.</given-names></name> <name><surname>Marnane</surname> <given-names>W. P.</given-names></name> <name><surname>Lightbody</surname> <given-names>G.</given-names></name> <name><surname>O&#x2019;Toole</surname> <given-names>J. M.</given-names></name></person-group> (<year>2021</year>). <article-title>Grading hypoxic-ischemic encephalopathy in neonatal EEG with convolutional neural networks and quadratic time&#x2013;frequency distributions.</article-title> <source><italic>J. Neural Eng.</italic></source> <volume>18</volume>:<fpage>046007</fpage>. <pub-id pub-id-type="doi">10.1088/1741-2552/abe8ae</pub-id> <pub-id pub-id-type="pmid">33618337</pub-id></mixed-citation></ref>
<ref id="B19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shellhaas</surname> <given-names>R. A.</given-names></name> <name><surname>Burns</surname> <given-names>J. W.</given-names></name> <name><surname>Hassan</surname> <given-names>F.</given-names></name> <name><surname>Carlson</surname> <given-names>M. D.</given-names></name> <name><surname>Barks</surname> <given-names>J. D. E.</given-names></name> <name><surname>Chervin</surname> <given-names>R. D.</given-names></name></person-group> (<year>2017</year>). <article-title>Neonatal sleep-wake analyses predict 18-month neurodevelopmental outcomes.</article-title> <source><italic>Sleep</italic></source> <volume>40</volume>:<fpage>zsx144</fpage>. <pub-id pub-id-type="doi">10.1093/sleep/zsx144</pub-id> <pub-id pub-id-type="pmid">28958087</pub-id></mixed-citation></ref>
<ref id="B20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Stevenson</surname> <given-names>N. J.</given-names></name> <name><surname>Oberdorfer</surname> <given-names>L.</given-names></name> <name><surname>Tataranno</surname> <given-names>M.-L.</given-names></name> <name><surname>Breakspear</surname> <given-names>M.</given-names></name> <name><surname>Colditz</surname> <given-names>P. B.</given-names></name> <name><surname>Vries</surname> <given-names>L. S.</given-names></name><etal/></person-group> (<year>2020</year>). <article-title>Automated cot-side tracking of functional brain age in preterm infants.</article-title> <source><italic>Ann. Clin. Transl. Neurol.</italic></source> <volume>7</volume> <fpage>891</fpage>&#x2013;<lpage>902</lpage>. <pub-id pub-id-type="doi">10.1002/acn3.51043</pub-id> <pub-id pub-id-type="pmid">32368863</pub-id></mixed-citation></ref>
<ref id="B21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Temko</surname> <given-names>A.</given-names></name> <name><surname>Marnane</surname> <given-names>W.</given-names></name> <name><surname>Boylan</surname> <given-names>G.</given-names></name> <name><surname>Lightbody</surname> <given-names>G.</given-names></name></person-group> (<year>2015</year>). <article-title>Clinical implementation of a neonatal seizure detection algorithm.</article-title> <source><italic>Decis. Support Syst.</italic></source> <volume>70</volume> <fpage>86</fpage>&#x2013;<lpage>96</lpage>. <pub-id pub-id-type="doi">10.1016/j.dss.2014.12.006</pub-id> <pub-id pub-id-type="pmid">25892834</pub-id></mixed-citation></ref>
<ref id="B22"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Usman</surname> <given-names>F.</given-names></name> <name><surname>Zandvoort</surname> <given-names>C. S.</given-names></name> <name><surname>Robinson</surname> <given-names>S.</given-names></name> <name><surname>Peck</surname> <given-names>M.</given-names></name> <name><surname>Cobo</surname> <given-names>M. M.</given-names></name> <name><surname>Adjei</surname> <given-names>T.</given-names></name><etal/></person-group> (<year>2025</year>). <article-title>Brain activity as a biomarker for personalised caffeine treatment in premature neonates.</article-title> <source><italic>medRxiv</italic></source> [<comment>Preprint</comment>] <pub-id pub-id-type="doi">10.1101/2025.05.02.25326856</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/635313/overview">Layla Banihashemi</ext-link>, University of Pittsburgh, United States</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2628590/overview">Rabindra Khadka</ext-link>, Oslo Metropolitan University, Norway</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2712461/overview">Johannes Mader</ext-link>, Medical University of Vienna, Austria</p></fn>
</fn-group>
</back>
</article>