<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="systematic-review" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2026.1734591</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Systematic Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Deep learning for detecting early gastric cancer with white-light endoscopy: a systematic review and meta-analysis</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Liu</surname>
<given-names>Jixiang</given-names>
</name>
<xref ref-type="aff" rid="aff1"/>
<uri xlink:href="https://loop.frontiersin.org/people/2033861"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Danyan</given-names>
</name>
<xref ref-type="aff" rid="aff1"/>
<uri xlink:href="https://loop.frontiersin.org/people/723072"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Zhuo</surname>
<given-names>Yudi</given-names>
</name>
<xref ref-type="aff" rid="aff1"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Zhang</surname>
<given-names>Shengsheng</given-names>
</name>
<xref ref-type="aff" rid="aff1"/>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
</contrib>
</contrib-group>
<aff id="aff1"><institution>Department of Gastroenterology, Beijing Traditional Chinese Medicine Hospital, Capital Medical University</institution>, <city>Beijing</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Shengsheng Zhang, <email xlink:href="mailto:zhangshengsheng@bjzhongyi.com">zhangshengsheng@bjzhongyi.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-29">
<day>29</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>9</volume>
<elocation-id>1734591</elocation-id>
<history>
<date date-type="received">
<day>29</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>17</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Liu, Li, Zhuo and Zhang.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Liu, Li, Zhuo and Zhang</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-29">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Background and objectives</title>
<p>The aim of this study is to evaluate the performance of DL algorithms in diagnosing early gastric cancer (EGC) using white light endoscopic images.</p>
</sec>
<sec>
<title>Methods</title>
<p>A systematic literature search was conducted in PubMed, Embase, Cochrane Library, and Web of Science up to July 25, 2025. Sensitivity and specificity were pooled for internal and external validation sets. The comparison between DL algorithms and expert endoscopists was performed using paired forest plots. Meta-regression was used to identify sources of heterogeneity.</p>
</sec>
<sec>
<title>Results</title>
<p>In the internal validation, 15 studies comprising 37,037 images (range: 433&#x2013;9,650) were included. Pooled sensitivity and specificity were 0.91 (95% CI: 0.82&#x2013;0.95) and 0.93 (95% CI: 0.87&#x2013;0.97), respectively. Meta-regression showed that heterogeneity in sensitivity and specificity was significantly associated with training dataset size. For external validation, 4 studies with 3,579 images (range: 200&#x2013;1,514) were included, yielding pooled sensitivity and specificity of 0.82 (95% CI: 0.61&#x2013;0.93) and 0.83 (95% CI: 0.74&#x2013;0.90), respectively. No significant difference was observed between deep learning models and expert endoscopists in diagnostic sensitivity and specificity.</p>
</sec>
<sec>
<title>Conclusion</title>
<p>Deep learning algorithms exhibit high diagnostic performance in detecting early gastric cancer using white-light endoscopy. The diagnostic accuracy of DL models is comparable to that of expert endoscopists, supporting their potential role as a clinical decision-support tool.</p>
</sec>
<sec>
<title>Systematic review registration</title>
<p><ext-link ext-link-type="uri" xlink:href="https://www.crd.york.ac.uk/PROSPERO/view/CRD420251112418">https://www.crd.york.ac.uk/PROSPERO/view/CRD420251112418</ext-link>, identifier CRD420251112418.</p>
</sec>
</abstract>
<kwd-group>
<kwd>artificial intelligence</kwd>
<kwd>deep learning</kwd>
<kwd>detection</kwd>
<kwd>early gastric cancer</kwd>
<kwd>endoscopy</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This research was supported by the National Administration of Traditional Chinese Medicine&#x2019;s &#x201C;Hundred Thousand Ten Thousand&#x201D; Talent Inheritance and Innovation Project (Qi Huang Scholars) National Leading Talent Support Plan for Traditional Chinese Medicine (No. [2021]203 of the Ministry of Traditional Chinese Medicine&#x2019;s Teacher Education and Personnel Work).</funding-statement>
</funding-group>
<counts>
<fig-count count="8"/>
<table-count count="3"/>
<equation-count count="0"/>
<ref-count count="62"/>
<page-count count="14"/>
<word-count count="8740"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Medicine and Public Health</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<title>Introduction</title>
<p>Gastric cancer (GC) is a major global health burden, ranking fifth in incidence and fourth in cancer-related mortality worldwide (<xref ref-type="bibr" rid="ref43">Sung et al., 2021</xref>). Early gastric cancer (EGC) is defined as adenocarcinoma that infiltrates the mucosa or submucosa of the stomach with or without lymph node metastases (T1, any N), which is associated with a favorable prognosis and a five-year survival rate of approximately 95% (<xref ref-type="bibr" rid="ref37">&#x00D6;hman et al., 1980</xref>; <xref ref-type="bibr" rid="ref18">GASTRIC (Global Advanced/Adjuvant Stomach Tumor Research International Collaboration) Group et al., 2013</xref>; <xref ref-type="bibr" rid="ref26">Katai et al., 2018</xref>; <xref ref-type="bibr" rid="ref55">Yang et al., 2021</xref>). Consequently, early detection of EGC is critical for improving patient clinical outcomes.</p>
<p>Upper gastrointestinal endoscopy has been established as the gold standard for the diagnosis of EGC (<xref ref-type="bibr" rid="ref32">Machlowska et al., 2020</xref>). Among its various imaging modalities, white-light endoscopy remains the preferred technique in routine clinical practice due to its widespread availability and ease of use (<xref ref-type="bibr" rid="ref35">Nagula et al., 2024</xref>). Evidence from South Korea has demonstrated that screening upper gastrointestinal endoscopy has significantly increased the detection of EGC and reduced mortality by approximately 50% (OR&#x202F;=&#x202F;0.53, 95% CI: 0.51&#x2013;0.56) (<xref ref-type="bibr" rid="ref24">Jun et al., 2017</xref>; <xref ref-type="bibr" rid="ref2">Arnold et al., 2020</xref>). However, EGC lesions often present with subtle mucosal changes, such as microsurface architectural disruption and color irregularities, making their detection challenging under standard white-light endoscopy during routine screening (<xref ref-type="bibr" rid="ref57">Zhang et al., 2011</xref>; <xref ref-type="bibr" rid="ref31">Liu et al., 2023</xref>). As a result, the accuracy of EGC detection is highly dependent on endoscopist expertise, resulting in variability in diagnostic performance. Indeed, previous studies have shown that senior endoscopists with more than 10&#x202F;years of experience achieved significantly higher diagnostic sensitivity in detecting EGC compared to junior endoscopists with only 2&#x2013;3&#x202F;years of training (<xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref56">Yuan et al., 2022</xref>).</p>
<p>To address the aforementioned challenges, deep learning (DL)-based artificial intelligence (AI) has been increasingly applied to medical imaging, showing substantial promise in improving diagnostic sensitivity and specificity (<xref ref-type="bibr" rid="ref13">Esteva et al., 2019</xref>; <xref ref-type="bibr" rid="ref17">Gandhi et al., 2025c</xref>). Compared to traditional machine learning, DL algorithms possess several advantages. First, they possess the ability to perform feature self-learning from medical image datasets, eliminating the need for manual feature extraction and avoiding potential performance degradation caused by inaccurate or inconsistent segmentation. Second, they can be trained in an end-to-end manner, mapping raw images to diagnostic outputs while jointly optimizing all components of the network (<xref ref-type="bibr" rid="ref4">Baldominos et al., 2019</xref>; <xref ref-type="bibr" rid="ref53">Wang et al., 2019b</xref>; <xref ref-type="bibr" rid="ref61">Zhou Z. et al., 2023</xref>). In recent years, DL algorithms have been widely investigated in the field of pathological image analysis. Numerous studies have consistently demonstrated high diagnostic accuracy in tumor detection across multiple cancer types, including breast, lung, and colorectal cancers, as well as glioma (<xref ref-type="bibr" rid="ref52">Wang et al., 2019a</xref>; <xref ref-type="bibr" rid="ref22">Im et al., 2021</xref>; <xref ref-type="bibr" rid="ref29">Li et al., 2022</xref>, <xref ref-type="bibr" rid="ref30">2025</xref>; <xref ref-type="bibr" rid="ref48">Thalakottor et al., 2023</xref>). In the diagnosis of EGC using endoscopic images, a previous meta-analysis found that conventional AI achieved a sensitivity of 86% and a specificity of 90%, demonstrating diagnostic accuracy comparable to that of experienced endoscopists (<xref ref-type="bibr" rid="ref8">Chen P.-C. et al., 2022</xref>). However, the aforementioned meta-analysis included a limited number of studies and did not specifically evaluate the performance of deep learning algorithms in detecting EGC under white-light endoscopy.</p>
<p>Therefore, this systematic review synthesizes the latest developments and analyzes the diagnostic performance of DL algorithms on white-light endoscopy image datasets in EGC diagnosis. Meanwhile, our study further compared the diagnostic performance for EGC between DL algorithms and expert endoscopists. The findings will provide evidence-based support for the clinical translation of DL algorithms in upper gastrointestinal endoscopy for EGC.</p>
</sec>
<sec sec-type="methods" id="sec2">
<title>Methods</title>
<p>This meta-analysis was conducted in full compliance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses of Diagnostic Test Accuracy (PRISMA-DTA) guidelines (<xref rid="SM1" ref-type="supplementary-material">Supplementary Table 1</xref>) (<xref ref-type="bibr" rid="ref33">Mdf et al., 2018</xref>). Additionally, the study protocol has been registered in the PROSPERO database (CRD420251112418).</p>
<sec id="sec3">
<title>Search strategy</title>
<p>We conducted a systematic literature search using the PubMed, Embase, Cochrane, and Web of Science databases, with the search completed on July 25, 2025. The search strategy involved three groups of keywords: AI-related terms (e.g., artificial intelligence, deep learning), examination-related terms (e.g., endoscopes, gastroscopy), and disease-related terms (e.g., stomach neoplasms, gastric cancer). Both free-text keywords and Medical Subject Headings (MeSH) terms were used to ensure precision. Detailed search strategies are available in <xref rid="SM1" ref-type="supplementary-material">Supplementary Table 2</xref>. Additionally, the references of included studies were reviewed to identify additional relevant literature.</p>
</sec>
<sec id="sec4">
<title>Inclusion and exclusion criteria</title>
<p>The studies were systematically selected according to the PITROS framework to ensure methodological clarity and reporting transparency. Participants (P): The participants in this study are patients diagnosed with EGC based on pathological examination. Index test (I): The index test involved the application of DL algorithms to analyze white-light endoscopic images for the automated detection of EGC. Target condition (T): The target condition was the presence of EGC. Diagnosis was based on histopathology, with patients categorized as EGC-positive or EGC-negative accordingly. Outcomes (O): The primary outcomes include sensitivity and specificity for the diagnosis of EGC. Secondary outcomes included a comparative assessment of sensitivity and specificity between DL algorithms and expert endoscopists in the diagnosis of EGC. Setting (S): The study setting includes retrospective or prospective data sources, covering public databases or local hospitals.</p>
<p>Exclusion criteria included studies on animals, non-original articles (e.g., reviews, case reports, meta-analyses, and letters to editors), and non-English publications due to accessibility issues. Furthermore, studies using conventional AI approaches that are unrelated to deep learning algorithms, such as classic machine learning techniques (e.g., support vector machines, logistic regression, and random forests), were excluded. In addition, studies utilizing endoscopic techniques other than white-light endoscopy, such as narrow-band imaging (NBI) or magnifying endoscopy, were excluded.</p>
</sec>
<sec id="sec5">
<title>Quality assessment</title>
<p>To ensure a rigorous evaluation of the methodological quality of the included studies, we utilized the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool to assess the risk of bias in predictive modeling (<xref ref-type="bibr" rid="ref54">Whiting et al., 2011</xref>). The quality evaluation criteria included four domains: patient selection, index test, reference standard, and flow and timing.</p>
</sec>
<sec id="sec6">
<title>Data extraction</title>
<p>Two independent reviewers (JXL and YDZ) screened the titles and abstracts of the remaining articles to identify potentially eligible studies, with a third reviewer (DYL) acting as an arbitrator to resolve any discrepancies. Extracted data were grouped into three categories: (1) study characteristics (first author, publication year, study design, country of origin, number of centers, diagnostic definition for EGC, and diagnostic algorithm); (2) image dataset composition (number of images in training, internal validation, external validation, DL and endoscopists comparative test set, and tile size); and (3) diagnostic performance outcomes (raw numbers of true positives, false positives, true negatives, and false negatives). For studies lacking information necessary for meta-analysis, we contacted the corresponding authors by email to request the missing data.</p>
</sec>
<sec id="sec7">
<title>Outcome measures</title>
<p>The primary outcome measures were sensitivity and specificity for internal and external validation sets. Sensitivity, also known as recall or the true positive rate, measures the probability of correctly identifying true EGC cases and is calculated as true positive (TP)/(TP&#x202F;+&#x202F;false negative (FN)). Specificity, or the true negative rate, reflects the probability of correctly identifying non-EGC cases and is calculated as true negative (TN)/(TN&#x202F;+&#x202F;false positive (FP)). For studies comparing the performance of endoscopists and DL algorithms in diagnosing EGC, the diagnostic data of expert endoscopists and DL algorithms will be extracted and entered.</p>
</sec>
<sec id="sec8">
<title>Statistical analysis</title>
<p>This study employed a bivariate random-effects model to perform the meta-analysis, which jointly pools sensitivity and specificity while accounting for their inherent negative correlation. This model was used to assess the diagnostic performance of deep learning for EGC detection on white-light endoscopy images and to generate a hierarchical summary receiver operating characteristic (HSROC) curve. Sensitivity and specificity were pooled separately for internal and external validation sets. Forest plots visually presented the study-level and pooled estimates, while the SROC curve provided an overall summary with a 95% confidence region and a 95% prediction region. The between-study variance for logit-transformed sensitivity and specificity was quantified using the tau<sup>2</sup> (<italic>&#x03C4;</italic><sup>2</sup>) statistic.</p>
<p>Heterogeneity across studies was evaluated using Higgins&#x2019; <italic>I</italic><sup>2</sup> statistic, with <italic>I</italic><sup>2</sup> values of 25, 50, and 75% indicating low, moderate, and high heterogeneity, respectively (<xref ref-type="bibr" rid="ref21">Huedo-Medina et al., 2006</xref>). Meta-regression analyses were conducted to identify sources of significant heterogeneity (<italic>I</italic><sup>2</sup>&#x202F;&#x003E;&#x202F;50%) (<xref ref-type="bibr" rid="ref51">van Houwelingen et al., 2002</xref>). Meta-regression variables included the number of centers (single or multiple), size of the training dataset (large-scale public datasets or small-scale institutional datasets), validation method (with or without cross-validation), tile size (&#x2264;448&#x202F;&#x00D7;&#x202F;448 or &#x003E;448&#x202F;&#x00D7;&#x202F;448), and risk of bias in patient selection (high risk or low risk). Potential publication bias was assessed using Deeks&#x2019; funnel plot asymmetry test. Furthermore, for comparative assessment of diagnostic performance, sensitivity and specificity were independently pooled for deep learning models and expert endoscopists. Paired forest plots were generated to facilitate direct, visual comparison of sensitivity and specificity across the two groups. Statistical analyses were performed using the Midas package in Stata (version 15.1) and the meta package in R, while risk of bias assessment was conducted with RevMan 5.4 from the Cochrane Collaboration. All statistical tests were two-sided, with <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05 considered statistically significant, and results were reported with 95% confidence intervals.</p>
</sec>
</sec>
<sec sec-type="results" id="sec9">
<title>Results</title>
<sec id="sec10">
<title>Study selection</title>
<p>The initial database search identified 721 potentially relevant articles. After removing 138 duplicate records, 583 unique articles underwent preliminary screening. Application of the predefined inclusion criteria led to the exclusion of 521 articles. Subsequently, a comprehensive full-text assessment resulted in the further exclusion of 47 studies due to insufficient or incomplete diagnostic data (TP, FP, FN, TN) or the use of non-white-light endoscopy techniques. Ultimately, 15 studies meeting the eligibility criteria were included in the meta-analysis to evaluate the diagnostic performance of DL algorithms (<xref ref-type="bibr" rid="ref41">Sakai et al., 2018</xref>; <xref ref-type="bibr" rid="ref10">Cho et al., 2019</xref>; <xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref60">Zhang et al., 2021</xref>; <xref ref-type="bibr" rid="ref47">Teramoto et al., 2022</xref>; <xref ref-type="bibr" rid="ref56">Yuan et al., 2022</xref>; <xref ref-type="bibr" rid="ref45">Takemoto et al., 2023</xref>; <xref ref-type="bibr" rid="ref11">Dong et al., 2023</xref>; <xref ref-type="bibr" rid="ref58">Zhang et al., 2023</xref>; <xref ref-type="bibr" rid="ref62">Zhou B. et al., 2023</xref>; <xref ref-type="bibr" rid="ref7">Chang et al., 2024</xref>; <xref ref-type="bibr" rid="ref20">Gong et al., 2024</xref>; <xref ref-type="bibr" rid="ref59">Zhang et al., 2024</xref>; <xref ref-type="bibr" rid="ref50">Ul Haq et al., 2024</xref>; <xref ref-type="bibr" rid="ref14">Feng et al., 2025</xref>). The literature selection process was summarized using a PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram, presented in <xref ref-type="fig" rid="fig1">Figure 1</xref>.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>PRISMA flow diagram illustrating the study selection process.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart depicting the identification and selection process for a meta-analysis. Initially, 721 records were identified from databases: PubMed (152), Embase (269), Web of Science (296), and Cochrane (4). After removing duplicates, 583 records remained. 521 articles were excluded due to irrelevant content or mismatched criteria, leaving 62 full-text articles for eligibility assessment. Subsequently, 47 articles were excluded due to unavailable data or non-white light endoscopy, resulting in 15 studies included in the meta-analysis.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec11">
<title>Study description and quality assessment</title>
<p>For internal validation, 15 studies involving 37,037 images (range: 433&#x2013;9,650) were included (<xref ref-type="bibr" rid="ref41">Sakai et al., 2018</xref>; <xref ref-type="bibr" rid="ref10">Cho et al., 2019</xref>; <xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref60">Zhang et al., 2021</xref>; <xref ref-type="bibr" rid="ref47">Teramoto et al., 2022</xref>; <xref ref-type="bibr" rid="ref56">Yuan et al., 2022</xref>; <xref ref-type="bibr" rid="ref45">Takemoto et al., 2023</xref>; <xref ref-type="bibr" rid="ref11">Dong et al., 2023</xref>; <xref ref-type="bibr" rid="ref58">Zhang et al., 2023</xref>; <xref ref-type="bibr" rid="ref62">Zhou B. et al., 2023</xref>; <xref ref-type="bibr" rid="ref7">Chang et al., 2024</xref>; <xref ref-type="bibr" rid="ref20">Gong et al., 2024</xref>; <xref ref-type="bibr" rid="ref59">Zhang et al., 2024</xref>; <xref ref-type="bibr" rid="ref50">Ul Haq et al., 2024</xref>; <xref ref-type="bibr" rid="ref14">Feng et al., 2025</xref>); for external validation, 4 studies with 3,579 images (range: 200&#x2013;1,514) were included (<xref ref-type="bibr" rid="ref10">Cho et al., 2019</xref>; <xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref55">Yang et al., 2021</xref>; <xref ref-type="bibr" rid="ref11">Dong et al., 2023</xref>; <xref ref-type="bibr" rid="ref20">Gong et al., 2024</xref>). The studies were published between 2018 and 2025. Regarding study design, 14 studies were retrospective, whereas only one study was prospective in its external validation cohort (<xref ref-type="bibr" rid="ref10">Cho et al., 2019</xref>). Only two studies utilized large-scale public datasets for training, while the remaining studies were trained using small-scale institutional datasets. All DL models employed in the studies were based on convolutional neural networks (CNNs). Study characteristics and diagnostic performance in internal and external validation are summarized in <xref ref-type="table" rid="tab1">Tables 1</xref>, <xref ref-type="table" rid="tab2">2</xref> and <xref rid="SM1" ref-type="supplementary-material">Supplementary Table 3</xref>, respectively. Notably, five studies included comparisons between DL algorithms and endoscopists in diagnostic performance (<xref ref-type="bibr" rid="ref10">Cho et al., 2019</xref>; <xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref60">Zhang et al., 2021</xref>; <xref ref-type="bibr" rid="ref56">Yuan et al., 2022</xref>; <xref ref-type="bibr" rid="ref45">Takemoto et al., 2023</xref>). The diagnostic performance of DL algorithms and endoscopists is presented in <xref rid="SM1" ref-type="supplementary-material">Supplementary Table 4</xref>.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Characteristics of the included studies.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Author</th>
<th align="center" valign="top" rowspan="2">Year</th>
<th align="left" valign="top" rowspan="2">Country</th>
<th align="left" valign="top" rowspan="2">Center</th>
<th align="center" valign="top" rowspan="2">Tile size</th>
<th align="left" valign="top" rowspan="2">Specific model</th>
<th align="left" valign="top" rowspan="2">Model type</th>
<th align="center" valign="top" colspan="2">Number of images (early gastric cancer vs. control)</th>
<th align="left" valign="top" rowspan="2">Endoscopist comparison</th>
</tr>
<tr>
<th align="center" valign="top">Training</th>
<th align="left" valign="top">Validation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Sakai et al.</td>
<td align="center" valign="middle">2018</td>
<td align="left" valign="middle">Japan</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">224 &#x002A; 224</td>
<td align="left" valign="middle">GoogLeNet</td>
<td align="left" valign="middle">CNN<xref ref-type="table-fn" rid="tfn1"><sup>a</sup></xref></td>
<td align="center" valign="middle">9,587 vs. 9,800</td>
<td align="left" valign="middle">NR<xref ref-type="table-fn" rid="tfn2"><sup>b</sup></xref></td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Cho et al.</td>
<td align="center" valign="middle">2019</td>
<td align="left" valign="middle">Korea</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">1,280 &#x002A; 640</td>
<td align="left" valign="middle">Inception-Resnet-v2</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">919 vs. 3,286</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">YES</td>
</tr>
<tr>
<td align="left" valign="middle">Tang et al.</td>
<td align="center" valign="middle">2020</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">416 &#x002A; 416</td>
<td align="left" valign="middle">Darknet-53</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">26,172 vs. 9,651</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">YES</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2021</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">NR</td>
<td align="left" valign="middle">ResNet34</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">6,139 vs. 15,078</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">YES</td>
</tr>
<tr>
<td align="left" valign="middle">Zhou et al.</td>
<td align="center" valign="middle">2022</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">512 &#x002A; 512</td>
<td align="left" valign="middle">EfficientDet-D2</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">1,390 vs. 2,232</td>
<td align="left" valign="middle">347 vs. 558</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Yuan et al.</td>
<td align="center" valign="middle">2022</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">640 &#x002A; 640</td>
<td align="left" valign="middle">YOLO</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">2,015 vs. 27,794</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">YES</td>
</tr>
<tr>
<td align="left" valign="middle">Teramoto et al.</td>
<td align="center" valign="middle">2022</td>
<td align="left" valign="middle">Japan</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">512 &#x002A; 512</td>
<td align="left" valign="middle">DenseNet-121</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">Imagenet database</td>
<td align="left" valign="middle">5-fold cross-validation</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Takemoto et al.</td>
<td align="center" valign="middle">2023</td>
<td align="left" valign="middle">Japan</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">224 &#x002A; 224</td>
<td align="left" valign="middle">GoogLeNet</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">534,926 vs. 593,874</td>
<td align="left" valign="middle">10-fold cross-validation</td>
<td align="left" valign="middle">YES</td>
</tr>
<tr>
<td align="left" valign="middle">Gong et al.</td>
<td align="center" valign="middle">2023</td>
<td align="left" valign="middle">Korea</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">512 &#x002A; 431</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">1,766 vs. 13,193</td>
<td align="left" valign="middle">221 vs. 1,650</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Dong et al.</td>
<td align="center" valign="middle">2023</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">NR</td>
<td align="left" valign="middle">YOLO-v3 and Resnet-50</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">1,933 vs. 1,679</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2023</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">NR</td>
<td align="left" valign="middle">Resnet50</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">2,070 vs. 7,966</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2024</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">1,080 &#x002A; 1,080</td>
<td align="left" valign="middle">Faster RCNN</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">Private database and public Kvasir-SEG dataset</td>
<td align="left" valign="middle">5-fold cross-validation</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Chang et al.</td>
<td align="center" valign="middle">2024</td>
<td align="left" valign="middle">Korea</td>
<td align="left" valign="middle">Multiple</td>
<td align="center" valign="middle">NR</td>
<td align="left" valign="middle">YOLO-v5 and EfficientNetB0</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">3,920 vs. 5,026</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Haq et al.</td>
<td align="center" valign="middle">2024</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">224&#x002A;224</td>
<td align="left" valign="middle">Faster RCNN</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">NR</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">No</td>
</tr>
<tr>
<td align="left" valign="middle">Feng et al.</td>
<td align="center" valign="middle">2025</td>
<td align="left" valign="middle">China</td>
<td align="left" valign="middle">Single</td>
<td align="center" valign="middle">448&#x002A;448</td>
<td align="left" valign="middle">ResNet18</td>
<td align="left" valign="middle">CNN</td>
<td align="center" valign="middle">3,400 vs. 8,400</td>
<td align="left" valign="middle">NR</td>
<td align="left" valign="middle">No</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="tfn1">
<label>a</label>
<p>CNN: convolutional neural network.</p>
</fn>
<fn id="tfn2">
<label>b</label>
<p>NR: Not Reported.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Diagnostic performance of the included studies.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top" rowspan="2">Author</th>
<th align="center" valign="top" rowspan="2">Year</th>
<th align="center" valign="top" colspan="4">Interval validation sets</th>
<th align="center" valign="top" colspan="4">External validation sets</th>
</tr>
<tr>
<th align="center" valign="top">TP</th>
<th align="center" valign="top">FP</th>
<th align="center" valign="top">TN</th>
<th align="center" valign="top">FN</th>
<th align="center" valign="top">TP</th>
<th align="center" valign="top">FP</th>
<th align="center" valign="top">TN</th>
<th align="center" valign="top">FN</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Sakai et al.</td>
<td align="center" valign="middle">2018</td>
<td align="center" valign="middle">3,723</td>
<td align="center" valign="middle">262</td>
<td align="center" valign="middle">4,735</td>
<td align="center" valign="middle">930</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Cho et al.</td>
<td align="center" valign="middle">2019</td>
<td align="center" valign="middle">97</td>
<td align="center" valign="middle">88</td>
<td align="center" valign="middle">559</td>
<td align="center" valign="middle">68</td>
<td align="center" valign="middle">13</td>
<td align="center" valign="middle">33</td>
<td align="center" valign="middle">136</td>
<td align="center" valign="middle">18</td>
</tr>
<tr>
<td align="left" valign="middle">Tang et al.</td>
<td align="center" valign="middle">2020</td>
<td align="center" valign="middle">3,967</td>
<td align="center" valign="middle">961</td>
<td align="center" valign="middle">4,303</td>
<td align="center" valign="middle">186</td>
<td align="center" valign="middle">678</td>
<td align="center" valign="middle">98</td>
<td align="center" valign="middle">659</td>
<td align="center" valign="middle">79</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2021</td>
<td align="center" valign="middle">92</td>
<td align="center" valign="middle">74</td>
<td align="center" valign="middle">766</td>
<td align="center" valign="middle">158</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Zhou et al.</td>
<td align="center" valign="middle">2022</td>
<td align="center" valign="middle">376</td>
<td align="center" valign="middle">76</td>
<td align="center" valign="middle">722</td>
<td align="center" valign="middle">69</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Yuan et al.</td>
<td align="center" valign="middle">2022</td>
<td align="center" valign="middle">177</td>
<td align="center" valign="middle">146</td>
<td align="center" valign="middle">1,124</td>
<td align="center" valign="middle">9</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Teramoto et al.</td>
<td align="center" valign="middle">2022</td>
<td align="center" valign="middle">531</td>
<td align="center" valign="middle">0</td>
<td align="center" valign="middle">1,845</td>
<td align="center" valign="middle">1</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Takemoto et al.</td>
<td align="center" valign="middle">2023</td>
<td align="center" valign="middle">387</td>
<td align="center" valign="middle">89</td>
<td align="center" valign="middle">307</td>
<td align="center" valign="middle">75</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Gong et al.</td>
<td align="center" valign="middle">2023</td>
<td align="center" valign="middle">164</td>
<td align="center" valign="middle">48</td>
<td align="center" valign="middle">1,602</td>
<td align="center" valign="middle">57</td>
<td align="center" valign="middle">165</td>
<td align="center" valign="middle">119</td>
<td align="center" valign="middle">1,104</td>
<td align="center" valign="middle">39</td>
</tr>
<tr>
<td align="left" valign="middle">Dong et al.</td>
<td align="center" valign="middle">2023</td>
<td align="center" valign="middle">104</td>
<td align="center" valign="middle">74</td>
<td align="center" valign="middle">244</td>
<td align="center" valign="middle">11</td>
<td align="center" valign="middle">117</td>
<td align="center" valign="middle">101</td>
<td align="center" valign="middle">211</td>
<td align="center" valign="middle">9</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2023</td>
<td align="center" valign="middle">365</td>
<td align="center" valign="middle">241</td>
<td align="center" valign="middle">1,373</td>
<td align="center" valign="middle">40</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Zhang et al.</td>
<td align="center" valign="middle">2024</td>
<td align="center" valign="middle">263</td>
<td align="center" valign="middle">20</td>
<td align="center" valign="middle">247</td>
<td align="center" valign="middle">13</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Chang et al.</td>
<td align="center" valign="middle">2024</td>
<td align="center" valign="middle">451</td>
<td align="center" valign="middle">96</td>
<td align="center" valign="middle">1,665</td>
<td align="center" valign="middle">26</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Haq et al.</td>
<td align="center" valign="middle">2024</td>
<td align="center" valign="middle">865</td>
<td align="center" valign="middle">21</td>
<td align="center" valign="middle">829</td>
<td align="center" valign="middle">26</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
<tr>
<td align="left" valign="middle">Feng et al.</td>
<td align="center" valign="middle">2025</td>
<td align="center" valign="middle">564</td>
<td align="center" valign="middle">68</td>
<td align="center" valign="middle">617</td>
<td align="center" valign="middle">40</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
<td align="center" valign="middle">NR</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>TP, true positive; TN, true negative; FP, false positive; FN, false negative; NR, Not Reported.</p>
</table-wrap-foot>
</table-wrap>
<p>The risk of bias, assessed using the revised QUADAS-2 tool, is illustrated in <xref ref-type="fig" rid="fig2">Figure 2</xref>. In the patient selection domain, four studies were classified as &#x201C;high&#x201D; due to insufficient reporting on patient recruitment (e.g., whether enrollment was conducted consecutively). All studies were deemed to have a low risk of bias in the index test, reference standard, and flow and timing domains.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Risk of bias and applicability concerns in the included studies, assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Chart depicting the risk of bias and applicability concerns across various studies. Evaluated areas include patient selection, index test, reference standard, and flow and timing. Symbols indicate low (green circle), unclear (yellow circle), and high (red circle) risk levels. Most studies show a low risk of bias and applicability concerns, with occasional unclear or high-risk assessments.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec12">
<title>Diagnostic performance of deep learning algorithms in the internal validation set for early gastric cancer detection</title>
<p>For the internal validation dataset, DL algorithms based on white endoscopy images achieved a sensitivity of 0.91 (95% CI: 0.82&#x2013;0.95) and a specificity of 0.93 (95% CI: 0.87&#x2013;0.97) in detecting EGC patients (<xref ref-type="fig" rid="fig3">Figure 3</xref>). The area under curve (AUC) was 0.97 (95% CI: 0.95&#x2013;0.98) (<xref ref-type="fig" rid="fig4">Figure 4a</xref>). With a pre-test probability of 36%, representing the average incidence rate across all studies included in the internal validation dataset, the Fagan nomogram demonstrated a positive likelihood ratio of 88% and a negative likelihood ratio of 5% (<xref ref-type="fig" rid="fig5">Figure 5a</xref>).</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Forest plot of sensitivity and specificity of deep learning algorithms for detecting early gastric cancer (EGC) in the internal validation set. Squares represent individual study estimates, with horizontal lines indicating 95% confidence intervals; the diamond denotes the pooled estimate.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Forest plots showing sensitivity and specificity with 95% confidence intervals for various studies. Sensitivity is plotted on the left and specificity on the right, both ranging from approximately 0.3 to 1. Combined values are shown at the bottom with respective metrics and heterogeneity statistics. Each study is represented by a square, and aggregate results are indicated by a diamond.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Summary receiver operating characteristic (SROC) curves of deep learning algorithms for detecting early gastric cancer (EGC) in the internal <bold>(a)</bold> and external <bold>(b)</bold> validation sets.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two ROC curve plots labeled &#x201C;a&#x201D; and &#x201C;b&#x201D; compare sensitivity and specificity. Both plots display observed data as blue circles and a summary operating point as a red diamond. Plot &#x201C;a&#x201D; shows a sensitivity of 0.91 and a specificity of 0.93 with an area under the curve (AUC) of 0.97. Plot &#x201C;b&#x201D; shows a sensitivity of 0.82 and a specificity of 0.83 with an AUC of 0.89. Solid, dashed, and dotted lines represent the SRCC curve, 95 percent confidence contour, and 95 percent prediction contour, respectively.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Fagan&#x2019;s nomogram illustrating the clinical utility of deep learning algorithms for detecting early gastric cancer (EGC) in the internal <bold>(a)</bold> and external <bold>(b)</bold> validation sets.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two probability nomograms labeled a and b, showing likelihood ratios between pre-test and post-test probabilities. Nomogram a has a prior probability of thirty-six percent, likelihood ratio positive thirteen, positive post-test probability eighty-eight percent, likelihood ratio negative zero point one, negative post-test probability five percent. Nomogram b has a likelihood ratio positive five, positive post-test probability seventy-three percent, likelihood ratio negative zero point two two, negative post-test probability eleven percent. Each nomogram features intersecting red and blue lines representing positive and negative likelihood paths.</alt-text>
</graphic>
</fig>
<p>High heterogeneity was observed in both sensitivity (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;99.33%, <italic>&#x03C4;</italic><sup>2</sup>&#x202F;=&#x202F;1.89) and specificity (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;99.13%, <italic>&#x03C4;</italic><sup>2</sup>&#x202F;=&#x202F;2.46) within the internal validation dataset. Meta-regression analysis revealed that heterogeneity in both sensitivity and specificity was significantly associated with the size of the training dataset (large-scale public datasets vs. small-scale institutional datasets, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.05) and validation method (cross-validation vs. without cross-validation, <italic>p</italic>&#x202F;&#x2264;&#x202F;0.05) (<xref ref-type="table" rid="tab3">Table 3</xref>). Level-one out sensitivity analysis did not identify any influential studies or potential sources of heterogeneity (<xref rid="SM1" ref-type="supplementary-material">Supplementary Table 5</xref>). In addition, after excluding studies with a high risk of bias, the sensitivity was 0.86 (95% CI: 0.72&#x2013;0.94) and the specificity was 0.90 (95% CI: 0.85&#x2013;0.93), yielding a summary AUC of 0.94 (95% CI: 0.92&#x2013;0.96).</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Meta-regression analysis of diagnostic performance of deep learning models for early gastric cancer (EGC) in internal validation cohorts.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Subgroup</th>
<th align="center" valign="top">Studies, <italic>n</italic></th>
<th align="center" valign="top">Sensitivity (95%CI)</th>
<th align="center" valign="top">Meta-regression <italic>p</italic>-value</th>
<th align="center" valign="top">Specificity (95%CI)</th>
<th align="center" valign="top">Meta-regression <italic>p</italic>-value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Center</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.96</td>
<td/>
<td align="char" valign="top" char=".">1.00</td>
</tr>
<tr>
<td align="left" valign="top">Single-center</td>
<td align="center" valign="top">9</td>
<td align="char" valign="top" char="(">0.92 (0.85&#x2013;0.99)</td>
<td/>
<td align="char" valign="top" char="(">0.95 (0.91&#x2013;0.99)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Multi-center</td>
<td align="center" valign="top">6</td>
<td align="char" valign="top" char="(">0.88 (0.75&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.89 (0.79&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Training dataset</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.01</td>
<td/>
<td align="char" valign="top" char=".">0.00</td>
</tr>
<tr>
<td align="left" valign="top">Large-scale public datasets</td>
<td align="center" valign="top">2</td>
<td align="char" valign="top" char="(">0.99 (0.97&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.99 (0.98&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Small-scale institutional datasets</td>
<td align="center" valign="top">13</td>
<td align="char" valign="top" char="(">0.87 (0.80&#x2013;0.95)</td>
<td/>
<td align="char" valign="top" char="(">0.91 (0.85&#x2013;0.96)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Validation method</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.05</td>
<td/>
<td align="char" valign="top" char=".">0.09</td>
</tr>
<tr>
<td align="left" valign="top">Cross-validation</td>
<td align="center" valign="top">3</td>
<td align="char" valign="top" char="(">0.97 (0.93&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.98 (0.94&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Without cross-validation</td>
<td align="center" valign="top">12</td>
<td align="char" valign="top" char="(">0.88 (0.79&#x2013;0.96)</td>
<td/>
<td align="char" valign="top" char="(">0.91 (0.85&#x2013;0.97)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Tile size</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.84</td>
<td/>
<td align="char" valign="top" char=".">0.33</td>
</tr>
<tr>
<td align="left" valign="top">&#x2264;448&#x002A;448</td>
<td align="center" valign="top">5</td>
<td align="char" valign="top" char="(">0.92 (0.84&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.96 (0.92&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">&#x003E;448&#x002A;448</td>
<td align="center" valign="top">6</td>
<td align="char" valign="top" char="(">0.92 (0.83&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.91 (0.79&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Risk of bias in patient selection</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.87</td>
<td/>
<td align="char" valign="top" char=".">0.99</td>
</tr>
<tr>
<td align="left" valign="top">High</td>
<td align="center" valign="top">4</td>
<td align="char" valign="top" char="(">0.87 (0.71&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.90 (0.78&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Unclear or low</td>
<td align="center" valign="top">11</td>
<td align="char" valign="top" char="(">0.92 (0.85&#x2013;0.98)</td>
<td/>
<td align="char" valign="top" char="(">0.94 (0.89&#x2013;0.99)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Control group composition</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.87</td>
<td/>
<td align="char" valign="top" char=".">0.99</td>
</tr>
<tr>
<td align="left" valign="top">Normal mucosa</td>
<td align="center" valign="top">4</td>
<td align="char" valign="top" char="(">0.87 (0.71&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.94 (0.89&#x2013;0.99)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Mixed normal and precancerous mucosa</td>
<td align="center" valign="top">11</td>
<td align="char" valign="top" char="(">0.92 (0.85&#x2013;0.98)</td>
<td/>
<td align="char" valign="top" char="(">0.90 (0.78&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">Year of publication</td>
<td/>
<td/>
<td align="char" valign="top" char=".">0.47</td>
<td/>
<td align="char" valign="top" char=".">0.68</td>
</tr>
<tr>
<td align="left" valign="top">&#x2264;2020</td>
<td align="center" valign="top">3</td>
<td align="char" valign="top" char="(">0.83 (0.61&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.89 (0.73&#x2013;1.00)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">&#x003E;2020</td>
<td align="center" valign="top">12</td>
<td align="char" valign="top" char="(">0.92 (0.86&#x2013;0.98)</td>
<td/>
<td align="char" valign="top" char="(">0.94 (0.89&#x2013;0.98)</td>
<td/>
</tr>
<tr>
<td align="left" valign="top">DL model types</td>
<td/>
<td/>
<td>0.80</td>
<td/>
<td>0.57</td>
</tr>
<tr>
<td align="left" valign="top">Image classification models</td>
<td align="center" valign="top">10</td>
<td align="char" valign="top" char="(">0.88 (0.79&#x2013;0.97)</td>
<td align="char" valign="top" char="."></td>
<td align="char" valign="top" char="(">0.93 (0.87&#x2013;0.99)</td>
<td align="char" valign="top" char="."></td>
</tr>
<tr>
<td align="left" valign="top">Lesion detection models</td>
<td align="center" valign="top">5</td>
<td align="char" valign="top" char="(">0.94 (0.88&#x2013;1.00)</td>
<td/>
<td align="char" valign="top" char="(">0.94 (0.86&#x2013;1.00)</td>
<td/>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec13">
<title>Diagnostic performance of deep learning algorithms in the external validation set for early gastric cancer detection</title>
<p>For the external validation dataset, DL algorithms based on white endoscopy images achieved a sensitivity of 0.82 (95% CI: 0.61&#x2013;0.93) and a specificity of 0.83 (95% CI: 0.74&#x2013;0.90) in detecting EGC patients (<xref ref-type="fig" rid="fig6">Figure 6</xref>). The AUC was 0.89 (95% CI: 0.86&#x2013;0.91) (<xref ref-type="fig" rid="fig4">Figure 4b</xref>). With a pre-test probability (prevalence) of 36%, the Fagan nomogram demonstrated a positive post-test probability of 73% and a negative post-test probability of 11% (<xref ref-type="fig" rid="fig5">Figure 5b</xref>). High heterogeneity was observed in both sensitivity (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;95.56%, <italic>&#x03C4;</italic><sup>2</sup>&#x202F;=&#x202F;1.09) and specificity (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;97.25%, <italic>&#x03C4;</italic><sup>2</sup>&#x202F;=&#x202F;0.31) within the external validation dataset. Due to the limited number of included studies, meta-regression analysis was not performed to explore potential sources of heterogeneity.</p>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Forest plot of sensitivity and specificity of deep learning algorithms for detecting early gastric cancer (EGC) in the external validation set. Squares represent individual study estimates, with horizontal lines indicating 95% confidence intervals; the diamond denotes the pooled estimate.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g006.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Forest plot comparing the sensitivity and specificity of four studies: Dong et al./2023, Gong et al./2023, Tang et al./2020, and Cho et al./2019. Sensitivity ranges from 0.42 to 0.93 with a combined value of 0.82. Specificity ranges from 0.68 to 0.90 with a combined value of 0.83. Red dashed lines indicate pooled estimates with confidence intervals shown as horizontal lines. The plot shows overall heterogeneity and p-values.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec14">
<title>Deep learning algorithms versus endoscopists: performance in early gastric cancer detection in the test set</title>
<p>In the comparison between the DL model and endoscopists on the test set, substantial heterogeneity was observed in diagnostic sensitivity (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;89.2%, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.0001) (<xref ref-type="fig" rid="fig7">Figure 7</xref>). A random-effects model was used for primary analysis, which showed no statistically significant difference between the two groups (pooled OR&#x202F;=&#x202F;2.21, 95% CI: 0.86&#x2013;5.69), indicating comparable sensitivity performance.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Forest plot comparing the sensitivity of artificial intelligence and endoscopists in detecting early gastric cancer (EGC) in the test set.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g007.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Forest plot comparing the diagnostic sensitivity of deep learning (DL) models versus endoscopists across five studies&#x2014;Cho et al. (2019), Tang et al. (2020), Zhang et al. (2021), Yuan et al. (2022), and Takemoto et al. (2023)&#x2014;using odds ratios (ORs). Each study&#x2019;s OR and 95% confidence interval are displayed, with weights based on a random-effects model. The pooled analysis yields a significant combined OR of 2.21, indicating higher sensitivity for DL models. Substantial heterogeneity is observed (I-squared = 89.2%; p &#x003C; 0.0001).</alt-text>
</graphic>
</fig>
<p>Similarly, for diagnostic specificity, significant heterogeneity was present (<italic>I</italic><sup>2</sup>&#x202F;=&#x202F;94.9%, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.0001) (<xref ref-type="fig" rid="fig8">Figure 8</xref>). The random-effects model revealed no significant difference between DL and endoscopists (pooled OR&#x202F;=&#x202F;0.66, 95% CI: 0.22&#x2013;1.97), suggesting similar specificity performance.</p>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>Forest plot comparing the specificity of artificial intelligence and endoscopists in detecting early gastric cancer (EGC) in the test set.</p>
</caption>
<graphic xlink:href="frai-09-1734591-g008.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Forest plot comparing the diagnostic specificity of deep learning (DL) models versus endoscopists across five studies&#x2014;Cho et al. (2019), Tang et al. (2020), Zhang et al. (2021), Yuan et al. (2022), and Takemoto et al. (2023)&#x2014;using odds ratios (ORs). Each study&#x2019;s OR and 95% confidence interval are displayed, with weights based on a randomeffects model. The pooled analysis yields a significant combined OR of 0.66, indicating lower specificity for DL models. Substantial heterogeneity is observed (I-squared = 94.9%; p &#x003C; 0.0001).</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec15">
<title>Publication bias</title>
<p>The Deeks&#x2019; funnel plot asymmetry test showed no significant publication bias in the internal validation dataset based on white light endoscopy images for DL (<italic>p</italic>&#x202F;&#x003E;&#x202F;0.05) (<xref rid="SM1" ref-type="supplementary-material">Supplementary Figure 1</xref>). In contrast, the Deeks&#x2019; funnel plot asymmetry test revealed significant publication bias in the external validation dataset, which consisted of only four studies utilizing white light endoscopy images (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05; <xref rid="SM1" ref-type="supplementary-material">Supplementary Figure 2</xref>).</p>
</sec>
</sec>
<sec sec-type="discussion" id="sec16">
<title>Discussion</title>
<p>To the best of our knowledge, this is the first meta-analysis to comprehensively evaluate the performance of DL algorithms in diagnosing EGC using white light endoscopic images. The results indicate that DL algorithms exhibit excellent diagnostic performance in the internal validation set, with a sensitivity of 0.91, a specificity of 0.93, and an AUC of 0.97. In the external validation set, the diagnostic sensitivity, specificity, and AUC were 0.82, 0.83, and 0.89, respectively, which were lower than those in the internal validation set. Furthermore, no significant differences were observed between DL algorithms and expert endoscopists in terms of diagnostic sensitivity or specificity. Meta-regression analysis indicates that the sample size of the training dataset contributes to the high heterogeneity in sensitivity and specificity observed in the internal validation sets. In summary, these results suggest that DL algorithms demonstrate good diagnostic performance in detecting EGC using white-light endoscopic images, indicating their potential as a reliable auxiliary diagnostic tool.</p>
<p>Sensitivity and specificity are key metrics for evaluating diagnostic performance. In this study, the DL model demonstrated high sensitivity and specificity in the internal validation set. High sensitivity indicates a low risk of missed diagnosis, facilitating the detection of EGC with atypical morphology or indistinct borders. High specificity reflects a low false-positive rate, conducive to reducing unnecessary biopsy procedures and thereby preventing overdiagnosis and overtreatment. The strong performance observed in the internal validation may be attributed to consistent data preprocessing, standardized image acquisition protocols, and uniform endoscopic imaging conditions (<xref ref-type="bibr" rid="ref30">Li et al., 2025</xref>). These factors help minimize technical variability, enabling the model to more accurately distinguish EGC from non-EGC findings. However, in the external validation set, both sensitivity and specificity were lower than those observed in the internal validation. This performance decline is likely due to real-world variations across institutions, such as differences in endoscopist expertise, types of endoscopic equipment, and image quality (<xref ref-type="bibr" rid="ref6">Campanella et al., 2019</xref>). These heterogeneities introduce noise and complexity that the model may not have fully accounted for during training. These findings underscore the importance of standardized data pipelines and the use of diverse, multi-center datasets during model development to improve model generalizability and robustness.</p>
<p>Currently, due to limitations in technical skills and clinical experience, trainee endoscopists exhibit significantly lower sensitivity and specificity in diagnosing EGC compared to expert endoscopists (<xref ref-type="bibr" rid="ref12">Ende et al., 2018</xref>; <xref ref-type="bibr" rid="ref46">Tang et al., 2020</xref>; <xref ref-type="bibr" rid="ref56">Yuan et al., 2022</xref>). This performance gap contributes to instability in clinical endoscopic practice and increases the risk of missed or incorrect diagnoses, especially in primary care hospitals. Previous studies revealed that, with AI assistance, trained novices can produce expert-level lung and cardiac ultrasound images that can be used to assess pathology after a short training session, thereby enhancing access to diagnosis in resource-constrained settings (<xref ref-type="bibr" rid="ref36">Narang et al., 2021</xref>; <xref ref-type="bibr" rid="ref5">Baloescu et al., 2025</xref>). In this study, our results demonstrate that DL algorithms achieve sensitivity and specificity comparable to those of expert endoscopists. Therefore, it is reasonable to hypothesize that AI may serve as an effective assistive tool to enhance the sensitivity and specificity of trainee endoscopists in the detection of EGC during white-light endoscopy screening, thereby minimizing the likelihood of missed or incorrect diagnoses and facilitating earlier detection and timely intervention.</p>
<p>In the internal validation of deep-learning algorithms, meta-regression analysis demonstrated that models trained on large-scale public datasets exhibited significantly superior diagnostic sensitivity and specificity compared to those trained on small-scale institutional datasets. This finding indicates that the size of the training dataset may be one of the key factors determining the diagnostic performance of the deep-learning algorithms. Previous studies revealed that merely expanding the size of the training dataset can improve the classification performance of the DL network (<xref ref-type="bibr" rid="ref27">Kiryati and Landau, 2021</xref>; <xref ref-type="bibr" rid="ref39">Pei et al., 2021</xref>). However, due to the challenges in acquiring and annotating medical imaging data, particularly in the three-dimensional context of endoscopic examinations, constructing large and high-quality training datasets was difficult (<xref ref-type="bibr" rid="ref44">Tajbakhsh et al., 2016</xref>; <xref ref-type="bibr" rid="ref9">Chen X. et al., 2022</xref>). In contrast, public datasets offered a viable pathway to overcome these difficulties. ImageNet is a large-scale hierarchical visual recognition database developed in the United States, comprising 14 million manually labeled images (<xref ref-type="bibr" rid="ref25">Kang et al., 2021</xref>). Kvasir-SEG is a publicly accessible high-quality gastrointestinal endoscopy dataset originating from Norway, comprising 1,000 images annotated with pixel-level segmentation masks (<xref ref-type="bibr" rid="ref23">Jha et al., 2019</xref>). Consequently, in this meta-analysis, deep-learning algorithms trained on ImageNet and Kvasir-SEG datasets achieve superior performance in EGC detection. Furthermore, although cross-validation is an important technique for evaluating model robustness, particularly in studies with small datasets, our analysis did not observe a significant influence of cross-validation on heterogeneity within the internal dataset (<xref ref-type="bibr" rid="ref1">Aggarwal et al., 2022</xref>). Similarly, factors including the number of participating centers, image size, and study quality did not contribute significantly to internal heterogeneity. However, this heterogeneity may stem from other potential factors such as clinical staging of EGC, image quality, and variations in the definition of EGC.</p>
<p>To our knowledge, this is the first meta-analysis specifically evaluating the diagnostic performance of DL algorithms for EGC. In contrast, a prior meta-analysis of 12 studies reported that AI&#x2014;encompassing both machine learning and DL algorithms&#x2014;achieved a sensitivity of 0.86 and a specificity of 0.90 in the diagnosis of EGC, values notably lower than the 0.91 and 0.93 observed in this study (<xref ref-type="bibr" rid="ref8">Chen P.-C. et al., 2022</xref>). This discrepancy may be attributed to differences in algorithmic model selection (DL versus a combination of machine learning and DL). At the algorithmic level, traditional machine learning methods rely on handcrafted feature engineering and exhibit limited generalizability, particularly when applied to complex and heterogeneous medical imaging data (<xref ref-type="bibr" rid="ref34">Moawad et al., 2022</xref>). In contrast, the DL models evaluated in this study enable end-to-end learning by automatically extracting hierarchical feature representations directly from raw images, thereby achieving enhanced robustness and higher diagnostic accuracy in complex visual recognition tasks (<xref ref-type="bibr" rid="ref53">Wang et al., 2019b</xref>).</p>
<p>With a pre-test probability of 36%, the Fagan nomogram demonstrated a positive post-test probability of 73% and a negative post-test probability of 11%. This provides a practical tool for clinicians: for a patient with a pre-test suspicion of 36%, a positive result from the DL model would increase the probability of EGC to 73%, warranting a confirmatory biopsy. Conversely, a negative result would lower the probability to 11%, potentially supporting a decision for surveillance rather than immediate intervention, depending on the clinical context. From a clinical implementation perspective, these findings support the role of DL-based systems as decision-support tools rather than standalone diagnostic solutions. Practical deployment would require targeted training for endoscopists on AI-assisted interpretation within endoscopy suites, alongside clearly defined safety workflows to ensure clinician oversight (<xref ref-type="bibr" rid="ref38">Olawuyi and Viriri, 2025</xref>). Moreover, regulatory approval is a prerequisite for clinical adoption. Similar to AI-based electrocardiogram detection systems, AI models for early gastric cancer detection require formal evaluation and regulatory clearance from authorities such as the FDA or CE bodies (<xref ref-type="bibr" rid="ref42">Singla et al., 2025</xref>). Such approval usually depends on robust external and prospective validation, which remains limited in current studies. From a methodological perspective, future improvements in DL-based EGC detection may benefit from incorporating Transformer-based architectures (e.g., Vision Transformer and Swin Transformer), which have shown strong performance in medical image analysis by capturing long-range spatial dependencies (<xref ref-type="bibr" rid="ref15">Gandhi et al., 2025a</xref>). In addition, generative data augmentation techniques could help mitigate data imbalance and enhance model robustness (<xref ref-type="bibr" rid="ref16">Gandhi et al., 2025b</xref>). The integration of multimodal learning frameworks, combining endoscopic video data with relevant clinical information, may further improve diagnostic accuracy and clinical relevance (<xref ref-type="bibr" rid="ref40">Qin et al., 2025</xref>). To address data privacy and enhance generalizability, federated learning offers a promising strategy for leveraging multicenter data without direct data sharing (<xref ref-type="bibr" rid="ref3">Assaf et al., 2025</xref>). Finally, adoption of standardized reporting guidelines, such as CONSORT-AI, DECIDE-AI, and STARD-AI, is essential to improve transparency, reproducibility, and clinical interpretability of future studies (<xref ref-type="bibr" rid="ref19">Goh et al., 2025</xref>).</p>
<p>In addition, it is important to note that the generalizability of these performance estimates may be further challenged by the lack of temporal validation in most included studies. Robust clinical prediction systems require testing on data from future time periods to ensure stability against shifts in clinical practice, equipment, or patient demographics. In our review, only one study employed a prospective external validation set (<xref ref-type="bibr" rid="ref41">Sakai et al., 2018</xref>). Future studies should prioritize this design to provide a more rigorous and clinically realistic assessment of model performance over time. Furthermore, the presence of publication bias in the external validation dataset likely stems from the limited number of available studies and potential selective reporting of higher-performing models in externally validated literature. The observed bias suggests that the overall diagnostic performance of DL models in external validation settings may be overestimated in the current literature. Therefore, the establishment of multi-center, large-scale external validation cohorts is essential for a comprehensive evaluation of DL model performance.</p>
<p>Several limitations of this meta-analysis should be acknowledged when interpreting the findings.</p>
<p>First, a fundamental limitation of this analysis is that all included studies utilized retrospective datasets for both model development and validation. This retrospective design inherently carries risks of selection bias and spectrum bias, where the case mix may not fully represent the broader clinical population encountered in practice. Therefore, while our meta-analysis suggests promising diagnostic potential, the reported high accuracy likely represents a &#x201C;best-case&#x201D; scenario. Forthcoming multi-center, prospective trials are crucial to rigorously evaluate model performance in unselected, consecutive patients under real-world conditions (<xref ref-type="bibr" rid="ref49">Tong et al., 2023</xref>). Second, there was variability in the definition of EGC across the included studies, and the inclusion criteria for control groups were inconsistent. This heterogeneity in the control population&#x2014;ranging from purely normal mucosa to a mix of benign lesions (e.g., gastric ulcers, low-grade epithelial neoplasia, gastric polyps)&#x2014;constitutes a potential source of classification bias. Such inconsistency may lead to systematic differences in model training and evaluation, as models trained against purely normal mucosa might achieve higher specificity in distinguishing cancer from normal tissue but potentially lower sensitivity for discriminating early cancer from challenging benign or precancerous conditions. Third, the current analysis was restricted to image-level evaluation of DL models due to incomplete patient-level data in the included studies. However, patient-level assessment better aligns with clinical practice. Image-based training risks overfitting to specific features within individual patients, which may limit the model&#x2019;s applicability to external datasets (<xref ref-type="bibr" rid="ref28">Lengerich et al., 2018</xref>). Fourth, all included studies focused on detecting gastric lesions from static, high-quality white-light endoscopic images, which inherently cannot reproduce the complexity of real-time endoscopy. In actual clinical practice, endoscopic observation is dynamic and often affected by motion blur caused by scope movement, variations in illumination, changes in viewing angle, and transient interference from mucus, blood, bubbles, or food residue. These <italic>in situ</italic> factors substantially increase diagnostic difficulty but were largely excluded from the training and validation datasets of the included studies. Consequently, the reported diagnostic performance of AI models derived from idealized image datasets may overestimate their effectiveness in real-world, real-time clinical settings. Future studies should therefore prioritize validation using video-based or real-time endoscopic data that better reflect routine clinical conditions.</p>
<p>In conclusion, our meta-analysis provides robust evidence that DL algorithms exhibit high diagnostic efficacy in detecting EGC from white-light endoscopic images. Moreover, the sensitivity and specificity of these algorithms are comparable to those of expert endoscopists. These findings highlight the potential for DL algorithms to serve as a clinical decision-support tool in routine practice.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec17">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref rid="SM1" ref-type="supplementary-material">Supplementary material</xref>, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="sec18">
<title>Author contributions</title>
<p>JL: Writing &#x2013; original draft, Data curation, Formal analysis. DL: Software, Writing &#x2013; review &#x0026; editing, Validation. YZ: Writing &#x2013; review &#x0026; editing, Data curation, Formal analysis. SZ: Writing &#x2013; review &#x0026; editing, Supervision.</p>
</sec>
<sec sec-type="COI-statement" id="sec19">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec20">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec21">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="sec22">
<title>Supplementary material</title>
<p>The Supplementary material for this article can be found online at: <ext-link xlink:href="https://www.frontiersin.org/articles/10.3389/frai.2026.1734591/full#supplementary-material" ext-link-type="uri">https://www.frontiersin.org/articles/10.3389/frai.2026.1734591/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Supplementary_file_1.docx" id="SM1" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aggarwal</surname><given-names>P.</given-names></name> <name><surname>Mishra</surname><given-names>N. K.</given-names></name> <name><surname>Fatimah</surname><given-names>B.</given-names></name> <name><surname>Singh</surname><given-names>P.</given-names></name> <name><surname>Gupta</surname><given-names>A.</given-names></name> <name><surname>Joshi</surname><given-names>S. D.</given-names></name></person-group> (<year>2022</year>). <article-title>COVID-19 image classification using deep learning: advances, challenges and opportunities</article-title>. <source>Comput. Biol. Med.</source> <volume>144</volume>:<fpage>105350</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compbiomed.2022.105350</pub-id>, <pub-id pub-id-type="pmid">35305501</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Arnold</surname><given-names>M.</given-names></name> <name><surname>Park</surname><given-names>J. Y.</given-names></name> <name><surname>Camargo</surname><given-names>M. C.</given-names></name> <name><surname>Lunet</surname><given-names>N.</given-names></name> <name><surname>Forman</surname><given-names>D.</given-names></name> <name><surname>Soerjomataram</surname><given-names>I.</given-names></name></person-group> (<year>2020</year>). <article-title>Is gastric cancer becoming a rare disease? A global assessment of predicted incidence trends to 2035</article-title>. <source>Gut</source> <volume>69</volume>, <fpage>823</fpage>&#x2013;<lpage>829</lpage>. doi: <pub-id pub-id-type="doi">10.1136/gutjnl-2019-320234</pub-id>, <pub-id pub-id-type="pmid">32001553</pub-id></mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Assaf</surname><given-names>J. F.</given-names></name> <name><surname>Ahuja</surname><given-names>A. S.</given-names></name> <name><surname>Kannan</surname><given-names>V.</given-names></name> <name><surname>Yazbeck</surname><given-names>H.</given-names></name> <name><surname>Krivit</surname><given-names>J.</given-names></name> <name><surname>Redd</surname><given-names>T. K.</given-names></name></person-group> (<year>2025</year>). <article-title>Applications of computer vision for infectious keratitis: a systematic review</article-title>. <source>Ophthalmol. Sci.</source> <volume>5</volume>:<fpage>100861</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.xops.2025.100861</pub-id>, <pub-id pub-id-type="pmid">40778364</pub-id></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Baldominos</surname><given-names>A.</given-names></name> <name><surname>Cervantes</surname><given-names>A.</given-names></name> <name><surname>Saez</surname><given-names>Y.</given-names></name> <name><surname>Isasi</surname><given-names>P.</given-names></name></person-group> (<year>2019</year>). <article-title>A comparison of machine learning and deep learning techniques for activity recognition using mobile devices</article-title>. <source>Sensors</source> <volume>19</volume>:<fpage>521</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s19030521</pub-id>, <pub-id pub-id-type="pmid">30691177</pub-id></mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Baloescu</surname><given-names>C.</given-names></name> <name><surname>Bailitz</surname><given-names>J.</given-names></name> <name><surname>Cheema</surname><given-names>B.</given-names></name> <name><surname>Agarwala</surname><given-names>R.</given-names></name> <name><surname>Jankowski</surname><given-names>M.</given-names></name> <name><surname>Eke</surname><given-names>O.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Artificial intelligence-guided lung ultrasound by nonexperts</article-title>. <source>JAMA Cardiol.</source> <volume>10</volume>, <fpage>245</fpage>&#x2013;<lpage>253</lpage>. doi: <pub-id pub-id-type="doi">10.1001/jamacardio.2024.4991</pub-id>, <pub-id pub-id-type="pmid">39813064</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Campanella</surname><given-names>G.</given-names></name> <name><surname>Hanna</surname><given-names>M. G.</given-names></name> <name><surname>Geneslaw</surname><given-names>L.</given-names></name> <name><surname>Miraflor</surname><given-names>A.</given-names></name> <name><surname>Werneck Krauss Silva</surname><given-names>V.</given-names></name> <name><surname>Busam</surname><given-names>K. J.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Clinical-grade computational pathology using weakly supervised deep learning on whole slide images</article-title>. <source>Nat. Med.</source> <volume>25</volume>, <fpage>1301</fpage>&#x2013;<lpage>1309</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41591-019-0508-1</pub-id>, <pub-id pub-id-type="pmid">31308507</pub-id></mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname><given-names>Y. H.</given-names></name> <name><surname>Shin</surname><given-names>C. M.</given-names></name> <name><surname>Lee</surname><given-names>H. D.</given-names></name> <name><surname>Park</surname><given-names>J.</given-names></name> <name><surname>Jeon</surname><given-names>J.</given-names></name> <name><surname>Cho</surname><given-names>S.-J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Real-world application of artificial intelligence for detecting pathologic gastric atypia and neoplastic lesions</article-title>. <source>J. Gastric Cancer</source> <volume>24</volume>, <fpage>327</fpage>&#x2013;<lpage>340</lpage>. doi: <pub-id pub-id-type="doi">10.5230/jgc.2024.24.e28</pub-id>, <pub-id pub-id-type="pmid">38960891</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>P.-C.</given-names></name> <name><surname>Lu</surname><given-names>Y.-R.</given-names></name> <name><surname>Kang</surname><given-names>Y.-N.</given-names></name> <name><surname>Chang</surname><given-names>C.-C.</given-names></name></person-group> (<year>2022</year>). <article-title>The accuracy of artificial intelligence in the endoscopic diagnosis of early gastric cancer: pooled analysis study</article-title>. <source>J. Med. Internet Res.</source> <volume>24</volume>:<fpage>e27694</fpage>. doi: <pub-id pub-id-type="doi">10.2196/27694</pub-id>, <pub-id pub-id-type="pmid">35576561</pub-id></mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>X.</given-names></name> <name><surname>Wang</surname><given-names>X.</given-names></name> <name><surname>Zhang</surname><given-names>K.</given-names></name> <name><surname>Fung</surname><given-names>K.-M.</given-names></name> <name><surname>Thai</surname><given-names>T. C.</given-names></name> <name><surname>Moore</surname><given-names>K.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Recent advances and clinical applications of deep learning in medical image analysis</article-title>. <source>Med. Image Anal.</source> <volume>79</volume>:<fpage>102444</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2022.102444</pub-id>, <pub-id pub-id-type="pmid">35472844</pub-id></mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cho</surname><given-names>B.-J.</given-names></name> <name><surname>Bang</surname><given-names>C. S.</given-names></name> <name><surname>Park</surname><given-names>S. W.</given-names></name> <name><surname>Yang</surname><given-names>Y. J.</given-names></name> <name><surname>Seo</surname><given-names>S. I.</given-names></name> <name><surname>Lim</surname><given-names>H.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Automated classification of gastric neoplasms in endoscopic images using a convolutional neural network</article-title>. <source>Endoscopy</source> <volume>51</volume>, <fpage>1121</fpage>&#x2013;<lpage>1129</lpage>. doi: <pub-id pub-id-type="doi">10.1055/a-0981-6133</pub-id>, <pub-id pub-id-type="pmid">31443108</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dong</surname><given-names>Z.</given-names></name> <name><surname>Wang</surname><given-names>J.</given-names></name> <name><surname>Li</surname><given-names>Y.</given-names></name> <name><surname>Deng</surname><given-names>Y.</given-names></name> <name><surname>Zhou</surname><given-names>W.</given-names></name> <name><surname>Zeng</surname><given-names>X.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Explainable artificial intelligence incorporated with domain knowledge diagnosing early gastric neoplasms under white light endoscopy</article-title>. <source>NPJ Digit. Med.</source> <volume>6</volume>:<fpage>64</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41746-023-00813-y</pub-id></mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ende</surname><given-names>A. R.</given-names></name> <name><surname>De Groen</surname><given-names>P.</given-names></name> <name><surname>Balmadrid</surname><given-names>B. L.</given-names></name> <name><surname>Hwang</surname><given-names>J. H.</given-names></name> <name><surname>Inadomi</surname><given-names>J.</given-names></name> <name><surname>Wojtera</surname><given-names>T.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Objective differences in colonoscopy technique between trainee and expert endoscopists using the colonoscopy force monitor</article-title>. <source>Dig. Dis. Sci.</source> <volume>63</volume>, <fpage>46</fpage>&#x2013;<lpage>52</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10620-017-4847-9</pub-id>, <pub-id pub-id-type="pmid">29147876</pub-id></mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Esteva</surname><given-names>A.</given-names></name> <name><surname>Robicquet</surname><given-names>A.</given-names></name> <name><surname>Ramsundar</surname><given-names>B.</given-names></name> <name><surname>Kuleshov</surname><given-names>V.</given-names></name> <name><surname>DePristo</surname><given-names>M.</given-names></name> <name><surname>Chou</surname><given-names>K.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>A guide to deep learning in healthcare</article-title>. <source>Nat. Med.</source> <volume>25</volume>, <fpage>24</fpage>&#x2013;<lpage>29</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41591-018-0316-z</pub-id>, <pub-id pub-id-type="pmid">30617335</pub-id></mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname><given-names>J.</given-names></name> <name><surname>Zhang</surname><given-names>Y.</given-names></name> <name><surname>Feng</surname><given-names>Z.</given-names></name> <name><surname>Ma</surname><given-names>H.</given-names></name> <name><surname>Gou</surname><given-names>Y.</given-names></name> <name><surname>Wang</surname><given-names>P.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>A prospective and comparative study on improving the diagnostic accuracy of early gastric cancer based on deep convolutional neural network real-time diagnosis system (with video)</article-title>. <source>Surg. Endosc.</source> <volume>39</volume>, <fpage>1874</fpage>&#x2013;<lpage>1884</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00464-025-11527-5</pub-id>, <pub-id pub-id-type="pmid">39843600</pub-id></mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gandhi</surname><given-names>V. C.</given-names></name> <name><surname>Gandhi</surname><given-names>P.</given-names></name> <name><surname>Ogundiran</surname><given-names>J. O.</given-names></name> <name><surname>Tshibola</surname><given-names>M. S. S.</given-names></name> <name><surname>Kapuya Bulaba Nyembwe</surname><given-names>J.-P.</given-names></name></person-group> (<year>2025a</year>). <article-title>Computational modeling and optimization of deep learning for multi-modal glaucoma diagnosis</article-title>. <source>Appl. Math.</source> <volume>5</volume>:<fpage>82</fpage>. doi: <pub-id pub-id-type="doi">10.3390/appliedmath5030082</pub-id></mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gandhi</surname><given-names>V. C.</given-names></name> <name><surname>Gandhi</surname><given-names>P. P.</given-names></name> <name><surname>Oza</surname><given-names>A. D.</given-names></name> <name><surname>Al-Nussairi</surname><given-names>A. K. J.</given-names></name> <name><surname>Hadi</surname><given-names>A. A.</given-names></name> <name><surname>Alamiery</surname><given-names>A. A.</given-names></name> <etal/></person-group>. (<year>2025b</year>). <article-title>Identifying glaucoma with deep learning by utilizing the VGG16 model for retinal image analysis</article-title>. <source>Intell. Based Med.</source> <volume>12</volume>:<fpage>100307</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ibmed.2025.100307</pub-id></mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Gandhi</surname><given-names>V. C.</given-names></name> <name><surname>Thakkar</surname><given-names>D.</given-names></name> <name><surname>Milanova</surname><given-names>M.</given-names></name></person-group> (<year>2025c</year>). &#x201C;<article-title>Unveiling Alzheimer&#x2019;s progression: AI-driven models for classifying stages of cognitive impairment through medical imaging</article-title>&#x201D; in <source>Pattern recognition. ICPR 2024 international workshops and challenges</source>. eds. <person-group person-group-type="editor"><name><surname>Palaiahnakote</surname><given-names>S.</given-names></name> <name><surname>Schuckers</surname><given-names>S.</given-names></name> <name><surname>Ogier</surname><given-names>J.-M.</given-names></name> <name><surname>Bhattacharya</surname><given-names>P.</given-names></name> <name><surname>Pal</surname><given-names>U.</given-names></name> <name><surname>Bhattacharya</surname><given-names>S.</given-names></name></person-group> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer Nature Switzerland</publisher-name>), <fpage>55</fpage>&#x2013;<lpage>87</lpage>.</mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="journal"><collab id="coll1">GASTRIC (Global Advanced/Adjuvant Stomach Tumor Research International Collaboration) Group</collab><person-group person-group-type="author"><name><surname>Oba</surname><given-names>K.</given-names></name> <name><surname>Paoletti</surname><given-names>X.</given-names></name> <name><surname>Bang</surname><given-names>Y.-J.</given-names></name> <name><surname>Bleiberg</surname><given-names>H.</given-names></name> <name><surname>Burzykowski</surname><given-names>T.</given-names></name> <etal/></person-group>. (<year>2013</year>). <article-title>Role of chemotherapy for advanced/recurrent gastric cancer: an individual-patient-data meta-analysis</article-title>. <source>Eur. J. Cancer Oxf. Engl. 1990</source> <volume>49</volume>, <fpage>1565</fpage>&#x2013;<lpage>1577</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ejca.2012.12.016</pub-id></mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Goh</surname><given-names>S.</given-names></name> <name><surname>Goh</surname><given-names>R. S. J.</given-names></name> <name><surname>Chong</surname><given-names>B.</given-names></name> <name><surname>Ng</surname><given-names>Q. X.</given-names></name> <name><surname>Koh</surname><given-names>G. C. H.</given-names></name> <name><surname>Ngiam</surname><given-names>K. Y.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Challenges in implementing artificial intelligence in breast cancer screening programs: systematic review and framework for safe adoption</article-title>. <source>J. Med. Internet Res.</source> <volume>27</volume>:<fpage>e62941</fpage>. doi: <pub-id pub-id-type="doi">10.2196/62941</pub-id>, <pub-id pub-id-type="pmid">40373301</pub-id></mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gong</surname><given-names>E. J.</given-names></name> <name><surname>Bang</surname><given-names>C. S.</given-names></name> <name><surname>Lee</surname><given-names>J. J.</given-names></name></person-group> (<year>2024</year>). <article-title>Computer-aided diagnosis in real-time endoscopy for all stages of gastric carcinogenesis: development and validation study</article-title>. <source>United Eur. Gastroenterol. J.</source> <volume>12</volume>, <fpage>487</fpage>&#x2013;<lpage>495</lpage>. doi: <pub-id pub-id-type="doi">10.1002/ueg2.12551</pub-id>, <pub-id pub-id-type="pmid">38400815</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huedo-Medina</surname><given-names>T. B.</given-names></name> <name><surname>S&#x00E1;nchez-Meca</surname><given-names>J.</given-names></name> <name><surname>Marin-Martinez</surname><given-names>F.</given-names></name> <name><surname>Botella</surname><given-names>J.</given-names></name></person-group> (<year>2006</year>). <article-title>Assessing heterogeneity in meta-analysis: q statistic or I2 index?</article-title> <source>Psychol. Methods</source> <volume>11</volume>:<fpage>193</fpage>. doi: <pub-id pub-id-type="doi">10.1037/1082-989X.11.2.193</pub-id></mixed-citation></ref>
<ref id="ref22"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Im</surname><given-names>S.</given-names></name> <name><surname>Hyeon</surname><given-names>J.</given-names></name> <name><surname>Rha</surname><given-names>E.</given-names></name> <name><surname>Lee</surname><given-names>J.</given-names></name> <name><surname>Choi</surname><given-names>H.-J.</given-names></name> <name><surname>Jung</surname><given-names>Y.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Classification of diffuse glioma subtype from clinical-grade pathological images using deep transfer learning</article-title>. <source>Sensors</source> <volume>21</volume>:<fpage>3500</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s21103500</pub-id>, <pub-id pub-id-type="pmid">34067934</pub-id></mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Jha</surname><given-names>D.</given-names></name> <name><surname>Smedsrud</surname><given-names>P. H.</given-names></name> <name><surname>Riegler</surname><given-names>M. A.</given-names></name> <name><surname>Halvorsen</surname><given-names>P.</given-names></name> <name><surname>de Lange</surname><given-names>T.</given-names></name> <name><surname>Johansen</surname><given-names>D.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Kvasir-SEG: a segmented polyp dataset</article-title>. <volume>arXiv preprint arXiv</volume>:<fpage>1911.07069</fpage>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1911.07069</pub-id></mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jun</surname><given-names>J. K.</given-names></name> <name><surname>Choi</surname><given-names>K. S.</given-names></name> <name><surname>Lee</surname><given-names>H.-Y.</given-names></name> <name><surname>Suh</surname><given-names>M.</given-names></name> <name><surname>Park</surname><given-names>B.</given-names></name> <name><surname>Song</surname><given-names>S. H.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>Effectiveness of the Korean national cancer screening program in reducing gastric cancer mortality</article-title>. <source>Gastroenterology</source> <volume>152</volume>, <fpage>1319</fpage>&#x2013;<lpage>1328.e7</lpage>. doi: <pub-id pub-id-type="doi">10.1053/j.gastro.2017.01.029</pub-id>, <pub-id pub-id-type="pmid">28147224</pub-id></mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kang</surname><given-names>D.</given-names></name> <name><surname>Gweon</surname><given-names>H. M.</given-names></name> <name><surname>Eun</surname><given-names>N. L.</given-names></name> <name><surname>Youk</surname><given-names>J. H.</given-names></name> <name><surname>Kim</surname><given-names>J.-A.</given-names></name> <name><surname>Son</surname><given-names>E. J.</given-names></name></person-group> (<year>2021</year>). <article-title>A convolutional deep learning model for improving mammographic breast-microcalcification diagnosis</article-title>. <source>Sci. Rep.</source> <volume>11</volume>:<fpage>23925</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-021-03516-0</pub-id>, <pub-id pub-id-type="pmid">34907330</pub-id></mixed-citation></ref>
<ref id="ref26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Katai</surname><given-names>H.</given-names></name> <name><surname>Ishikawa</surname><given-names>T.</given-names></name> <name><surname>Akazawa</surname><given-names>K.</given-names></name> <name><surname>Isobe</surname><given-names>Y.</given-names></name> <name><surname>Miyashiro</surname><given-names>I.</given-names></name> <name><surname>Oda</surname><given-names>I.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Five-year survival analysis of surgically resected gastric cancer cases in Japan: a retrospective analysis of more than 100,000 patients from the nationwide registry of the Japanese gastric cancer association (2001-2007)</article-title>. <source>Gastric Cancer</source> <volume>21</volume>, <fpage>144</fpage>&#x2013;<lpage>154</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10120-017-0716-7</pub-id>, <pub-id pub-id-type="pmid">28417260</pub-id></mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kiryati</surname><given-names>N.</given-names></name> <name><surname>Landau</surname><given-names>Y.</given-names></name></person-group> (<year>2021</year>). <article-title>Dataset growth in medical image analysis research</article-title>. <source>J. Imaging</source> <volume>7</volume>:<fpage>155</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jimaging7080155</pub-id>, <pub-id pub-id-type="pmid">34460791</pub-id></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lengerich</surname><given-names>B. J.</given-names></name> <name><surname>Aragam</surname><given-names>B.</given-names></name> <name><surname>Xing</surname><given-names>E. P.</given-names></name></person-group> (<year>2018</year>). <article-title>Personalized regression enables sample-specific pan-cancer analysis</article-title>. <source>Bioinformatics</source> <volume>34</volume>:<fpage>294496</fpage>. doi: <pub-id pub-id-type="doi">10.1101/294496</pub-id></mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>Z.</given-names></name> <name><surname>Liu</surname><given-names>F.</given-names></name> <name><surname>Yang</surname><given-names>W.</given-names></name> <name><surname>Peng</surname><given-names>S.</given-names></name> <name><surname>Zhou</surname><given-names>J.</given-names></name></person-group> (<year>2022</year>). <article-title>A survey of convolutional neural networks: analysis, applications, and prospects</article-title>. <source>IEEE Trans. Neural Netw. Learn. Syst.</source> <volume>33</volume>, <fpage>6999</fpage>&#x2013;<lpage>7019</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TNNLS.2021.3084827</pub-id>, <pub-id pub-id-type="pmid">34111009</pub-id></mixed-citation></ref>
<ref id="ref30"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>H.</given-names></name> <name><surname>Qin</surname><given-names>J.</given-names></name> <name><surname>Li</surname><given-names>Z.</given-names></name> <name><surname>Ouyang</surname><given-names>R.</given-names></name> <name><surname>Chen</surname><given-names>Z.</given-names></name> <name><surname>Huang</surname><given-names>S.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Systematic review and meta-analysis of deep learning for MSI-H in colorectal cancer whole slide images</article-title>. <source>NPJ Digit. Med.</source> <volume>8</volume>:<fpage>456</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41746-025-01848-z</pub-id>, <pub-id pub-id-type="pmid">40681867</pub-id></mixed-citation></ref>
<ref id="ref31"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>X.</given-names></name> <name><surname>Wang</surname><given-names>X.</given-names></name> <name><surname>Mao</surname><given-names>T.</given-names></name> <name><surname>Yin</surname><given-names>X.</given-names></name> <name><surname>Wei</surname><given-names>Z.</given-names></name> <name><surname>Fu</surname><given-names>J.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Characteristic analysis of early gastric cancer after <italic>Helicobacter pylori</italic> eradication: a multicenter retrospective propensity score-matched study</article-title>. <source>Ann. Med.</source> <volume>55</volume>:<fpage>2231852</fpage>. doi: <pub-id pub-id-type="doi">10.1080/07853890.2023.2231852</pub-id>, <pub-id pub-id-type="pmid">37450336</pub-id></mixed-citation></ref>
<ref id="ref32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Machlowska</surname><given-names>J.</given-names></name> <name><surname>Baj</surname><given-names>J.</given-names></name> <name><surname>Sitarz</surname><given-names>M.</given-names></name> <name><surname>Maciejewski</surname><given-names>R.</given-names></name> <name><surname>Sitarz</surname><given-names>R.</given-names></name></person-group> (<year>2020</year>). <article-title>Gastric cancer: epidemiology, risk factors, classification, genomic characteristics and treatment strategies</article-title>. <source>Int. J. Mol. Sci.</source> <volume>21</volume>:<fpage>4012</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ijms21114012</pub-id>, <pub-id pub-id-type="pmid">32512697</pub-id></mixed-citation></ref>
<ref id="ref33"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mdf</surname><given-names>M.</given-names></name> <name><surname>D</surname><given-names>M.</given-names></name> <name><surname>Bd</surname><given-names>T.</given-names></name> <name><surname>Ta</surname><given-names>M.</given-names></name> <name><surname>Pm</surname><given-names>B.</given-names></name> <name><surname>T</surname><given-names>C.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: the PRISMA-DTA statement</article-title>. <source>JAMA</source> <volume>319</volume>. doi: <pub-id pub-id-type="doi">10.1001/jama.2017.19163</pub-id></mixed-citation></ref>
<ref id="ref34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Moawad</surname><given-names>A. W.</given-names></name> <name><surname>Fuentes</surname><given-names>D. T.</given-names></name> <name><surname>ElBanan</surname><given-names>M. G.</given-names></name> <name><surname>Shalaby</surname><given-names>A. S.</given-names></name> <name><surname>Guccione</surname><given-names>J.</given-names></name> <name><surname>Kamel</surname><given-names>S.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Artificial intelligence in diagnostic radiology: where do we stand, challenges, and opportunities</article-title>. <source>J. Comput. Assist. Tomogr.</source> <volume>46</volume>, <fpage>78</fpage>&#x2013;<lpage>90</lpage>. doi: <pub-id pub-id-type="doi">10.1097/RCT.0000000000001247</pub-id></mixed-citation></ref>
<ref id="ref35"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nagula</surname><given-names>S.</given-names></name> <name><surname>Parasa</surname><given-names>S.</given-names></name> <name><surname>Laine</surname><given-names>L.</given-names></name> <name><surname>Shah</surname><given-names>S. C.</given-names></name></person-group> (<year>2024</year>). <article-title>AGA clinical practice update on high-quality upper endoscopy: expert review</article-title>. <source>Clin. Gastroenterol. Hepatol.</source> <volume>22</volume>, <fpage>933</fpage>&#x2013;<lpage>943</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cgh.2023.10.034</pub-id></mixed-citation></ref>
<ref id="ref36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Narang</surname><given-names>A.</given-names></name> <name><surname>Bae</surname><given-names>R.</given-names></name> <name><surname>Hong</surname><given-names>H.</given-names></name> <name><surname>Thomas</surname><given-names>Y.</given-names></name> <name><surname>Surette</surname><given-names>S.</given-names></name> <name><surname>Cadieu</surname><given-names>C.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Utility of a deep-learning algorithm to guide novices to acquire echocardiograms for limited diagnostic use</article-title>. <source>JAMA Cardiol.</source> <volume>6</volume>, <fpage>624</fpage>&#x2013;<lpage>632</lpage>. doi: <pub-id pub-id-type="doi">10.1001/jamacardio.2021.0185</pub-id>, <pub-id pub-id-type="pmid">33599681</pub-id></mixed-citation></ref>
<ref id="ref37"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>&#x00D6;hman</surname><given-names>U.</given-names></name> <name><surname>Em&#x00E5;s</surname><given-names>S.</given-names></name> <name><surname>Rubio</surname><given-names>C.</given-names></name></person-group> (<year>1980</year>). <article-title>Relation between early and advanced gastric cancer</article-title>. <source>Am. J. Surg.</source> <volume>140</volume>, <fpage>351</fpage>&#x2013;<lpage>355</lpage>. doi: <pub-id pub-id-type="doi">10.1016/0002-9610(80)90166-X</pub-id>, <pub-id pub-id-type="pmid">6158879</pub-id></mixed-citation></ref>
<ref id="ref38"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Olawuyi</surname><given-names>O.</given-names></name> <name><surname>Viriri</surname><given-names>S.</given-names></name></person-group> (<year>2025</year>). <article-title>Deep learning techniques for prostate cancer analysis and detection: survey of the state of the art</article-title>. <source>J. Imaging</source> <volume>11</volume>:<fpage>254</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jimaging11080254</pub-id>, <pub-id pub-id-type="pmid">40863464</pub-id></mixed-citation></ref>
<ref id="ref39"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pei</surname><given-names>Y.</given-names></name> <name><surname>Luo</surname><given-names>Z.</given-names></name> <name><surname>Yan</surname><given-names>Y.</given-names></name> <name><surname>Yan</surname><given-names>H.</given-names></name> <name><surname>Jiang</surname><given-names>J.</given-names></name> <name><surname>Li</surname><given-names>W.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Data augmentation: using channel-level recombination to improve classification performance for motor imagery EEG</article-title>. <source>Front. Hum. Neurosci.</source> <volume>15</volume>:<fpage>645952</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnhum.2021.645952</pub-id>, <pub-id pub-id-type="pmid">33776673</pub-id></mixed-citation></ref>
<ref id="ref40"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Qin</surname><given-names>Y.</given-names></name> <name><surname>Chang</surname><given-names>J.</given-names></name> <name><surname>Li</surname><given-names>L.</given-names></name> <name><surname>Wu</surname><given-names>M.</given-names></name></person-group> (<year>2025</year>). <article-title>Enhancing gastroenterology with multimodal learning: the role of large language model chatbots in digestive endoscopy</article-title>. <source>Front. Med.</source> <volume>12</volume>:<fpage>1583514</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmed.2025.1583514</pub-id>, <pub-id pub-id-type="pmid">40470039</pub-id></mixed-citation></ref>
<ref id="ref41"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sakai</surname><given-names>Y.</given-names></name> <name><surname>Takemoto</surname><given-names>S.</given-names></name> <name><surname>Hori</surname><given-names>K.</given-names></name> <name><surname>Nishimura</surname><given-names>M.</given-names></name> <name><surname>Ikematsu</surname><given-names>H.</given-names></name> <name><surname>Yano</surname><given-names>T.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network</article-title>. <source>Conf. Proc. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. IEEE Eng. Med. Biol. Soc. Annu. Conf</source> <volume>2018</volume>, <fpage>4138</fpage>&#x2013;<lpage>4141</lpage>. doi: <pub-id pub-id-type="doi">10.1109/EMBC.2018.8513274</pub-id></mixed-citation></ref>
<ref id="ref42"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Singla</surname><given-names>N.</given-names></name> <name><surname>Ghosh</surname><given-names>A.</given-names></name> <name><surname>Dhingra</surname><given-names>M.</given-names></name> <name><surname>Pal</surname><given-names>U.</given-names></name> <name><surname>Dasgupta</surname><given-names>A.</given-names></name> <name><surname>Ghosh</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>A pilot study of breast cancer histopathological image classification using Google teachable machine: a no-code artificial intelligence approach</article-title>. <source>Cureus</source> <volume>17</volume>:<fpage>e87301</fpage>. doi: <pub-id pub-id-type="doi">10.7759/cureus.87301</pub-id>, <pub-id pub-id-type="pmid">40761997</pub-id></mixed-citation></ref>
<ref id="ref43"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sung</surname><given-names>H.</given-names></name> <name><surname>Ferlay</surname><given-names>J.</given-names></name> <name><surname>Siegel</surname><given-names>R. L.</given-names></name> <name><surname>Laversanne</surname><given-names>M.</given-names></name> <name><surname>Soerjomataram</surname><given-names>I.</given-names></name> <name><surname>Jemal</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries</article-title>. <source>CA Cancer J. Clin.</source> <volume>71</volume>, <fpage>209</fpage>&#x2013;<lpage>249</lpage>. doi: <pub-id pub-id-type="doi">10.3322/caac.21660</pub-id>, <pub-id pub-id-type="pmid">33538338</pub-id></mixed-citation></ref>
<ref id="ref44"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tajbakhsh</surname><given-names>N.</given-names></name> <name><surname>Shin</surname><given-names>J. Y.</given-names></name> <name><surname>Gurudu</surname><given-names>S. R.</given-names></name> <name><surname>Hurst</surname><given-names>R. T.</given-names></name> <name><surname>Kendall</surname><given-names>C. B.</given-names></name> <name><surname>Gotway</surname><given-names>M. B.</given-names></name> <etal/></person-group>. (<year>2016</year>). <article-title>Convolutional neural networks for medical image analysis: full training or fine tuning?</article-title> <source>IEEE Trans. Med. Imaging</source> <volume>35</volume>, <fpage>1299</fpage>&#x2013;<lpage>1312</lpage>. doi: <pub-id pub-id-type="doi">10.1109/TMI.2016.2535302</pub-id>, <pub-id pub-id-type="pmid">26978662</pub-id></mixed-citation></ref>
<ref id="ref45"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Takemoto</surname><given-names>S.</given-names></name> <name><surname>Hori</surname><given-names>K.</given-names></name> <name><surname>Yoshimasa</surname><given-names>S.</given-names></name> <name><surname>Nishimura</surname><given-names>M.</given-names></name> <name><surname>Nakajo</surname><given-names>K.</given-names></name> <name><surname>Inaba</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Computer-aided demarcation of early gastric cancer: a pilot comparative study with endoscopists</article-title>. <source>J. Gastroenterol.</source> <volume>58</volume>, <fpage>741</fpage>&#x2013;<lpage>750</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00535-023-02001-x</pub-id></mixed-citation></ref>
<ref id="ref46"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tang</surname><given-names>D.</given-names></name> <name><surname>Wang</surname><given-names>L.</given-names></name> <name><surname>Ling</surname><given-names>T.</given-names></name> <name><surname>Lv</surname><given-names>Y.</given-names></name> <name><surname>Ni</surname><given-names>M.</given-names></name> <name><surname>Zhan</surname><given-names>Q.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Development and validation of a real-time artificial intelligence-assisted system for detecting early gastric cancer: a multicentre retrospective diagnostic study</article-title>. <source>EBioMedicine</source> <volume>62</volume>:<fpage>103146</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ebiom.2020.103146</pub-id></mixed-citation></ref>
<ref id="ref47"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Teramoto</surname><given-names>A.</given-names></name> <name><surname>Shibata</surname><given-names>T.</given-names></name> <name><surname>Yamada</surname><given-names>H.</given-names></name> <name><surname>Hirooka</surname><given-names>Y.</given-names></name> <name><surname>Saito</surname><given-names>K.</given-names></name> <name><surname>Fujita</surname><given-names>H.</given-names></name></person-group> (<year>2022</year>). <article-title>Detection and characterization of gastric cancer using cascade deep learning model in endoscopic images</article-title>. <source>Diagnostics</source> <volume>12</volume>:<fpage>1996</fpage>. doi: <pub-id pub-id-type="doi">10.3390/diagnostics12081996</pub-id>, <pub-id pub-id-type="pmid">36010346</pub-id></mixed-citation></ref>
<ref id="ref48"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thalakottor</surname><given-names>L. A.</given-names></name> <name><surname>Shirwaikar</surname><given-names>R. D.</given-names></name> <name><surname>Pothamsetti</surname><given-names>P. T.</given-names></name> <name><surname>Mathews</surname><given-names>L. M.</given-names></name></person-group> (<year>2023</year>). <article-title>Classification of histopathological images from breast cancer patients using deep learning: a comparative analysis</article-title>. <source>Crit. Rev. Biomed. Eng.</source> <volume>51</volume>, <fpage>41</fpage>&#x2013;<lpage>62</lpage>. doi: <pub-id pub-id-type="doi">10.1615/CritRevBiomedEng.2023047793</pub-id>, <pub-id pub-id-type="pmid">37581350</pub-id></mixed-citation></ref>
<ref id="ref49"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tong</surname><given-names>Z.</given-names></name> <name><surname>Wang</surname><given-names>Y.</given-names></name> <name><surname>Bao</surname><given-names>X.</given-names></name> <name><surname>Deng</surname><given-names>Y.</given-names></name> <name><surname>Lin</surname><given-names>B.</given-names></name> <name><surname>Su</surname><given-names>G.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Development of a whole-slide-level segmentation-based dMMR/pMMR deep learning detector for colorectal cancer</article-title>. <source>iScience</source> <volume>26</volume>:<fpage>108468</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.isci.2023.108468</pub-id>, <pub-id pub-id-type="pmid">38077136</pub-id></mixed-citation></ref>
<ref id="ref50"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ul Haq</surname><given-names>E.</given-names></name> <name><surname>Yong</surname><given-names>Q.</given-names></name> <name><surname>Yuan</surname><given-names>Z.</given-names></name> <name><surname>Jianjun</surname><given-names>H.</given-names></name> <name><surname>Ul Haq</surname><given-names>R.</given-names></name> <name><surname>Qin</surname><given-names>X.</given-names></name></person-group> (<year>2024</year>). <article-title>Accurate multiclassification and segmentation of gastric cancer based on a hybrid cascaded deep learning model with a vision transformer from endoscopic images</article-title>. <source>Inf. Sci.</source> <volume>670</volume>:<fpage>120568</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ins.2024.120568</pub-id></mixed-citation></ref>
<ref id="ref51"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>van Houwelingen</surname><given-names>H. C.</given-names></name> <name><surname>Arends</surname><given-names>L. R.</given-names></name> <name><surname>Stijnen</surname><given-names>T.</given-names></name></person-group> (<year>2002</year>). <article-title>Advanced methods in meta-analysis: multivariate approach and meta-regression</article-title>. <source>Stat. Med.</source> <volume>21</volume>, <fpage>589</fpage>&#x2013;<lpage>624</lpage>. doi: <pub-id pub-id-type="doi">10.1002/sim.1040</pub-id>, <pub-id pub-id-type="pmid">11836738</pub-id></mixed-citation></ref>
<ref id="ref52"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>S.</given-names></name> <name><surname>Wang</surname><given-names>T.</given-names></name> <name><surname>Yang</surname><given-names>L.</given-names></name> <name><surname>Yang</surname><given-names>D. M.</given-names></name> <name><surname>Fujimoto</surname><given-names>J.</given-names></name> <name><surname>Yi</surname><given-names>F.</given-names></name> <etal/></person-group>. (<year>2019a</year>). <article-title>ConvPath: a software tool for lung adenocarcinoma digital pathological image analysis aided by a convolutional neural network</article-title>. <source>EBioMedicine</source> <volume>50</volume>, <fpage>103</fpage>&#x2013;<lpage>110</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ebiom.2019.10.033</pub-id>, <pub-id pub-id-type="pmid">31767541</pub-id></mixed-citation></ref>
<ref id="ref53"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>S.</given-names></name> <name><surname>Yang</surname><given-names>D. M.</given-names></name> <name><surname>Rong</surname><given-names>R.</given-names></name> <name><surname>Zhan</surname><given-names>X.</given-names></name> <name><surname>Fujimoto</surname><given-names>J.</given-names></name> <name><surname>Liu</surname><given-names>H.</given-names></name> <etal/></person-group>. (<year>2019b</year>). <article-title>Artificial intelligence in lung cancer pathology image analysis</article-title>. <source>Cancer</source> <volume>11</volume>:<fpage>1673</fpage>. doi: <pub-id pub-id-type="doi">10.3390/cancers11111673</pub-id>, <pub-id pub-id-type="pmid">31661863</pub-id></mixed-citation></ref>
<ref id="ref54"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Whiting</surname><given-names>P. F.</given-names></name> <name><surname>Rutjes</surname><given-names>A. W. S.</given-names></name> <name><surname>Westwood</surname><given-names>M. E.</given-names></name> <name><surname>Mallett</surname><given-names>S.</given-names></name> <name><surname>Deeks</surname><given-names>J. J.</given-names></name> <name><surname>Reitsma</surname><given-names>J. B.</given-names></name> <etal/></person-group>. (<year>2011</year>). <article-title>QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies</article-title>. <source>Ann. Intern. Med.</source> <volume>155</volume>, <fpage>529</fpage>&#x2013;<lpage>536</lpage>. doi: <pub-id pub-id-type="doi">10.7326/0003-4819-155-8-201110180-00009</pub-id>, <pub-id pub-id-type="pmid">22007046</pub-id></mixed-citation></ref>
<ref id="ref55"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>K.</given-names></name> <name><surname>Lu</surname><given-names>L.</given-names></name> <name><surname>Liu</surname><given-names>H.</given-names></name> <name><surname>Wang</surname><given-names>X.</given-names></name> <name><surname>Gao</surname><given-names>Y.</given-names></name> <name><surname>Yang</surname><given-names>L.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>A comprehensive update on early gastric cancer: defining terms, etiology, and alarming risk factors</article-title>. <source>Expert Rev. Gastroenterol. Hepatol.</source> <volume>15</volume>, <fpage>255</fpage>&#x2013;<lpage>273</lpage>. doi: <pub-id pub-id-type="doi">10.1080/17474124.2021.1845140</pub-id>, <pub-id pub-id-type="pmid">33121300</pub-id></mixed-citation></ref>
<ref id="ref56"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yuan</surname><given-names>X. L.</given-names></name> <name><surname>Zhou</surname><given-names>Y.</given-names></name> <name><surname>Liu</surname><given-names>W.</given-names></name> <name><surname>Luo</surname><given-names>Q.</given-names></name> <name><surname>Zeng</surname><given-names>X. H.</given-names></name> <name><surname>Yi</surname><given-names>Z.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Artificial intelligence for diagnosing gastric lesions under white-light endoscopy</article-title>. <source>Surg. Endosc.</source> <volume>36</volume>, <fpage>9444</fpage>&#x2013;<lpage>9453</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00464-022-09420-6</pub-id></mixed-citation></ref>
<ref id="ref57"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>J.</given-names></name> <name><surname>Guo</surname><given-names>S.-B.</given-names></name> <name><surname>Duan</surname><given-names>Z.-J.</given-names></name></person-group> (<year>2011</year>). <article-title>Application of magnifying narrow-band imaging endoscopy for diagnosis of early gastric cancer and precancerous lesion</article-title>. <source>BMC Gastroenterol.</source> <volume>11</volume>:<fpage>135</fpage>. doi: <pub-id pub-id-type="doi">10.1186/1471-230X-11-135</pub-id>, <pub-id pub-id-type="pmid">22168239</pub-id></mixed-citation></ref>
<ref id="ref58"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>L.</given-names></name> <name><surname>Lu</surname><given-names>Z.</given-names></name> <name><surname>Yao</surname><given-names>L.</given-names></name> <name><surname>Dong</surname><given-names>Z.</given-names></name> <name><surname>Zhou</surname><given-names>W.</given-names></name> <name><surname>He</surname><given-names>C.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Effect of a deep learning&#x2013;based automatic upper GI endoscopic reporting system: a randomized crossover study (with video)</article-title>. <source>Gastrointest. Endosc.</source> <volume>98</volume>, <fpage>181</fpage>&#x2013;<lpage>190.e10</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.gie.2023.02.025</pub-id>, <pub-id pub-id-type="pmid">36849056</pub-id></mixed-citation></ref>
<ref id="ref59"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>K.</given-names></name> <name><surname>Wang</surname><given-names>H.</given-names></name> <name><surname>Cheng</surname><given-names>Y.</given-names></name> <name><surname>Liu</surname><given-names>H.</given-names></name> <name><surname>Gong</surname><given-names>Q.</given-names></name> <name><surname>Zeng</surname><given-names>Q.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Early gastric cancer detection and lesion segmentation based on deep learning and gastroscopic images</article-title>. <source>Sci. Rep.</source> <volume>14</volume>:<fpage>7847</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-58361-8</pub-id></mixed-citation></ref>
<ref id="ref60"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>L.</given-names></name> <name><surname>Zhang</surname><given-names>Y.</given-names></name> <name><surname>Wang</surname><given-names>L.</given-names></name> <name><surname>Wang</surname><given-names>J.</given-names></name> <name><surname>Liu</surname><given-names>Y.</given-names></name></person-group> (<year>2021</year>). <article-title>Diagnosis of gastric lesions through a deep convolutional neural network</article-title>. <source>Dig. Endosc.</source> <volume>33</volume>, <fpage>788</fpage>&#x2013;<lpage>796</lpage>. doi: <pub-id pub-id-type="doi">10.1111/den.13844</pub-id>, <pub-id pub-id-type="pmid">32961597</pub-id></mixed-citation></ref>
<ref id="ref61"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname><given-names>Z.</given-names></name> <name><surname>Qian</surname><given-names>X.</given-names></name> <name><surname>Hu</surname><given-names>J.</given-names></name> <name><surname>Chen</surname><given-names>G.</given-names></name> <name><surname>Zhang</surname><given-names>C.</given-names></name> <name><surname>Zhu</surname><given-names>J.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>An artificial intelligence-assisted diagnosis modeling software (AIMS) platform based on medical images and machine learning: a development and validation study</article-title>. <source>Quant. Imaging Med. Surg.</source> <volume>13</volume>, <fpage>7504</fpage>&#x2013;<lpage>7522</lpage>. doi: <pub-id pub-id-type="doi">10.21037/qims-23-20</pub-id>, <pub-id pub-id-type="pmid">37969634</pub-id></mixed-citation></ref>
<ref id="ref62"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname><given-names>B.</given-names></name> <name><surname>Rao</surname><given-names>X.</given-names></name> <name><surname>Xing</surname><given-names>H.</given-names></name> <name><surname>Ma</surname><given-names>Y.</given-names></name> <name><surname>Wang</surname><given-names>F.</given-names></name> <name><surname>Rong</surname><given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>A convolutional neural network-based system for detecting early gastric cancer in white-light endoscopy</article-title>. <source>Scand. J. Gastroenterol.</source> <volume>58</volume>, <fpage>157</fpage>&#x2013;<lpage>162</lpage>. doi: <pub-id pub-id-type="doi">10.1080/00365521.2022.2113427</pub-id>, <pub-id pub-id-type="pmid">36000979</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/12494/overview">Thomas Hartung</ext-link>, Johns Hopkins University, United States</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1433626/overview">Kai Zhang</ext-link>, Chongqing Chang&#x2019;an Industrial Co. Ltd, China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2838355/overview">Vaibhav C. Gandhi</ext-link>, Charutar Vidya Mandal University, India</p>
</fn>
</fn-group>
<fn-group>
<fn fn-type="abbr" id="abbrev1">
<label>Abbreviations:</label>
<p>AI, artificial intelligence; AUC, area under curve; CNNs, convolutional neural networks; DL, deep learning; EGC, early gastric cancer; FN, false negative; FP, false positive; GC, Gastric cancer; MeSH, Medical Subject Headings; NBI, narrow-band imaging; QUADAS-2, Quality Assessment of Diagnostic Accuracy Studies-2; SROC, summary receiver operating characteristic; TP, true positive; TN, true negative.</p>
</fn>
</fn-group>
</back>
</article>