<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="review-article" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Radiol.</journal-id><journal-title-group>
<journal-title>Frontiers in Radiology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Radiol.</abbrev-journal-title></journal-title-group>
<issn pub-type="epub">2673-8740</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fradi.2025.1733003</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Applications, image analysis, and interpretation of computer vision in medical imaging</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes"><name><surname>Matsuzaka</surname><given-names>Yasunari</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="cor1">&#x002A;</xref><uri xlink:href="https://loop.frontiersin.org/people/3256693/overview"/><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
<contrib contrib-type="author"><name><surname>Iyoda</surname><given-names>Masayuki</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref><uri xlink:href="https://loop.frontiersin.org/people/1521846/overview" /><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role><role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role></contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Microbiology and Immunology, Showa Medical University Graduate School of Medicine</institution>, <city>Shinagawa-ku</city>, <state>Tokyo</state>, <country country="jp">Japan</country></aff>
<aff id="aff2"><label>2</label><institution>Division of Nephrology, Department of Medicine, Showa Medical University Graduate School of Medicine</institution>, <city>Shinagawa-ku</city>, <state>Tokyo</state>, <country country="jp">Japan</country></aff>
<author-notes>
<corresp id="cor1"><label>&#x002A;</label><bold>Correspondence:</bold> Yasunari Matsuzaka <email xlink:href="mailto:yasunari.matsuzaka@showa-u.ac.jp">yasunari.matsuzaka@showa-u.ac.jp</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-09"><day>09</day><month>01</month><year>2026</year></pub-date>
<pub-date publication-format="electronic" date-type="collection"><year>2025</year></pub-date>
<volume>5</volume><elocation-id>1733003</elocation-id>
<history>
<date date-type="received"><day>27</day><month>10</month><year>2025</year></date>
<date date-type="rev-recd"><day>12</day><month>12</month><year>2025</year></date>
<date date-type="accepted"><day>16</day><month>12</month><year>2025</year></date>
</history>
<permissions>
<copyright-statement>&#x00A9; 2026 Matsuzaka and Iyoda.</copyright-statement>
<copyright-year>2026</copyright-year><copyright-holder>Matsuzaka and Iyoda</copyright-holder><license><ali:license_ref start_date="2026-01-09">https://creativecommons.org/licenses/by/4.0/</ali:license_ref><license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p></license>
</permissions>
<abstract>
<p>This review summarizes the current advances, applications, and research prospects of computer vision in advancing medical imaging. Computer vision in healthcare has revolutionized medical practice by increasing diagnostic accuracy, improving patient care, and increasing operational efficiency. Likewise, deep learning algorithms have advanced medical image analysis, significantly improved healthcare outcomes and transforming diagnostic processes. Specifically, convolutional neural networks are crucial for modern medical image segmentation, enabling the accurate, efficient analysis of various imaging modalities and helping enhance computer-aided diagnosis and treatment planning. Computer vision algorithms have demonstrated remarkable capabilities in detecting various diseases. Artificial intelligence (AI) systems can identify lung nodules in chest computed tomography scans at a sensitivity comparable to that of experienced radiologists. Computer vision can analyze brain scans to detect problems such as aneurysms and tumors or areas affected by diseases such as Alzheimer&#x0027;s. In summary, computer vision in medical imaging is significantly improving diagnostic accuracy, efficiency, and patient outcomes across a range of medical specialties.</p>
</abstract>
<kwd-group>
<kwd>application programming interfaces</kwd>
<kwd>computer vision</kwd>
<kwd>convolutional neural networks</kwd>
<kwd>deep learning</kwd>
<kwd>machine learning</kwd>
</kwd-group><funding-group><funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement></funding-group><counts>
<fig-count count="5"/>
<table-count count="2"/><equation-count count="9"/><ref-count count="296"/><page-count count="33"/><word-count count="1110"/></counts><custom-meta-group><custom-meta><meta-name>section-at-acceptance</meta-name><meta-value>Artificial Intelligence in Radiology</meta-value></custom-meta></custom-meta-group>
</article-meta>
</front>
<body><sec id="s1" sec-type="intro"><label>1</label><title>Introduction</title>
<p>Computer vision engineering is crucial to the advancement of medical imaging technologies and applications (<xref ref-type="bibr" rid="B1">1</xref>). This interdisciplinary field combines computer science, mathematics, and healthcare expertise to develop innovative solutions for analyzing and interpreting medical images (<xref ref-type="bibr" rid="B2">2</xref>). Computer vision in medical imaging encompasses several important tasks. In image classification, deep learning (DL) and convolutional neural networks (CNNs) have significantly improved medical image classification, such as identifying abnormalities in chest x-rays (CXRs) and detecting cancerous lesions (<xref ref-type="bibr" rid="B3">3</xref>). Image segmentation, or the division of medical images into distinct regions or objects of interest, is particularly useful in robotic surgery training and instrument segmentation during procedures. For object detection and recognition, computer vision algorithms can locate and identify specific structures, anomalies, or instruments within medical images, aiding in diagnosis and surgical planning (<xref ref-type="bibr" rid="B4">4</xref>). When selecting computer vision algorithms for a project, it is needed to carefully evaluate several key criteria to ensure the most appropriate solution. A comprehensive breakdown of the main selection criteria included core selection factors, such as task requirements, accuracy vs. speed trade-offs, and computational resources (<xref ref-type="bibr" rid="B5">5</xref>). Algorithm choice depends on three factors: task type, processing speed needs, and hardware constraints. Different algorithms excel at different tasks like classification, detection, segmentation, or feature extraction (<xref ref-type="bibr" rid="B6">6</xref>). Also, YOLO and ORB (Oriented FAST and Rotated BRIEF) deliver real-time performance for speed-critical applications like autonomous driving, while U-Net and Mask R-CNN provide pixel-level precision for medical imaging and segmentation tasks (<xref ref-type="bibr" rid="B7">7</xref>). Further, traditional algorithms (SIFT, SURF, HOG) work without training data on resource-constrained devices (<xref ref-type="bibr" rid="B8">8</xref>). Deep learning models like CNNs require substantial computational power, while lightweight algorithms like ORB are designed for embedded systems and mobile devices. Furthermore, as technical considerations, there are training data requirements, robustness and invariance, and feature complexity. Traditional algorithms like SIFT and edge detection do not require training data, making them suitable when labeled datasets are unavailable (<xref ref-type="bibr" rid="B9">9</xref>). Deep learning approaches need large amounts of annotated data but can achieve higher accuracy for complex tasks (<xref ref-type="bibr" rid="B10">10</xref>). SIFT offers scale and rotation invariance, making it reliable for matching tasks across different viewing conditions. Simple edge detection works well for boundary identification, while complex object detection in cluttered scenes may require sophisticated deep learning architectures like Faster R-CNN or YOLO (<xref ref-type="bibr" rid="B11">11</xref>). Moreover, as practical deployment factors, there are real-time requirements, interpretability, model size and memory, and development resources (<xref ref-type="bibr" rid="B12">12</xref>). If application needs immediate processing (autonomous vehicles, live video analysis), prioritize algorithms optimized for speed even if they sacrifice some accuracy. Some applications require understanding why the algorithm made certain decisions. Traditional methods offer more interpretable results, while deep learning models can be &#x201C;black boxes&#x201D; that are harder to explain. For deployment on edge devices or mobile platforms, consider the model&#x0027;s memory footprint and whether it can run efficiently without cloud connectivity (<xref ref-type="bibr" rid="B13">13</xref>). Evaluate the availability of pre-trained models, frameworks, and community support. Modern deep learning frameworks offer transfer learning capabilities that can significantly reduce development time (<xref ref-type="bibr" rid="B14">14</xref>). The optimal algorithm choice involves balancing these criteria based on specific use case, constraints, and priorities. Many modern applications use hybrid approaches that combine traditional and deep learning methods to leverage the strengths of both (<xref ref-type="bibr" rid="B15">15</xref>).</p>
<p>Recent developments in computer vision have greatly enhanced the capabilities of medical imaging. The use of DL techniques, particularly CNNs, has revolutionized medical image analysis, improving the processing accuracy and efficiency of complex visual data (<xref ref-type="bibr" rid="B16">16</xref>). Advanced algorithms have upgraded the interpretation of 3D medical imaging data, enhancing depth perception and spatial understanding in applications such as computed tomography (CT) and magnetic resonance imaging (MRI) scans (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B18">18</xref>). Optimized algorithms and hardware acceleration are used to analyze medical imaging data in real time, which is crucial for applications such as live surgical guidance (<xref ref-type="bibr" rid="B19">19</xref>).</p>
<p>Despite this significant progress, several challenges remain in this area. Understanding these challenges is essential for developing robust, trustworthy, and clinically valuable AI systems. Medical image labeling is expensive, time-consuming, and requires expert participation from physicians, radiologists, and specialists. Unlike natural image analysis with large-scale labeled datasets such as ImageNet, medical image analysis faces a major challenge of lacking labeled data to construct reliable and robust models (<xref ref-type="bibr" rid="B20">20</xref>). Image quality, resolution, artifacts, and variability across imaging devices and protocols pose challenges for standardized workflows (<xref ref-type="bibr" rid="B21">21</xref>). Domain shift is widespread among different medical image datasets due to different scanners, scanning parameters, and subject cohorts (<xref ref-type="bibr" rid="B22">22</xref>). Existing visual backbones lack appropriate priority for reliable generalization in medical settings, with models showing over 63&#x0025; average performance drop when introducing confounds (<xref ref-type="bibr" rid="B23">23</xref>). Five core elements define interpretability: localization, visual recognizability, physical attribution, model transparency, and actionability (<xref ref-type="bibr" rid="B24">24</xref>). Deep learning models are criticized for their &#x201C;black-box&#x201D; nature which undermines clinicians&#x0027; trust in high-stakes healthcare domains (<xref ref-type="bibr" rid="B25">25</xref>). The direct impact of medical image analysis on patient care prioritizes accuracy and reliability, with severe consequences of adversarial attacks introducing profound ethical considerations (<xref ref-type="bibr" rid="B26">26</xref>). VGG16 achieved 96&#x0025; accuracy on clean MRI data but dropped to 32&#x0025; under FGSM attacks and 13&#x0025; under PGD attacks (<xref ref-type="bibr" rid="B27">27</xref>). Foundation models consistently underdiagnose marginalized groups, with even higher rates in intersectional subgroups such as Black female patients.</p>
<p>Public datasets are available, but more comprehensive and diverse medical imaging datasets are needed to train robust models (<xref ref-type="bibr" rid="B20">20</xref>). Moreover, despite the potential of computer vision, the use of such applications in frontline healthcare settings is limited, indicating a research&#x2013;implementation gap (<xref ref-type="bibr" rid="B28">28</xref>). Computer scientists, medical professionals, and healthcare institutions should closely collaborate to develop clinically relevant and applicable solutions and advance this field. Thus, computer vision in medical imaging is rapidly evolving and has immense potential to improve healthcare outcomes. As technologies and interdisciplinary collaborations strengthen, more innovative applications will improve diagnosis, treatment planning, and patient care across medical specialties. This review summarizes the current advances, applications, and prospects of computer vision in advancing medical imaging technologies.</p>
</sec>
<sec id="s2"><label>2</label><title>Transformation of medical image analysis via DL</title>
<p>DL is revolutionizing medical image analysis by providing powerful tools that improve diagnosis, treatment planning, and patient care across multiple medical specialties (<xref ref-type="bibr" rid="B29">29</xref>). This transformation is evident in several key areas. In automated analysis and detection, DL algorithms, particularly CNNs, can remarkably identify and categorize anomalies in medical images automatically (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B30">30</xref>). These algorithms can analyze various types of medical images, including x-rays, MRI scans, CT scans, and ultrasound images, providing healthcare professionals with fast, accurate insights. In addition, to improve accuracy and efficiency, DL models leverage large amounts of annotated data to learn complex patterns and relationships within medical images, facilitating accurate detection, localization, and diagnosis of diseases and abnormalities (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B31">31</xref>). This capability enables faster, more accurate interpretation of medical images, thereby improving patient outcomes and healthcare workflow efficiency.</p>
<p>DL techniques have numerous applications in medical image analysis, such as in medical image segmentation, which is crucial for tasks such as tumor delineation; locating and identifying specific structures, anomalies, or instruments in medical images (object detection); accurately categorizing various medical conditions based on image analysis (disease classification), and improving image quality or reconstructing images from limited data (image reconstruction) (<xref ref-type="fig" rid="F1">Figure&#x00A0;1</xref>) (<xref ref-type="bibr" rid="B32">32</xref>). DL algorithms are valuable tools for healthcare professionals, assisting in early disease detection, decision support for radiologists, assessment of disease progression and treatment response, and personalized treatment planning (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B33">33</xref>).</p>
<fig id="F1" position="float"><label>Figure&#x00A0;1</label>
<caption><p>Deep learning-based segmentation algorithm. Three algorithms of the segmentation, semantic instance, and panoptic are included.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fradi-05-1733003-g001.tif"><alt-text content-type="machine-generated">Flowchart of deep learning-based segmentation algorithms. It categorizes into three types: Semantic Segmentation, Instance Segmentation, and Panoptic Segmentation. Semantic Segmentation includes methods like CNN + CRF, FCN (VGG16), U-Net, and DeepLab V3. Instance Segmentation features DeepMask, Instance-Aware, Mask R-CNN, and CenterMask. Panoptic Segmentation involves Panoptic FPN, DetectoRS, UPS-Net, and Panoptic Deeplab.</alt-text>
</graphic>
</fig>
<p>The use of DL for medical image analysis is evolving, with ongoing research focusing on improving model interpretability and explainability, developing more robust and generalizable algorithms, integrating multimodal data for comprehensive analyses, and addressing privacy and security challenges (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B34">34</xref>). As these advances continue, DL will be increasingly important in transforming medical imaging and healthcare delivery, ultimately leading to more personalized, accurate, and efficient patient care (<xref ref-type="bibr" rid="B29">29</xref>).</p>
</sec>
<sec id="s3"><label>3</label><title>Main challenges in deploying computer vision in healthcare</title>
<sec id="s3a"><label>3.1</label><title>Data-related challenges</title>
<p>Obtaining high-quality, diverse representative medical imaging datasets is difficult due to ethical, legal, and logistical issues (<xref ref-type="bibr" rid="B35">35</xref>). Rare diseases are often underrepresented, hindering the training of robust models. In terms of privacy and security, healthcare data are highly sensitive, exposed to high risks of data breaches and misuse (<xref ref-type="bibr" rid="B36">36</xref>). Strong data protection and compliance with regulations, such as the Health Insurance Portability and Accountability Act (HIPAA) and the General Data Protection Regulation (GDPR), are essential to protect personal data and maintain trust in the digital age (<xref ref-type="bibr" rid="B37">37</xref>). These regulations provide comprehensive frameworks for handling sensitive data, with HIPAA focusing on healthcare information in the United States and GDPR focusing on personal data protection in the European Union (EU). Both emphasize several key data protection principles: Data must be processed lawfully, fairly, and transparently (lawfulness, fairness, and transparency): data should only be collected for specific, legitimate purposes (data minimization); personal data should be kept accurate and up to date (accuracy); data should only be kept for as long as necessary (data retention); and appropriate data security measures must be taken (integrity and confidentiality).</p>
<p>Organizations should comply with relevant regulations by implementing robust access controls and authentication mechanisms, using encryption for sensitive data storage and transmission, and conducting regular risk assessments and security audits. They should also provide comprehensive data protection training to their staff and establish clear data handling and breach notification policies. By adhering to these principles and strengthening their data protection measures, organizations can ensure compliance with HIPAA and GDPR, protect individuals&#x0027; privacy rights, and mitigate the risks associated with data breaches and unauthorized data access.</p>
<p>As for important regulatory philosophy differences, philosophy in United States (FDA) is risk-based pragmatism with innovation support, such as flexible pathways based on device risk and novelty, substantial equivalence concept (510(k)) enables iterative innovation, post-market surveillance increasingly data-driven (Real-World Evidence), PCCP framework represents adaptive regulation acknowledging AI&#x0027;s evolving nature, and &#x201C;Light touch&#x201D; for lower-risk devices; rigorous review for breakthrough technologies (<xref ref-type="table" rid="T1">Table&#x00A0;1</xref>) (<xref ref-type="bibr" rid="B38">38</xref>). In European Union, it is precautionary principle with fundamental rights protection, including comprehensive, cross-sectoral approach (MDR&#x2009;&#x002B;&#x2009;AI Act&#x2009;&#x002B;&#x2009;GDPR), high-risk AI systems subject to strict global requirements, human rights and ethical considerations central to framework, transparency and explainability legally mandated, and consumer protection paramount, even if slowing market entry. In Japan, philosophy is quality-first approach with measured innovation adoption thorough validation preferred over rapid deployment, cultural emphasis on safety and meticulous review, strong alignment with international standards (ISO, IMDRF), and preference for evidence from Japanese populations reflects local validation focus. Recent acceleration efforts (DASH for SaMD 2) show commitment to competitiveness.</p>
<table-wrap id="T1" position="float"><label>Table&#x00A0;1</label>
<caption><p>Medical image AI deployment: regulatory comparison (US vs. EU vs. Japan).</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Regulatory aspect</th>
<th valign="top" align="center">United States (FDA)</th>
<th valign="top" align="center">European Union (MDR/IVDR&#x2009;&#x002B;&#x2009;AI Act)</th>
<th valign="top" align="center">Japan (PMDA)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Primary regulatory bodies</td>
<td valign="top" align="left">Food and Drug Administration (FDA), Center for Devices and Radiological Health (CDRH)</td>
<td valign="top" align="left">European Commission, National Competent Authorities, Notified Bodies</td>
<td valign="top" align="left">Ministry of Health, Labour and Welfare (MHLW), Pharmaceuticals and Medical Devices Agency (PMDA)</td>
</tr>
<tr>
<td valign="top" align="left">Key legislation</td>
<td valign="top" align="left">Federal Food, Drug, and Cosmetic Act (FD&#x0026;C Act); 21st Century Cures Act (2016)</td>
<td valign="top" align="left">Medical Devices Regulation (EU 2017/745), AI Act (EU 2024/1689), GDPR</td>
<td valign="top" align="left">Pharmaceuticals and Medical Devices Act (PMD Act, 2014)</td>
</tr>
<tr>
<td valign="top" align="left">Classification system</td>
<td valign="top" align="left">Risk-based: Class I (low), II (moderate), III (high risk)</td>
<td valign="top" align="left">Risk-based: Class I, IIa, IIb, III (Medical Devices); Minimal, Limited, High-risk, Unacceptable (AI Act)</td>
<td valign="top" align="left">Risk-based: Class I (General), II (Controlled), III (Highly Controlled), IV (Highly Controlled)</td>
</tr>
<tr>
<td valign="top" align="left">Approval pathways</td>
<td valign="top" align="left">510(k) (predicate device comparison), <italic>de novo</italic> (novel low-to-moderate risk), PMA (Premarket Approval for high-risk), Breakthrough Devices Program</td>
<td valign="top" align="left">CE Marking via: Self-certification (Class I), Notified Body assessment (IIa&#x2013;III); Combined conformity assessment under MDR/IVDR and AI Act for high-risk AI</td>
<td valign="top" align="left">Todokede (notification for Class I), Ninsho (certification for Class II/III), Shonin (approval for Class III/IV)</td>
</tr>
<tr>
<td valign="top" align="left">AI-specific framework</td>
<td valign="top" align="left">Predetermined Change Control Plan (PCCP) finalized December 2024; allows pre-approved algorithm modifications without new submissions</td>
<td valign="top" align="left">AI Act (effective August 2024, full implementation by August 2027); high-risk AI systems require strict compliance including human oversight, transparency, data governance</td>
<td valign="top" align="left">Post-Approval Change Management Protocol (PACMP, March 2023); DASH for SaMD 2 strategy; Two-stage approval system for SaMD</td>
</tr>
<tr>
<td valign="top" align="left">Typical approval timeline</td>
<td valign="top" align="left">510(k): 3&#x2013;6 months <italic>de novo</italic>: 6&#x2013;12 months PMA: 12&#x2013;18&#x002B; months Most AI devices use 510(k) pathway</td>
<td valign="top" align="left">Class IIa: 6&#x2013;12 months Class IIb/III: 12&#x2013;24&#x002B; months Current bottleneck: Notified Body capacity constraints</td>
<td valign="top" align="left">Class I: 1&#x2013;2 months Class II/III: 6&#x2013;12 months Class IV: 12&#x2013;18 months SaMD Priority Review: 6 months (target from 2024)</td>
</tr>
<tr>
<td valign="top" align="left">Approved AI devices</td>
<td valign="top" align="left">1,250&#x002B; AI-enabled devices (as of July 2025); &#x223C;712 in radiology; exponential growth from 6 (2015) to 223 (2023)</td>
<td valign="top" align="left">Thousands approved under MDR; exact numbers vary by member state; most radiology AI devices Class IIa or higher</td>
<td valign="top" align="left">Limited compared to US/EU; only 3 therapeutic apps approved by Sept 2023 vs. 50&#x002B; in US/EU; 15&#x0025; increase in AI imaging device approvals 2018&#x2013;2023</td>
</tr>
<tr>
<td valign="top" align="left">Post-market surveillance</td>
<td valign="top" align="left">Medical Device Reporting (MDR); Real-World Evidence (RWE) emphasis; Performance monitoring required under PCCP</td>
<td valign="top" align="left">Stringent post-market surveillance under MDR; Vigilance reporting; AI Act requires continuous monitoring of high-risk systems</td>
<td valign="top" align="left">Good Vigilance Practice (GVP) Ordinance; Safety management measures; Enhanced monitoring for SaMD updates</td>
</tr>
<tr>
<td valign="top" align="left">Algorithm update management</td>
<td valign="top" align="left">PCCP Framework (2024): - Pre-approved modifications without new submission - Must follow exact protocol - QMS documentation required - Deviations require new submission</td>
<td valign="top" align="left">AI Act Requirements: - Predefined change protocols - Continuous oversight required - Must maintain conformity throughout TPLC - Combined MDR/AI Act assessment</td>
<td valign="top" align="left">PACMP (2023): - Predefined parameters for updates - Risk-mitigated modifications - Post-approval within safety boundaries - 30-day review for IDATEN system changes</td>
</tr>
<tr>
<td valign="top" align="left">Data requirements</td>
<td valign="top" align="left">Clinical validation required; increasing acceptance of RWE; bridging studies may be accepted for global data</td>
<td valign="top" align="left">High-quality datasets mandated under AI Act; GDPR compliance required; data from EU populations often preferred</td>
<td valign="top" align="left">Japanese population data often required; bridging studies from global trials accepted; stricter requirements for novel devices</td>
</tr>
<tr>
<td valign="top" align="left">Transparency &#x0026; explainability</td>
<td valign="top" align="left">Transparency Guidance (June 2024): - Human-centered design principles - User interface clarity - Model description in submissions - Not mandated but strongly recommended</td>
<td valign="top" align="left">AI Act Mandates: - High transparency for high-risk systems - Explainability requirements - Fundamental rights considerations - Documentation of AI decision-making process</td>
<td valign="top" align="left">Transparency requirements aligned with international standards (JIS T 62366-1:2022); Human factors engineering required as of April 2024</td>
</tr>
<tr>
<td valign="top" align="left">Data privacy framework</td>
<td valign="top" align="left">HIPAA (sector-specific): - Applies to covered entities only - Permits data sharing for treatment/payment - Separate state privacy laws - No comprehensive federal AI privacy law</td>
<td valign="top" align="left">GDPR (cross-sectoral): - Comprehensive data protection - Applies to all health data - Strict consent requirements - Right to explanation - Data minimization principles</td>
<td valign="top" align="left">Act on Protection of Personal Information (APPI): - Similar to GDPR but less stringent - Special provisions for sensitive medical data - Focus on appropriate data handling</td>
</tr>
<tr>
<td valign="top" align="left">Liability framework</td>
<td valign="top" align="left">Mixed liability model: - Product liability (Restatement Third of Torts) - Medical malpractice for physicians - Manufacturer responsibility unclear for adaptive AI - Case-by-case determination</td>
<td valign="top" align="left">Strict liability model: - Product Liability Directive (85/374/EEC) under revision - New AI Liability Directive (2024/2853) - No-fault product liability - Burden of proof on manufacturer - Penalties under AI Act</td>
<td valign="top" align="left">Mixed liability model: - Product Liability Law - Medical malpractice framework - Manufacturer accountability for defects - Healthcare provider responsibility for clinical use</td>
</tr>
<tr>
<td valign="top" align="left">Human oversight requirements</td>
<td valign="top" align="left">Recommended but not mandated; clinical decision support software must allow independent physician review</td>
<td valign="top" align="left">AI Act Mandates: - Human oversight required for high-risk AI - Cannot fully replace human judgment - Override capability necessary - Recognized as risk mitigation factor</td>
<td valign="top" align="left">Encouraged through human factors engineering requirements; physician final decision-making authority maintained</td>
</tr>
<tr>
<td valign="top" align="left">Documentation language</td>
<td valign="top" align="left">English</td>
<td valign="top" align="left">Native languages of member states&#x2009;&#x002B;&#x2009;English increasingly accepted</td>
<td valign="top" align="left">Japanese required (all documentation); English submissions under pilot for certain applications (as of Sept 2024)</td>
</tr>
<tr>
<td valign="top" align="left">International harmonization</td>
<td valign="top" align="left">Participates in IMDRF (International Medical Device Regulators Forum); collaborative guidance with UK MHRA and Health Canada</td>
<td valign="top" align="left">Leader in international AI governance; IMDRF participation; AI Act influences global standards</td>
<td valign="top" align="left">Active IMDRF participant; aligns with ICH, ICMRA, MDSAP; ISO 13485 compliance</td>
</tr>
<tr>
<td valign="top" align="left">Innovation support mechanisms</td>
<td valign="top" align="left">- Breakthrough Devices Program (accelerated review) - Pre-submission meetings - Q-Submission program - Real-World Evidence pilots</td>
<td valign="top" align="left">- Innovation support via EU funds - Regulatory sandboxes (member state dependent) - SME support initiatives - Notified Body guidance</td>
<td valign="top" align="left">- DASH for SaMD 2 program - Priority review for SaMD - Two-stage approval system - Pre-submission consultations - Expanded review team for SaMD</td>
</tr>
<tr>
<td valign="top" align="left">Key challenges</td>
<td valign="top" align="left">- 510(k) predicate pathway complexity - Unclear guidance for truly novel AI - Time-intensive for complex devices - Post-market surveillance requirements</td>
<td valign="top" align="left">- Notified Body capacity constraints - Dual compliance (MDR&#x2009;&#x002B;&#x2009;AI Act) - Regulatory fragmentation across member states - Lengthy certification timelines - High compliance costs</td>
<td valign="top" align="left">- Language barrier (Japanese documentation) - Japanese population data requirements - Slower adoption than US/EU - Limited number of approved SaMD - Rigorous QMS requirements</td>
</tr>
<tr>
<td valign="top" align="left">Deployment timeline impact</td>
<td valign="top" align="left">Fast-to-Market: - 510(k) enables rapid market entry - Established predicate pathways - Largest approved device database - Iterative updates via PCCP</td>
<td valign="top" align="left">Moderate Pace: - CE marking timeframe variable - Notified Body availability critical - Dual assessment (MDR&#x2009;&#x002B;&#x2009;AI Act) adds complexity - Single market access advantage</td>
<td valign="top" align="left">Deliberate Pace: - Focus on quality over speed - Comprehensive review process - SaMD priority review improving timelines - Cultural emphasis on thorough validation</td>
</tr>
<tr>
<td valign="top" align="left">Clinical trial requirements</td>
<td valign="top" align="left">- Pivotal clinical trials for PMA pathway - May accept foreign clinical data - Good Clinical Practice (GCP) compliance - IDE required for investigational devices</td>
<td valign="top" align="left">- Clinical evaluation required under MDR - Clinical Trials Regulation (CTR) - GCP compliance mandatory - May require EU-specific data</td>
<td valign="top" align="left">- GCP compliance required - Often requires Japanese population data - Bridging studies common - Clinical data from outside Japan may be accepted with justification</td>
</tr>
<tr>
<td valign="top" align="left">Bias &#x0026; fairness requirements</td>
<td valign="top" align="left">Emphasis on bias mitigation in PCCP guidance; recommendations for diverse datasets; no explicit mandates</td>
<td valign="top" align="left">AI Act requires: - High-quality, representative datasets - Bias monitoring and mitigation - Fundamental rights assessment - Population diversity considerations</td>
<td valign="top" align="left">Aligned with international standards; increasing focus on data quality and representation</td>
</tr>
<tr>
<td valign="top" align="left">Cybersecurity requirements</td>
<td valign="top" align="left">FDA cybersecurity guidance; premarket and postmarket requirements; Software Bill of Materials (SBOM)</td>
<td valign="top" align="left">Cyber Resilience Act; NIS2 Directive; cybersecurity mandatory for high-risk AI systems</td>
<td valign="top" align="left">Aligned with international cybersecurity standards; QMS includes cybersecurity provisions</td>
</tr>
<tr>
<td valign="top" align="left">Cost implications</td>
<td valign="top" align="left">Moderate: - 510(k): &#x0024;10,000-&#x0024;50,000 - <italic>de novo</italic>: &#x0024;50,000-&#x0024;150,000 - PMA: &#x0024;200,000-&#x0024;1M&#x002B;- User fees&#x2009;&#x002B;&#x2009;compliance costs</td>
<td valign="top" align="left">High: - Notified Body fees: &#x20AC;50,000-&#x20AC;300,000&#x002B; - Dual compliance (MDR&#x2009;&#x002B;&#x2009;AI Act) - Multiple market authorizations if multi-state - Legal/consulting fees substantial</td>
<td valign="top" align="left">Moderate-High: - Translation costs significant - Japanese representative/MAH required - Consultant fees for navigation - QMS certification costs</td>
</tr>
<tr>
<td valign="top" align="left">Market access strategy</td>
<td valign="top" align="left">Single approval for entire US market (330M&#x002B; population); largest single medical device market</td>
<td valign="top" align="left">Single CE mark grants access to 27 EU member states (450M&#x002B; population); some national requirements remain</td>
<td valign="top" align="left">Third-largest medical device market (&#x0024;30B by 2025); gateway to broader Asia-Pacific region</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3b"><label>3.2</label><title>Technical challenges</title>
<p>Artificial intelligence (AI) models in healthcare often struggle to generalize across different patient populations and healthcare settings (<xref ref-type="bibr" rid="B39">39</xref>) due to several factors, including variations in equipment, procedures, and patient demographics. AI models often perform inconsistently across demographic groups (<xref ref-type="bibr" rid="B40">40</xref>). For example, models trained primarily on middle-aged adults may inaccurately diagnose conditions in pediatric or geriatric populations due to differences in imaging characteristics and health profiles that are underrepresented in their training datasets (<xref ref-type="bibr" rid="B41">41</xref>). In addition, different hospitals and clinics may use varying protocols, equipment, and data collection methods (<xref ref-type="bibr" rid="B42">42</xref>). A recent study highlighted the effect of sample size and patient characteristics, such as age, comorbidities, and insurance type, on the performance of a clinical language model (ClinicLLM) trained on data from multiple hospitals. The results showed that models often generalized poorly in hospitals with smaller samples or for patients with certain insurance types (<xref ref-type="bibr" rid="B43">43</xref>). The insufficient representation in training datasets also significantly limits the generalizability of AI models (data representation issues) (<xref ref-type="bibr" rid="B44">44</xref>). Privacy constraints often prevent access to comprehensive datasets that include diverse demographics, leading to biases, which can compromise model performance in different populations (<xref ref-type="bibr" rid="B45">45</xref>). In addition, when a model learns too much about the training data, including noise and anomalies, its ability to generalize new data is impaired (overfitting and model uncertainty) (<xref ref-type="fig" rid="F2">Figure&#x00A0;2</xref>) (<xref ref-type="bibr" rid="B46">46</xref>). This issue is particularly relevant in clinical settings, where the variability of patient conditions can differ considerably from that of the training environment.</p>
<fig id="F2" position="float"><label>Figure&#x00A0;2</label>
<caption><p>Underfitting and overfitting in regression, classification, and deep learning. (middle) Blue and orange dots represent classes in binary classification. (bottom) In deep learning, orange lines are training datasets, and blue lines are validation datasets.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fradi-05-1733003-g002.tif"><alt-text content-type="machine-generated">Graphs illustrating underfitting, just right, and overfitting in regression, classification, and deep learning. Regression shows a straight line, a curve, and a complex curve. Classification displays a straight line, a curved boundary, and a complex boundary. Deep learning shows error vs. epoch curves where the overfitting one fluctuates.</alt-text>
</graphic>
</fig>
<p>Several strategies can be used to improve the generalization capabilities of AI models in healthcare (<xref ref-type="bibr" rid="B47">47</xref>). Models can be tailored to specific hospital settings (local fine-tuning) to improve their adaptability to the unique characteristics of the patient population in each facility, thus improving model performance. In addition, the ranges of patient demographics and health conditions in training datasets can be broadened to mitigate bias and improve overall model performance in diverse settings (diverse data inclusion) (<xref ref-type="bibr" rid="B48">48</xref>). In addition, AI systems should be assessed constantly for biases and disparities (continuous monitoring and adaptation) (<xref ref-type="bibr" rid="B49">49</xref>). User feedback should be used to refine AI technologies continuously and ensure they remain effective in evolving cultural contexts and patient needs (<xref ref-type="bibr" rid="B50">50</xref>). In summary, addressing the challenges of model performance and generalization in healthcare AI requires a multifaceted approach that considers the complexities of different patient populations and healthcare environments. The effectiveness of AI applications in healthcare can be considerably improved by focusing on local adaptations, inclusive data practices, and continuous evaluation.</p>
<p>Integrating new computer vision systems into healthcare IT infrastructure is challenging due to the complexity and siloing of existing systems (<xref ref-type="bibr" rid="B51">51</xref>). Healthcare organizations often operate with fragmented data across different departments and systems (<xref ref-type="bibr" rid="B52">52</xref>). This isolation hinders information sharing, leading to inefficiencies in patient care and decision-making. Approximately 80&#x0025; of healthcare data are unstructured and reside in disparate systems, worsening the problem of data silos (<xref ref-type="bibr" rid="B53">53</xref>). Furthermore, many healthcare providers rely on outdated technologies that were not designed for interoperability. These legacy systems increase the difficulty of integrating new technologies, such as computer vision systems, because they often cannot communicate effectively with modern applications. In addition, the lack of standardized protocols and the wide variety of data formats prevent seamless communication between different healthcare systems (<xref ref-type="bibr" rid="B54">54</xref>). Certain regulations also mandate strict privacy measures during information transfer (interoperability issues). Moreover, the integration of new technologies requires a robust IT infrastructure capable of supporting complex data exchange (technical barriers) (<xref ref-type="bibr" rid="B55">55</xref>). Without modern, flexible systems that can accommodate such integration, healthcare organizations face significant hurdles in achieving interoperability (<xref ref-type="bibr" rid="B56">56</xref>).</p>
<p>These integration challenges can be overcome using certain strategies. Implementing standards such as Fast Healthcare Interoperability Resources can enhance communication between different systems for manageable integration (adopting standards) (<xref ref-type="bibr" rid="B57">57</xref>). In addition, application programming interfaces (APIs) can enable different software applications to communicate effectively for the smooth integration of new technologies with existing systems (use of APIs). Furthermore, investing in comprehensive data platforms that centralize data management can help eliminate silos and improve interoperability between different healthcare applications (integrated data platforms) (<xref ref-type="bibr" rid="B58">58</xref>). Thus, integrating new computer vision systems into complex, siloed healthcare IT infrastructure requires a solution to challenges related to legacy systems, interoperability, and regulatory compliance. Strategies such as the use of standardized protocols and APIs can help facilitate seamless integration.</p>
<p>Computer vision applications require combining software algorithms and specialized hardware, presenting challenges for optimal implementation (<xref ref-type="bibr" rid="B59">59</xref>). Moreover, implementing computer vision systems can be expensive due to their need for powerful hardware and complex software setups. This often entails hiring skilled personnel for development and maintenance, incurring added costs. The adaptability of computer vision systems can be both a strength and a challenge (<xref ref-type="bibr" rid="B60">60</xref>). Although they can learn from data, they also require large datasets for training and regular updates to maintain their performance in dynamic environments. Therefore, computer vision applications can be implemented successfully by ensuring the balanced integration of advanced hardware and sophisticated software algorithms and considering the challenges posed by resource requirements and system complexity.</p>
</sec>
<sec id="s3c"><title>3.3 Ethical and interpretability challeng<bold>es</bold></title>
<p>DL models, often called black boxes, present critical challenges regarding explainability and interpretability (<xref ref-type="bibr" rid="B61">61</xref>). These challenges can hinder trust and adoption, particularly in sensitive domains, such as clinical settings, where decisions can profoundly affect patient care. The term &#x201C;black box&#x201D; refers to the opaque nature of DL models; their internal decision-making processes are not easily understood by humans (<xref ref-type="bibr" rid="B62">62</xref>). This lack of transparency increases the difficulty of diagnosing problems when such models produce unexpected or harmful results. For example, when an autonomous vehicle fails to stop for a pedestrian, the reason for this decision of the model is nearly impossible to understand due to its complex internal workings (<xref ref-type="bibr" rid="B63">63</xref>). This opacity can lead to misplaced trust in AI systems, as users may be unable to determine the reliability or fairness of such systems&#x0027; decisions.</p>
<p>Explainability is the provision of clear reasons for AI models&#x0027; decisions, allowing users to understand why particular results are produced (<xref ref-type="bibr" rid="B64">64</xref>). Interpretability refers to understanding how a model processes inputs to arrive at its outputs. Both concepts are critical in domains such as healthcare, where understanding the rationale behind a model&#x0027;s decisions can directly affect patient outcomes. An interpretable model allows users to understand its decision-making process, promoting accountability and transparency (responsibility) (<xref ref-type="bibr" rid="B65">65</xref>). When users understand how a model works and why it makes certain decisions, their trust in AI systems increases substantially (trust) (<xref ref-type="bibr" rid="B64">64</xref>). Several methods can be used to address the black-box problem. Tools such as heat maps can illustrate the features that most influence a model&#x0027;s decisions (visualization techniques). In addition, breaking down complex models into simpler components can help clarify how decisions are made (decomposition methods). Providing users with examples like their input can also explain how a model arrives at its conclusions (example-based explanations). The black-box nature of DL models is therefore a significant barrier to their adoption in critical applications, such as healthcare (<xref ref-type="bibr" rid="B66">66</xref>). Their explainability and interpretability should be improved to foster trust and ensure their responsible use. As AI evolves, transparency should be prioritized for its successful integration into high-stakes environments.</p>
<p>Automated decision-making in healthcare, particularly using AI, raises ethical concerns (<xref ref-type="bibr" rid="B67">67</xref>), which primarily revolve around bias, accountability, and its effect on patient autonomy. Algorithmic bias, which can occur in different stages of AI development, including data collection, model training, and development, is a crucial ethical problem (<xref ref-type="bibr" rid="B40">40</xref>). Biased data can lead to suboptimal clinical decisions and exacerbate healthcare disparities, particularly affecting marginalized populations (<xref ref-type="bibr" rid="B68">68</xref>, <xref ref-type="bibr" rid="B69">69</xref>). For example, AI systems trained on nonrepresentative datasets may produce recommendations that inadequately serve patient groups, leading to inequitable treatment outcomes (<xref ref-type="bibr" rid="B68">68</xref>). In addition, automation bias, where healthcare providers may rely excessively on AI recommendations, may cause errors of omission or commission. This bias can result from cognitive complacency, or clinicians choosing to rely on automated systems rather than exercising their clinical judgment (<xref ref-type="bibr" rid="B70">70</xref>). Such reliance can be dangerous, particularly if the AI system produces incorrect or misleading results.</p>
<p>Accountability is critical in the context of AI-driven healthcare decisions. When an AI system makes a mistake, such as an incorrect diagnosis, questions arise about the responsible party: the algorithm developers, healthcare providers (system users), or regulatory bodies (overseers of its implementation) (<xref ref-type="bibr" rid="B71">71</xref>). These algorithms&#x0027; lack of transparency further complicates the problem. Many AI systems operate as black boxes, so the bases for their recommendations are difficult for clinicians to understand (<xref ref-type="bibr" rid="B72">72</xref>). This opacity can undermine the trust between healthcare providers and patients. These accountability concerns can be addressed by establishing clear frameworks that delineate the responsibilities of stakeholders involved in AI use (<xref ref-type="bibr" rid="B73">73</xref>). These obligations include rigorously testing algorithms, regularly monitoring their accuracy, and ensuring that healthcare providers retain the ultimate decision-making authority.</p>
<p>Automated decision-making also poses challenges related to patient autonomy. Patients should be informed about how AI will influence their care decisions and be allowed to participate in discussions about their treatment options (<xref ref-type="bibr" rid="B74">74</xref>). AI systems&#x0027; prioritization of certain medical outcomes over individual patient preferences, such as quality of life over survival rates, can undermine patient autonomy and lead to their dissatisfaction with care (<xref ref-type="bibr" rid="B75">75</xref>). In addition, ensuring that patients understand the implications of AI involvement in their care is essential for informed consent (<xref ref-type="bibr" rid="B76">76</xref>). Patients should be aware of the prospective use of their data and the potential risks associated with automated decision-making (<xref ref-type="bibr" rid="B69">69</xref>). Thus, although automated decision-making in healthcare has the transformative potential to improve patient outcomes and operational efficiency, it presents complex ethical challenges. Addressing biases in AI systems, clarifying their accountability structures, and safeguarding patient autonomy are necessary to ensure that these technologies are implemented ethically and effectively (<xref ref-type="bibr" rid="B77">77</xref>). Developers, healthcare providers, regulators, and patients should cooperate in navigating these ethical concerns to foster trust and improve the quality of care delivered using AI.</p>
</sec>
<sec id="s3d"><label>3.4</label><title>Validation and regulatory challenges</title>
<p>Clinical validation of computer vision systems in healthcare is a complex, time-consuming process needed to ensure quality, safety, and efficacy (<xref ref-type="bibr" rid="B51">51</xref>). It entails thorough tests in real-world clinical settings, such as interventional or clinical trials evaluating the performance of AI systems, comparisons with relevant systems and assessment of meaningful endpoints, and minimization of bias in study design and implementation (<xref ref-type="bibr" rid="B78">78</xref>). Validation includes the reporting of performance metrics in internal and independent external test data and the benchmarking of system performance against care standards and other AI systems (<xref ref-type="bibr" rid="B79">79</xref>). Evaluation should continue after the initial validation. Performance and effects on care, including safety and effectiveness, should be monitored to understand expected and unexpected outcomes. Algorithmic audits should be conducted to understand the mechanisms of adverse events or errors. Several factors contribute to the complexity and considerable time consumption of clinical validation. Ensuring the high quality, diversity, and representativeness of data for AI model training and testing is challenging (data quality). Furthermore, the seamless integration of such models into existing healthcare systems and workflows can be difficult (interoperability and integration) (<xref ref-type="bibr" rid="B80">80</xref>). The evolving regulatory landscape for AI-based medical devices is likewise an ongoing challenge (regulatory compliance) (<xref ref-type="bibr" rid="B56">56</xref>), and validating AI systems in different clinical settings and patient populations is critical but complex (real-world performance) (<xref ref-type="bibr" rid="B81">81</xref>). Addressing bias, fairness, and patient privacy problems adds a layer of complexity to validation (ethical considerations). Despite these challenges, rigorous clinical validation is essential to build confidence in AI-based computer vision systems and ensure their safe, effective implementation in healthcare (<xref ref-type="bibr" rid="B82">82</xref>).</p>
<p>Meeting the stringent regulatory requirements for medical devices and AI in healthcare is indeed challenging and resource intensive. The continuous evolution of regulations requires constant vigilance and adaptation (ever-evolving regulations) (<xref ref-type="bibr" rid="B83">83</xref>). Documentation requirements, such as device master and design history files, are difficult to comply with (complex documentation), and compliance strategies need to be harmonized to navigate different international regulatory frameworks (global market variability). Regulating AI in healthcare presents unique difficulties, including data governance and bias mitigation, ensuring AI system transparency and accountability, and implementing effective risk management and postmarket surveillance (<xref ref-type="bibr" rid="B83">83</xref>). Moreover, keeping up to date with regulations and maintaining compliance requires considerable time and financial investment (resource constraints). Digitization of medical devices raises additional security concerns related to patient data protection (cybersecurity concerns). Regulators also must encourage innovation while ensuring patient safety and device efficacy (balancing innovation and safety). Regulators are adapting their approaches to address these challenges. For example, the EU AI Act uses a risk-based approach to regulate AI systems, categorizing them according to potential risk level (<xref ref-type="bibr" rid="B83">83</xref>). In addition, some regulators are exploring the use of AI itself to improve compliance processes, such as using machine learning (ML) algorithms to analyze clinical and operational data in real time (<xref ref-type="bibr" rid="B77">77</xref>).</p>
</sec>
<sec id="s3e"><label>3.5</label><title>Practical implementation challenges</title>
<p>Obtaining expert annotations for medical images is difficult due to financial and time costs (<xref ref-type="bibr" rid="B84">84</xref>), among other factors. Its need for specialized equipment, software, and highly skilled personnel incurs high costs, especially for small medical institutions (<xref ref-type="bibr" rid="B85">85</xref>). Furthermore, depending on the complexity of regions of interest and local anatomical structures, annotating an image can take minutes to hours (<xref ref-type="bibr" rid="B86">86</xref>). Medical images can be large, with some scans generating gigabytes of data (<xref ref-type="bibr" rid="B87">87</xref>), making their annotation process resource intensive and difficult to manage effectively. Time constraints often prevent medical professionals from allocating sufficient time for image annotation, especially in cases requiring rapid diagnosis and treatment (<xref ref-type="bibr" rid="B88">88</xref>). Medical images can be complex, requiring interpretation by highly skilled experts trained in specific medical imaging techniques (<xref ref-type="bibr" rid="B89">89</xref>). Moreover, applying DL to medical image analysis requires unprecedented amounts of labeled training data, incurring financial and time costs (large datasets) (<xref ref-type="bibr" rid="B90">90</xref>). These factors constitute a bottleneck in the development of AI-based medical imaging and limit the clinical applicability of DL.</p>
<p>The performance of ML models often decline gradually over time&#x2014;a phenomenon called model decay, or AI aging (<xref ref-type="bibr" rid="B91">91</xref>). A recent study by researchers from the Massachusetts Institute of Technology, Harvard, and other institutions found that 91&#x0025; of ML models degrade over time (<xref ref-type="bibr" rid="B92">92</xref>). This degradation is due to several factors, including temporal changes in the statistical properties of input data (data drift), relationship changes between input and output variables (concept drift), and a temporal decline in performance since model training (model aging). These issues should be addressed via continuous model monitoring and updating. Continuous model monitoring involves tracking of performance metrics and model behavior under real-world conditions; detection of data drift, concept drift, and outliers; and analysis of the root causes of performance problems. By implementing these practices, organizations can proactively manage model decay and maintain the performance reliability of their models over time.</p>
<p>Computer vision systems benefit healthcare, but healthcare professionals require extensive training to use and interpret them effectively (<xref ref-type="bibr" rid="B51">51</xref>). Such technology also requires proper understanding and integration into clinical workflows to improve diagnostic accuracy and efficiency (<xref ref-type="bibr" rid="B93">93</xref>). Healthcare professionals need to understand the basics of computer vision algorithms, including their capabilities and limitations (<xref ref-type="bibr" rid="B94">94</xref>). Consequently, they will be able to interpret results appropriately and avoid overreliance on automated systems. Training should focus on interpreting the outputs of computer vision systems, especially for complex medical images, such as x-rays, MRIs, and CT scans (<xref ref-type="bibr" rid="B17">17</xref>). Professionals should combine algorithmic insights with their clinical expertise to optimize patient care (<xref ref-type="bibr" rid="B95">95</xref>). Healthcare workers need training to integrate computer vision tools into their daily routines for seamless adoption and maximum efficiency gains (<xref ref-type="bibr" rid="B96">96</xref>). Their training should include the use of software interfaces and incorporation of system insights into decision making. As computer vision technology rapidly evolves, ongoing training is essential to keep healthcare professionals abreast of new developments and applications (<xref ref-type="bibr" rid="B97">97</xref>). Consequently, they can use recent advances to improve patient outcomes. By investing in comprehensive training programs, healthcare institutions can maximize the benefits of computer vision technology while maintaining high standards of patient care and safety (<xref ref-type="bibr" rid="B96">96</xref>). Addressing the abovementioned challenges entails collaboration between healthcare providers, AI researchers, regulators, and other stakeholders to develop robust, ethical, clinically relevant computer vision solutions for healthcare.</p>
</sec>
</sec>
<sec id="s4"><label>4</label><title>Improvement in medical image segmentation using CNNs</title>
<sec id="s4a"><label>4.1</label><title>CNN models</title>
<p>CNNs have considerably improved medical image segmentation. They can automatically learn and extract relevant features from medical images, overcoming the limitations of traditional algorithms, which rely on manually designed features (automatic feature extraction) (<xref ref-type="bibr" rid="B98">98</xref>). DL-based CNN models have state-of-the-art accuracy in various medical image segmentation tasks, often at levels comparable to those of expert radiologists (improved accuracy) (<xref ref-type="bibr" rid="B87">87</xref>). CNNs use key formulas for image segmentation. A fundamental operation in CNNs is convolution, which is expressed mathematically as<disp-formula id="disp-formula1"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM1"><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">f</mml:mi><mml:mo mathvariant="bold">&#x2217;</mml:mo><mml:mi mathvariant="bold-italic">g</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mo mathvariant="bold">,</mml:mo><mml:mi mathvariant="bold-italic">y</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>=</mml:mo><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">=</mml:mo></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">&#x2212;</mml:mo><mml:mi mathvariant="bold-italic">a</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">a</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:msubsup><mml:mo movablelimits="false">&#x2211;</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">j</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">=</mml:mo></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">&#x2212;</mml:mo><mml:mi mathvariant="bold-italic">b</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">b</mml:mi></mml:mrow></mml:msubsup><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">f</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">i</mml:mi><mml:mo mathvariant="bold">,</mml:mo><mml:mi mathvariant="bold-italic">j</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mspace width="0.25em"/><mml:mo>&#x22C5;</mml:mo><mml:mspace width="0.25em"/><mml:mrow><mml:mi mathvariant="bold-italic">g</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mo mathvariant="bold">&#x2212;</mml:mo><mml:mi mathvariant="bold-italic">i</mml:mi><mml:mo mathvariant="bold">,</mml:mo><mml:mi mathvariant="bold-italic">y</mml:mi><mml:mo mathvariant="bold">&#x2212;</mml:mo><mml:mi mathvariant="bold-italic">j</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:math><label>(1)</label></disp-formula>where <bold><italic>f</italic></bold> is the input image, <bold><italic>g</italic></bold> is the kernel or filter, and <bold><italic>x</italic></bold> and <bold><italic>y</italic></bold> are the output pixel coordinates.</p>
<p>Another important formula in CNNs for image segmentation is the activation function, commonly a rectified linear unit (ReLU), which is expressed mathematically as<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM2"><mml:mrow><mml:mi mathvariant="bold-italic">R</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi><mml:mi mathvariant="bold-italic">L</mml:mi><mml:mi mathvariant="bold-italic">U</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mspace width="0.25em"/><mml:mo>=</mml:mo><mml:mspace width="0.25em"/><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mn mathvariant="bold">0</mml:mn><mml:mo mathvariant="bold">,</mml:mo><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(2)</label></disp-formula></p>
<p>This function allows the network to learn complex patterns by introducing nonlinearity.</p>
<p>Max pooling is often used for pooling operations that reduce the spatial dimensions of feature maps.<disp-formula id="disp-formula2"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM3"><mml:mrow><mml:mi mathvariant="bold-italic">M</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mi mathvariant="bold-italic">P</mml:mi><mml:mi mathvariant="bold-italic">o</mml:mi><mml:mi mathvariant="bold-italic">o</mml:mi><mml:mi mathvariant="bold-italic">l</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">X</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mspace width="0.25em"/><mml:mo>=</mml:mo><mml:mspace width="0.25em"/><mml:mrow><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi><mml:mi mathvariant="bold-italic">i</mml:mi><mml:mo mathvariant="bold">,</mml:mo><mml:mi mathvariant="bold-italic">j</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(3)</label></disp-formula>where <bold><italic>X</italic></bold> is a subregion of the feature map. These formulas work together in CNNs to learn features, reduce dimensionality, and segment images into meaningful regions or objects.</p>
<p>Advanced CNN architectures, such as U-Net, can capture both local and global image features for the accurate segmentation of complex anatomical structures (multiscale feature learning) (<xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B99">99</xref>). U-Net is widely used for image segmentation tasks. Its key formulas are as follows:</p>
<p>Convolutional layers:<disp-formula id="disp-formula3"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM4"><mml:mrow><mml:mi mathvariant="bold-italic">y</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mo>=</mml:mo><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">f</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">W</mml:mi><mml:mo mathvariant="bold">&#x2217;</mml:mo><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mo>+</mml:mo><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">b</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(4)</label></disp-formula>where <bold><italic>y</italic></bold> is the output, <bold><italic>x</italic></bold> is the input, <bold><italic>W</italic></bold> is the conventional kernel, <bold><italic>b</italic></bold> is the bias, and <bold><italic>f</italic></bold> is the activation function (typically ReLU).</p>
<p>Upsampling:<disp-formula id="disp-formula4"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM5"><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub><mml:mspace width=".1em"/><mml:mo>=</mml:mo><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">U</mml:mi><mml:mi mathvariant="bold-italic">p</mml:mi><mml:mi mathvariant="bold-italic">S</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">p</mml:mi><mml:mi mathvariant="bold-italic">l</mml:mi><mml:mi mathvariant="bold-italic">e</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(5)</label></disp-formula><disp-formula id="disp-formula5"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM6"><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mo>=</mml:mo><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">C</mml:mi><mml:mi mathvariant="bold-italic">o</mml:mi><mml:mi mathvariant="bold-italic">n</mml:mi><mml:mi mathvariant="bold-italic">c</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub><mml:mrow><mml:mo mathvariant="bold">,</mml:mo></mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(6)</label></disp-formula>where <bold><italic>x</italic><sub>1</sub></bold> is the unsampled feature map, <bold><italic>x</italic><sub>2</sub></bold> is the skip connection from the encoder, and <bold><italic>Concat</italic></bold> is the concentration operation.<disp-formula id="disp-formula6"><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM7"><mml:mrow><mml:mi mathvariant="normal">Final</mml:mi><mml:mspace width=".1em"/><mml:mi mathvariant="normal">convolution</mml:mi></mml:mrow><mml:mo>&#x003A;</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">y</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mo>=</mml:mo><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">&#x03C3;</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="bold-italic">W</mml:mi></mml:mrow><mml:mo>&#x002A;</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">+</mml:mo></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">b</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(7)</label></disp-formula>where <italic>&#x03C3;</italic> is typically a sigmoid (softmax) activation function for binary segmentation (multiclass segmentation). Together, these formulas form a U-shaped architecture that allows U-Net to capture both local and global features for accurate image segmentation.</p>
<p>Modern CNNs can use 3D image information for a comprehensive analysis of volumetric medical imaging data, such as MRIs and CT scans (3D image processing) (<xref ref-type="bibr" rid="B87">87</xref>, <xref ref-type="bibr" rid="B98">98</xref>). Fully convolutional architectures enable end-to-end learning for pixel-wise classification, improving overall segmentation. CNNs are also versatile, applicable to different medical imaging modalities and segmentation tasks through transfer learning (adaptability) (<xref ref-type="bibr" rid="B98">98</xref>). Advanced CNN models can segment images with multiple objects, occlusions, or background noise for enhanced accuracy in challenging clinical scenarios (handling of complex cases). Using these capabilities, CNNs have become a cornerstone of modern medical image segmentation, enabling accurate, efficient analyses of medical image data in various clinical applications.</p>
</sec>
<sec id="s4b"><label>4.2</label><title>U-Net for medical image segmentation</title>
<p>U-Net has revolutionized biomedical image segmentation (<xref ref-type="bibr" rid="B100">100</xref>). It is the most widely used image segmentation architecture in medical imaging due to its flexibility, optimized modular design, and success in various applications. The U-shaped architecture of U-Net consists of a contracting encoder path and an expanding decoder path (<xref ref-type="bibr" rid="B101">101</xref>). High-resolution features from the contracting path are combined with unsampled features using skip links, enabling precise localization. Furthermore, U-Net uses limited training data efficiently through extensive data augmentation, and arbitrarily large images are seamlessly integrated using the overlap tile strategy (<xref ref-type="bibr" rid="B102">102</xref>). Other advantages of U-Net include its accurate segmentation of small targets, and performance superiority to previous methods in various biomedical image segmentation challenges (<xref ref-type="bibr" rid="B100">100</xref>).</p>
<p>Since its inception, U-Net has influenced the study of many variations and improvements. 3D U-Net extends its architecture to handle 3D volumetric data, which is crucial for analyzing biomedical image stacks (<xref ref-type="bibr" rid="B103">103</xref>). Attention mechanisms enhance feature selection and segmentation accuracy (<xref ref-type="bibr" rid="B104">104</xref>). Dense modules improve feature reuse and gradient flow. Feature enhancement techniques are used to extract meaningful features from medical images. The loss function is improved to optimize network performance for specific segmentation tasks. Generalization enhancements have been implemented to improve the model&#x0027;s ability to work across different medical imaging modalities. Certain variations were developed to address the handling of different imaging modalities, improve accuracy for small or complex structures, and optimize segmentation performance with limited data. U-Net and its variants have been successfully applied to various medical image segmentation tasks, including neuronal structure segmentation in electron microscopy stacks, cell tracking in light microscopy, caries detection in dental radiography, and organ and tumor segmentation, in various imaging modalities (<xref ref-type="bibr" rid="B105">105</xref>). The continued development and refinement of U-Net-based architectures has significantly advanced the field of medical image segmentation, providing powerful tools for automated analysis and diagnosis in healthcare (<xref ref-type="bibr" rid="B106">106</xref>).</p>
</sec>
<sec id="s4c"><label>4.3</label><title>Comparison of U-Net with other segmentation algorithms</title>
<p>U-Net is the most widely used and successful image segmentation architecture in medical imaging due to its flexibility, optimized modular design, and effectiveness across different modalities (<xref ref-type="bibr" rid="B104">104</xref>). Compared with other segmentation algorithms, U-Net offers several advantages in aspects such as performance, efficiency, versatility, adaptability, robustness, and computational efficiency. U-Net consistently outperforms existing methods in various biomedical image segmentation tasks (<xref ref-type="bibr" rid="B107">107</xref>). It performs high-accuracy retinal layer segmentation in optical coherence tomography (OCT) images, often matching or exceeding the performance of more complex variants (e.g., B1). U-Net trains rapidly and can work effectively with very few training images (<xref ref-type="bibr" rid="B85">85</xref>), which is particularly beneficial in medical imaging, where large, annotated datasets are often scarce. U-Net has also been successfully applied to a wide range of medical imaging tasks, including neural structure segmentation, cell tracking, caries detection, and organ and tumor segmentation, across different imaging modalities (<xref ref-type="bibr" rid="B108">108</xref>). Since its introduction, U-Net has influenced the development of numerous variants and enhancements, allowing researchers to adapt its architecture to specific medical imaging challenges (<xref ref-type="bibr" rid="B109">109</xref>). These variations address problems such as the handling of 3D volumetric data, incorporation of attention mechanisms, and improvement of feature extraction.</p>
<p>Vanilla U-Net often performs comparably to more complex variants across different datasets and pathologies, suggesting its robustness and generalizability (<xref ref-type="bibr" rid="B109">109</xref>). Some U-Net variants offer marginal performance improvements while increasing complexity and slowing down inference. The original U-Net architecture balances performance and computational efficiency well (<xref ref-type="bibr" rid="B110">110</xref>). Thus, its combination of accuracy, efficiency, and adaptability has made it a preferred choice for medical image segmentation tasks. Its success in various applications and datasets, coupled with its ability to perform well with limited training data, distinguishes it from many other segmentation algorithms in medical imaging.</p>
<p>U-Net has remained a dominant architecture for image segmentation since its introduction in 2015, particularly excelling in medical imaging despite the emergence of more complex models like transformers (<xref ref-type="bibr" rid="B111">111</xref>). Understanding why U-Net continues to offer specific advantages requires examining its architectural design principles, computational characteristics, and practical deployment considerations (<xref ref-type="bibr" rid="B112">112</xref>).</p>
<sec id="s4c1"><label>4.3.1</label><title>The skip connection architecture: preserving spatial information</title>
<p>The defining feature of U-Net is its skip connections, which directly address a fundamental problem in encoder-decoder architectures (<xref ref-type="bibr" rid="B113">113</xref>). Skip connections mitigate loss of fine-grained details by allowing the decoder to access high-resolution feature maps from the encoder, helping reconstruct the segmentation map with greater precision and ensuring fine details are preserved (<xref ref-type="bibr" rid="B114">114</xref>). U-Net is an encoder-decoder segmentation network with skip connections, where an encoder extracts more general features the deeper it goes, while skip connections reintroduce detailed features into the decoder (<xref ref-type="bibr" rid="B115">115</xref>). This architecture enables U-Net to segment using features that are both detailed and general.</p>
</sec>
<sec id="s4c2"><label>4.3.2</label><title>Bridging the semantic gap</title>
<p>The underlying hypothesis behind improved U-Net variants is that models can more effectively capture fine-grained details when high-resolution feature maps from the encoder are gradually enriched before fusion with semantically rich feature maps from the decoder (<xref ref-type="bibr" rid="B116">116</xref>). Skip connections in U-Net directly fast-forward high-resolution feature maps from encoder to decoder, resulting in concatenation of semantically dissimilar feature maps (<xref ref-type="bibr" rid="B117">117</xref>). This semantic dissimilarity is both a limitation and an advantage. While advanced variants like UNet&#x002B;&#x002B; address this through nested skip connections, standard U-Net&#x0027;s simpler approach often proves sufficient for many medical imaging tasks where preserving fine anatomical boundaries is critical (<xref ref-type="bibr" rid="B111">111</xref>).</p>
</sec>
<sec id="s4c3"><label>4.3.3</label><title>Gradient flow and training stability</title>
<p>Skip connections improve gradient flow during backpropagation, as gradients can diminish when propagated through deep networks&#x2014;the vanishing gradient problem. This property makes U-Net easier to train than deeper architectures, particularly important when working with limited medical imaging datasets where training stability is paramount.</p>
</sec>
<sec id="s4c4"><label>4.3.4</label><title>Computational efficiency: the practical advantage</title>
<p>U-Net&#x0027;s efficiency stems from its relatively modest computational requirements compared to modern transformer-based models, making it ideal for resource-constrained medical environments (<xref ref-type="bibr" rid="B118">118</xref>).</p>
</sec>
<sec id="s4c5"><label>4.3.5</label><title>Parameter and FLOPs analysis</title>
<p>SD-U-Net requires approximately 8&#x00D7; fewer FLOPs compared to U-Net, is 81 milliseconds faster in prediction speed for 256&#x2009;&#x00D7;&#x2009;256&#x2009;&#x00D7;&#x2009;1 inputs on an NVIDIA Tesla K40C GPU, and is 23&#x00D7; smaller (<xref ref-type="bibr" rid="B119">119</xref>). This demonstrates that even the baseline U-Net is significantly more efficient than many modern architectures. Lightweight U-Net variants reduce parameters to 12.4&#x0025; and FLOPs to 12.8&#x0025; of standard U-Net while increasing inference speed by nearly 5 times, showing that U-Net&#x0027;s architecture is amenable to further optimization without sacrificing performance. U-Net model capacities range from 1.4M to 137M parameters with corresponding FLOPs computed for 128&#x2009;&#x00D7;&#x2009;128&#x2009;&#x00D7;&#x2009;128 inputs, demonstrating that optimal architecture is highly task-specific with smaller models often performing competitively (<xref ref-type="bibr" rid="B120">120</xref>).</p>
</sec>
<sec id="s4c6"><label>4.3.6</label><title>The &#x201C;bigger is not always better&#x201D; paradigm</title>
<p>The &#x201C;bigger is better&#x201D; paradigm has significant limitations in medical imaging, as optimal U-Net architecture is highly dependent on specific task characteristics, with architectural scaling yielding distinct benefits tied to image resolution, anatomical complexity, and segmentation classes (<xref ref-type="bibr" rid="B120">120</xref>). Research reveals three key insights: increasing resolution stages provides limited benefits for datasets with larger voxel spacing; deeper networks offer limited advantages for anatomically complex shapes; and wider networks provide minimal advantages for tasks with limited segmentation classes. This task-aware approach enables better balance between accuracy and computational cost.</p>
</sec>
<sec id="s4c7"><label>4.3.7</label><title>Data efficiency: excelling with limited samples</title>
<p>Medical imaging faces a chronic data scarcity problem, and U-Net&#x0027;s architecture is particularly well-suited for learning from limited samples (<xref ref-type="bibr" rid="B121">121</xref>).</p>
</sec>
<sec id="s4c8"><label>4.3.8</label><title>Few-shot learning compatibility</title>
<p>Few-shot learning techniques have been successfully adapted to rapidly generalize to new tasks with only a few samples, leveraging prior knowledge, with MAML demonstrating high performance and generalization even on small datasets (<xref ref-type="bibr" rid="B122">122</xref>). In 10-shot settings, enhanced 3D U-Net with MAML achieved mean dice coefficients of 93.70&#x0025;, 85.98&#x0025;, 81.20&#x0025;, and 89.58&#x0025; for liver, spleen, right kidney, and left kidney segmentation respectively (<xref ref-type="bibr" rid="B122">122</xref>). U-Net&#x0027;s success in few-shot scenarios stems from its efficient parameter usage and strong inductive biases for spatial hierarchies (<xref ref-type="bibr" rid="B123">123</xref>). U-Net is like auto-encoder, learning latent representation and reconstructing output with the same size as input, providing strong feature extraction capability essential for medical image segmentation (<xref ref-type="bibr" rid="B124">124</xref>).</p>
</sec>
<sec id="s4c9"><label>4.3.9</label><title>No pre-training required</title>
<p>Unlike Vision Transformers that require massive pre-training datasets to achieve competitive performance, U-Net can be trained effectively from scratch on small medical imaging datasets (<xref ref-type="bibr" rid="B125">125</xref>). This makes it particularly valuable when domain-specific data is limited and transfer learning from natural images provides limited benefits.</p>
</sec>
<sec id="s4c10"><label>4.3.10</label><title>Robustness considerations: the skip connection trade-off</title>
<p>Recent research has revealed important nuances about U-Net&#x0027;s robustness characteristics that highlight when its advantages are most pronounced (<xref ref-type="bibr" rid="B126">126</xref>). Skip connections offer performance benefits, usually at the expense of robustness losses, depending on texture disparity between foreground and background and the range of texture variations present in the training set. Training on narrow texture ranges harms robustness in models with more skip connections, while the robustness gap between architectures reduces when trained on larger texture disparity ranges (<xref ref-type="bibr" rid="B127">127</xref>). This finding suggests that U-Net excels when: (1) training data encompasses diverse texture variations, (2) texture disparity between target and background is moderate, or (3) the task prioritizes pixel-level accuracy over robustness to texture perturbations (<xref ref-type="bibr" rid="B128">128</xref>). For robustness-critical applications, careful consideration of skip connection design is warranted (<xref ref-type="bibr" rid="B129">129</xref>).</p>
</sec>
<sec id="s4c11"><label>4.3.11</label><title>Architectural simplicity: maintainability and interpretability</title>
<p>U-Net&#x0027;s straightforward encoder-decoder structure with skip connections offers advantages that extend beyond raw performance metrics (<xref ref-type="bibr" rid="B130">130</xref>).</p>
</sec>
<sec id="s4c12"><label>4.3.12</label><title>Ease of implementation and modification</title>
<p>The architecture&#x0027;s simplicity makes it easy to implement, debug, and modify for specific applications. Researchers can readily adapt U-Net by changing the encoder backbone, adjusting network depth, or incorporating attention mechanisms without fundamentally altering its core structure (<xref ref-type="bibr" rid="B131">131</xref>). UNet&#x002B;&#x002B; introduces redesigned skip connections that enable flexible feature fusion in decoders, an improvement over restrictive skip connections in U-Net that require fusion of only same-scale feature maps (<xref ref-type="bibr" rid="B115">115</xref>). The fact that improvements can be readily incorporated into the U-Net framework demonstrates its architectural flexibility (<xref ref-type="bibr" rid="B125">125</xref>).</p>
</sec>
<sec id="s4c13"><label>4.3.13</label><title>Interpretability</title>
<p>Compared to transformer-based models, U-Net&#x0027;s convolutional operations and skip connections are more interpretable (<xref ref-type="bibr" rid="B98">98</xref>). Researchers can visualize feature maps at different encoder and decoder levels, understand what spatial features are being preserved through skip connections, and identify which resolution stages contribute most to segmentation accuracy (<xref ref-type="bibr" rid="B132">132</xref>).</p>
</sec>
<sec id="s4c14"><label>4.3.14</label><title>Domain-specific advantages in medical imaging</title>
<p>U-Net was originally designed for biomedical image segmentation, and its architecture embodies several design choices particularly suited for medical imaging tasks (<xref ref-type="bibr" rid="B120">120</xref>).</p>
</sec>
<sec id="s4c15"><label>4.3.15</label><title>Handling class imbalance</title>
<p>Medical images often have severe class imbalance with diseased tissue occupying small regions (<xref ref-type="bibr" rid="B133">133</xref>). U-Net&#x0027;s architecture, combined with appropriate loss functions like Dice loss, handles this imbalance effectively (<xref ref-type="bibr" rid="B134">134</xref>). The skip connections ensure that small target regions maintain adequate representation throughout the network.</p>
</sec>
<sec id="s4c16"><label>4.3.16</label><title>Boundary precision</title>
<p>Skip connections help preserve spatial accuracy by bringing forward detailed features from earlier layers, especially useful when models need to distinguish boundaries in segmentation tasks (<xref ref-type="bibr" rid="B135">135</xref>). In medical imaging, accurate delineation of tumor boundaries or organ edges is often more critical than achieving the highest overall pixel accuracy (<xref ref-type="bibr" rid="B136">136</xref>).</p>
</sec>
<sec id="s4c17"><label>4.3.17</label><title>Multi-scale feature integration</title>
<p>Dense and nested skip connections aim to enhance information exchange between encoder and decoder layers, allowing comprehensive exploration of multi-scale features (<xref ref-type="bibr" rid="B115">115</xref>). This multi-scale capability is essential for medical images where pathologies can appear at vastly different scales.</p>
</sec>
<sec id="s4c18"><label>4.3.18</label><title>Practical deployment: point-of-care and edge devices</title>
<p>The transition of medical imaging from laboratory settings to bedside environments creates unique deployment requirements where U-Net excels. LV-UNet model sizes and computation complexities are suitable for edge device and point-of-care scenarios (<xref ref-type="bibr" rid="B137">137</xref>). Lightweight U-Net variants can run on resource-constrained devices including mobile phones, portable ultrasound machines, and edge computing platforms without requiring cloud connectivity or high-end GPUs (<xref ref-type="bibr" rid="B138">138</xref>). Segmentation of a 512&#x2009;&#x00D7;&#x2009;512 image takes less than a second on a modern GPU using U-Net architecture, enabling real-time clinical applications (<xref ref-type="bibr" rid="B139">139</xref>). This speed, combined with low memory requirements, makes U-Net ideal for interactive segmentation tools where clinicians need immediate feedback.</p>
</sec>
<sec id="s4c19"><label>4.3.19</label><title>When U-Net remains the optimal choice</title>
<p>U-Net offers specific advantages that make it the preferred architecture as follows (<xref ref-type="bibr" rid="B106">106</xref>);
<list list-type="simple">
<list-item><label>(1)</label>
<p>Limited Training Data: Small medical imaging datasets (hundreds rather than thousands of images) where transformers would overfit</p></list-item>
<list-item><label>(2)</label>
<p>Computational Constraints: Deployment on edge devices, mobile platforms, or resource-limited clinical settings</p></list-item>
<list-item><label>(3)</label>
<p>Real-Time Requirements: Applications requiring sub-second inference times without GPU acceleration</p></list-item>
<list-item><label>(4)</label>
<p>Boundary Precision: Tasks where accurate delineation of fine structures is more important than classification accuracy</p></list-item>
<list-item><label>(5)</label>
<p>Interpretability Needs: Clinical applications requiring explainable predictions and feature visualization</p></list-item>
<list-item><label>(6)</label>
<p>Development Speed: Projects with limited time or expertise for implementing complex architectures</p></list-item>
<list-item><label>(7)</label>
<p>Stable Texture Environments: When training and deployment data have consistent texture characteristics.</p></list-item>
</list>The architecture&#x0027;s longevity is not merely historical inertia but reflects genuine technical advantages for specific problem domains. While transformers and foundation models push the boundaries of what&#x0027;s possible with massive datasets and computational resources, U-Net continues to provide an optimal balance of simplicity, efficiency, and effectiveness for practical medical image segmentation tasks where data is limited, resources are constrained, and reliability is paramount (<xref ref-type="bibr" rid="B140">140</xref>).</p>
</sec>
</sec>
<sec id="s4d"><label>4.4</label><title>Challenges in applying U-Net to different medical imaging modalities</title>
<p>The application of U-Net to different medical imaging modalities faces several challenges. Medical image datasets are scarce and difficult to obtain compared with ordinary computer vision datasets (<xref ref-type="bibr" rid="B106">106</xref>). This scarcity is due to privacy concerns, limited availability of annotated data, diversity of imaging modalities (e.g., x-ray, MRI, CT, and ultrasound), and the need for annotation by medical professionals (<xref ref-type="bibr" rid="B141">141</xref>). Improving image quality and standardizing imaging protocols across different modalities are also critical. Problems include variations in image acquisition from different medical devices, lack of uniform standards for annotation and CT/MRI machine performance, and inconsistencies in image quality, which affect model generalization. In addition, medical imaging requires extremely high accuracy for disease diagnosis. Related concerns include the difficulty of distinguishing boundaries between multiple cells and organs, the need for pixel- or voxel-level segmentation, and continuous parameter adjustment for achieving and maintaining accuracy. Furthermore, professionals lack confidence in applying DL model predictions to medical images. Specific problems include the poor interpretability of U-Net (<xref ref-type="bibr" rid="B106">106</xref>). In addition, gaining the trust and acceptance of expert physicians for clinical applications is difficult, and achieving interpretability while maintaining performance is a complex objective.</p>
</sec>
</sec>
<sec id="s5"><label>5</label><title>Applications for computer vision in real-world medical imaging</title>
<p>Revolutionizing healthcare diagnosis and treatment, computer vision has numerous real-world applications in medical imaging (<xref ref-type="bibr" rid="B94">94</xref>). In radiology and diagnostic imaging, computer vision algorithms help radiologists detect and classify abnormalities in x-rays, CT scans, and MRIs. For example, DL models can identify lung nodules in CXRs with high accuracy, aiding in the early detection of lung cancer (<xref ref-type="bibr" rid="B142">142</xref>). Computer vision is also used in digital pathology to analyze tissue samples and detect cancer cells (<xref ref-type="bibr" rid="B143">143</xref>). These systems can quickly process large volumes of slides, improving the efficiency and accuracy of cancer diagnosis. In addition, AI-powered systems are used to analyze retinal images to detect eye diseases, such as diabetic retinopathy, glaucoma, and age-related macular degeneration (AMD) (<xref ref-type="bibr" rid="B144">144</xref>&#x2013;<xref ref-type="bibr" rid="B146">146</xref>). This technology enables early diagnosis and intervention, potentially preventing vision loss. Computer vision algorithms can also analyze skin lesions in photographs, helping dermatologists identify potential skin cancers and other dermatological conditions (<xref ref-type="bibr" rid="B147">147</xref>). In surgical planning and guidance, the 3D reconstruction of medical images helps surgeons plan complex procedures (<xref ref-type="bibr" rid="B148">148</xref>). During surgery, computer vision is used in augmented reality systems to overlay critical information on the surgeon&#x0027;s view (<xref ref-type="bibr" rid="B149">149</xref>). In cardiovascular imaging, echocardiograms and coronary CT angiography images are analyzed using AI algorithms to detect heart abnormalities and assess cardiovascular risk (<xref ref-type="bibr" rid="B150">150</xref>). In neuroimaging, computer vision techniques are used to analyze brain scans to help diagnose neurological disorders, such as Alzheimer&#x0027;s disease, multiple sclerosis, and brain tumors (<xref ref-type="bibr" rid="B80">80</xref>, <xref ref-type="bibr" rid="B151">151</xref>&#x2013;<xref ref-type="bibr" rid="B153">153</xref>).</p>
<p>AI-based systems can also detect and classify polyps in real time during colonoscopy, improving the accuracy of colorectal cancer screening (<xref ref-type="bibr" rid="B154">154</xref>). Computer-aided detection (CAD) systems assist radiologists in identifying potential breast cancer lesions on mammograms, increasing detection rates and reducing false negatives (<xref ref-type="bibr" rid="B155">155</xref>). AI algorithms are used in emergency medicine to analyze CT scans and detect critical conditions rapidly, such as intracranial hemorrhage, accelerating treatment in time-sensitive situations (<xref ref-type="bibr" rid="B156">156</xref>). These applications demonstrate the value of computer vision in enhancing medical imaging across multiple specialties, where it improves diagnostic accuracy, treatment planning, and patient outcomes.</p>
</sec>
<sec id="s6"><label>6</label><title>Computer vision techniques for cancer detection</title>
<p>Computer vision techniques contribute substantially to cancer detection by improving the accuracy, efficiency, and accessibility of diagnostic procedures. These techniques use advanced image processing algorithms and ML to analyze various types of medical images, including x-rays, CT scans, MRIs, and histopathology slides (<xref ref-type="bibr" rid="B157">157</xref>). Computer vision algorithms can detect subtle abnormalities in medical images, enabling early cancer detection/diagnosis (<xref ref-type="bibr" rid="B4">4</xref>). This capability is critical for improving treatment outcomes and patient prognoses. AI-powered systems can analyze large volumes of medical imaging data with high accuracy, often outperforming traditional diagnostic methods (improved accuracy) (<xref ref-type="bibr" rid="B158">158</xref>). For example, HiDisc, developed by researchers at the University of Michigan, has demonstrated almost 88&#x0025; accuracy in certain cases. CAD systems can rapidly analyze medical images, reducing the time required for diagnosis from days to minutes (improved efficiency) (<xref ref-type="bibr" rid="B159">159</xref>). Consequently, healthcare professionals can focus more on patient care. AI-powered systems also act as &#x201C;second pairs of eyes&#x201D; for clinicians, helping them identify potential cancerous lesions or tumors that may be missed during routine examinations. Computer vision models can apply the expertise of top oncologists to image analysis, making high-quality cancer detection more accessible in rural areas and developing countries (democratization of expertise) (<xref ref-type="bibr" rid="B4">4</xref>).</p>
<p>AI algorithms help detect lung nodules in CXRs and analyze mammograms for breast cancer screening (<xref ref-type="bibr" rid="B160">160</xref>). Regarding digital pathology, computer vision is used to analyze tissue samples and detect cancer cells in histopathology slides (<xref ref-type="bibr" rid="B143">143</xref>). Advanced techniques, such as transfer learning, ensemble learning, and vision transformers (ViTs), are used to integrate information from different imaging modalities, such as CT, MRI, and positron emission tomography scans, for comprehensive cancer detection (multimodal analysis) (<xref ref-type="bibr" rid="B4">4</xref>).</p>
<p>The ViT architecture involves several key formulas. Its key formula for self-attention is<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="DM8"><mml:msubsup><mml:mrow><mml:mi mathvariant="bold-italic">s</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow><mml:mrow><mml:mspace width=".1em"/><mml:mspace width=".1em"/><mml:mi mathvariant="normal">&#x2032;</mml:mi></mml:mrow></mml:msubsup><mml:mspace width=".1em"/><mml:mrow><mml:mo mathvariant="bold">=</mml:mo></mml:mrow><mml:mspace width=".1em"/><mml:mrow><mml:mi mathvariant="bold-italic">s</mml:mi><mml:mi mathvariant="bold-italic">o</mml:mi><mml:mi mathvariant="bold-italic">f</mml:mi><mml:mi mathvariant="bold-italic">t</mml:mi><mml:mi mathvariant="bold-italic">m</mml:mi><mml:mi mathvariant="bold-italic">a</mml:mi><mml:mi mathvariant="bold-italic">x</mml:mi></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">q</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">k</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>/</mml:mo><mml:mo stretchy="false">&#x221A;</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">d</mml:mi><mml:mo mathvariant="bold">,</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">q</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">k</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>/</mml:mo><mml:mo stretchy="false">&#x221A;</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">d</mml:mi><mml:mo mathvariant="bold">,</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">q</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">1</mml:mn></mml:mrow></mml:msub><mml:mo>&#x22C5;</mml:mo><mml:msub><mml:mrow><mml:mi mathvariant="bold-italic">k</mml:mi></mml:mrow><mml:mrow><mml:mn mathvariant="bold">3</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo>/</mml:mo><mml:mo stretchy="false">&#x221A;</mml:mo><mml:mrow><mml:mi mathvariant="bold-italic">d</mml:mi><mml:mo mathvariant="bold">,</mml:mo></mml:mrow><mml:mo>&#x2026;</mml:mo></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:math><label>(8)</label></disp-formula>where <bold><italic>q<sub>1</sub></italic></bold> is the query vector; <bold><italic>k<sub>1</sub></italic></bold>, <bold><italic>k</italic></bold><bold><italic><sub>2</sub></italic></bold>, and <bold><italic>k</italic></bold><bold><italic><sub>3</sub></italic></bold> are the key vectors; <bold><italic>d</italic></bold> is the dimension of the key vectors; and <bold><italic>s<sub>1</sub></italic></bold>&#x2032; is the attention weight vector.</p>
<p>The input image <bold><italic>x</italic></bold> &#x2208; RH&#x2009;&#x00D7;&#x2009;W&#x2009;&#x00D7;&#x2009;C is divided into patches, where H, W, and C are the height, width, and channels, respectively. These patches are then flattened and linearly embedded, with positional embeddings added to preserve spatial information. The resulting sequence is fed into a standard transformer encoder that applies multihead self-attention and feedforward layers. Final classification is performed using a softmax function on the output of the transformer encoder.</p>
<p>Although computer vision techniques offer considerable potential for cancer detection, challenges remain, including the need to improve the availability of high-quality labeled datasets, the diagnosis of rare cancers, and model explainability and generalization (<xref ref-type="bibr" rid="B4">4</xref>). Ongoing research aims to address these challenges and further enhance the role of computer vision in cancer detection and diagnosis.</p>
</sec>
<sec id="s7"><label>7</label><title>Application of ML algorithms to medical images</title>
<sec id="s7a"><label>7.1</label><title>U-Net with ML algorithms for image segmentation</title>
<p>U-Net was used to segment brain images in the Brain Tumor Segmentation challenge (<xref ref-type="bibr" rid="B161">161</xref>), where it helped identify and delineate brain tumors accurately in MRI scans. In addition, the architecture was used for liver image segmentation in the &#x201C;siliver07&#x201D; challenge for liver segmentation in CT scans (<xref ref-type="bibr" rid="B162">162</xref>), helping detect liver boundaries and identify potential lesions accurately. U-Net variants are also used for the semantic segmentation of OCT scans of the posterior segment of the eye. Specifically, they help identify different regions in OCT images of healthy eyes and eyes with conditions such as Stargardt&#x0027;s disease and AMD (<xref ref-type="bibr" rid="B163">163</xref>). Furthermore, U-Net is adopted to predict protein binding sites, which is critical for understanding protein&#x2013;protein interactions and drug discovery (<xref ref-type="bibr" rid="B164">164</xref>), and to estimate fluorescent staining (in image-to-image translation), which is valuable in cellular imaging and analysis (<xref ref-type="bibr" rid="B165">165</xref>).</p>
<p>U-Net is implemented to analyze the micrographs of materials to characterize them and study their properties at microscopic scales (<xref ref-type="bibr" rid="B166">166</xref>). Pixelwise regression using U-Net is used for pansharpening, which combines high-resolution panchromatic and low-resolution multispectral satellite imagery (<xref ref-type="bibr" rid="B167">167</xref>). In addition, the U-Net architecture is used in diffusion models for iterative image denoising; in this application, it underlies modern image generation models, such as DALL-E, Midjourney, and Stable Diffusion (<xref ref-type="bibr" rid="B168">168</xref>). As for medical image reconstruction, U-Net variations have improved the quality and resolution of medical imaging techniques. 3D U-Net is used to learn dense volumetric segmentation from sparse annotation, which is particularly useful in 3D medical imaging (<xref ref-type="bibr" rid="B169">169</xref>). These applications demonstrate the versatility of U-Net in tackling different image analysis tasks in various domains, with each implementation being tailored to task-specific requirements.</p>
</sec>
<sec id="s7b"><label>7.2</label><title>You-only-look-once (YOLO) algorithms in detecting structures, anomalies, and instruments within medical images</title>
<p>YOLOv5 is used to detect and classify pressure ulcers in medical images, potentially improving patient outcomes and reducing healthcare costs (<xref ref-type="bibr" rid="B170">170</xref>). In CXR analysis, YOLOv5 and faster R-CNN were used to identify various abnormalities, with YOLOv5 demonstrating superior accuracy (email protected] 0.616 vs. faster R-CNN&#x0027;s 0.580) (<xref ref-type="bibr" rid="B171">171</xref>). YOLO algorithms are effective tools for tumor detection and localization in various medical imaging modalities (<xref ref-type="bibr" rid="B172">172</xref>). YOLOv8 enables real-time processing when applied to MRI images for brain tumor detection (<xref ref-type="bibr" rid="B172">172</xref>). YOLO models can identify tumors, lesions, and other clinically relevant medical objects in various imaging modalities (<xref ref-type="bibr" rid="B173">173</xref>). Likewise, YOLO can detect and localize anatomical structures, helping diagnose cardiovascular, neurological, and other conditions (<xref ref-type="bibr" rid="B170">170</xref>). YOLOv8 introduces several enhancements, including a refined spine network and neck portion (influenced by YOLOv7&#x0027;s efficient layer aggregation network design), the adoption of a decoupled head structure, a shift from an anchor-based approach to an anchor-free one, and improved data augmentation (<xref ref-type="bibr" rid="B172">172</xref>). Ultralytics&#x0027; YOLOv11 further enhances medical imaging, particularly for tumor detection, by enabling radiologists to handle higher case volumes with consistent quality and improving real-time object detection and interpretation of complex images (<xref ref-type="bibr" rid="B174">174</xref>).</p>
<p>However, despite their successes, YOLO algorithms face certain challenges in medical imaging, including the need for large and balanced datasets, high computational requirements, limited flexibility with highly heterogeneous images, suboptimal sensitivity and precision for images such as MRIs, and possible inadequacy of standard loss functions for precise boundary localization in medical images (<xref ref-type="bibr" rid="B172">172</xref>, <xref ref-type="bibr" rid="B175">175</xref>). Researchers are addressing these limitations by developing more flexible convolutional networks to adapt to complex image features, improving sensitivity and accuracy for specific medical imaging modalities, or designing specialized loss functions for better boundary localization in medical images (<xref ref-type="bibr" rid="B172">172</xref>). These ongoing improvements aim to improve the performance of YOLO algorithms in medical applications, particularly in handling images with complex backgrounds and unclear boundaries.</p>
</sec>
</sec>
<sec id="s8"><label>8</label><title>Active learning, reinforcement learning, and federated learning for medical images</title>
<sec id="s8a"><label>8.1</label><title>Potential of active learning, reinforcement learning, and federated learning</title>
<p>Active learning, reinforcement learning, and federated learning provide innovative solutions to various challenges in medical imaging. Active learning is particularly useful for medical imaging because of the considerable cost and time consumption of data labeling (<xref ref-type="bibr" rid="B176">176</xref>). Active learning has been applied to the classification of CXR images for COVID-19 detection, significantly reducing the required labeling (<xref ref-type="bibr" rid="B177">177</xref>). For example, structure-and-boundary-consistency-based active learning was developed for medical image segmentation (<xref ref-type="bibr" rid="B178">178</xref>); open-set active learning was used for nucleus detection in medical images (<xref ref-type="bibr" rid="B179">179</xref>); and foundation-model-driven active barely supervised learning (FM-ABS) was adopted for 3D medical image segmentation (<xref ref-type="bibr" rid="B180">180</xref>).</p>
<p>Reinforcement learning, especially deep reinforcement learning, has several applications in medical imaging (<xref ref-type="bibr" rid="B181">181</xref>). It has been used to identify key anatomical points in medical images, detect and localize objects or lesions in medical scans, align medical images from different modalities or time points, determine optimal viewing planes in 3D imaging, optimize ML model parameters for medical imaging tasks, and segment and analyze surgical motion in medical procedures.</p>
<p>Federated learning is gaining traction in medical imaging because it enables collaborative model training while maintaining privacy (<xref ref-type="bibr" rid="B182">182</xref>, <xref ref-type="bibr" rid="B183">183</xref>). Multiple medical institutions can collaboratively train models without sharing raw patient data. Federated MRI reconstruction techniques are used to improve image quality across multiple centers. Federated learning is used to segment tumors in medical images from multiple institutions. Additionally, Fed-CBT generates representative connectivity maps from multicenter brain imaging data, whereas FedFocus was developed for COVID-19 detection in CXRs across multiple healthcare centers. These applications demonstrate the potential of active, reinforcement, and federated learning to address critical challenges in medical imaging, including data scarcity, decision-making complexity, and privacy concerns (<xref ref-type="bibr" rid="B184">184</xref>).</p>
</sec>
<sec id="s8b"><label>8.2</label><title>Problems in medical imaging</title>
<p>ML and computer vision algorithms are used in medical imaging to solve broad-ranging problems and enhance diagnostic capabilities and patient care. These algorithms are primarily used for disease detection and diagnosis, image segmentation, early disease detection, image analysis and interpretation, surgical assistance, patient monitoring, workflow optimization, and multimodal analysis. AI systems can identify various conditions, such as tumors, lesions, and anatomical abnormalities, in medical images, such as x-rays, MRIs, CT scans, and ultrasound images (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B29">29</xref>). For example, they can detect breast cancer in mammograms with higher accuracy (fewer false positives and false negatives) than human radiologists (<xref ref-type="bibr" rid="B185">185</xref>).</p>
<p>DL algorithms, particularly CNNs, accurately and efficiently segment medical images, helping isolate specific structures or regions of interest (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B31">31</xref>). These algorithms recognize subtle patterns in medical images and patient records, enabling the early detection of diseases such as cancer, Alzheimer&#x0027;s, and cardiovascular disease, often before symptoms develop. AI models can automatically extract meaningful features from medical images for efficient, accurate interpretation. This entails disease classification, assessment of disease progression, and evaluation of treatment response (<xref ref-type="bibr" rid="B17">17</xref>). Computer vision algorithms can assist in surgical planning and guidance through a real-time analysis of medical images during surgery (<xref ref-type="bibr" rid="B186">186</xref>). These technologies enable the real-time monitoring of patient conditions, which is particularly useful for treating chronic conditions, postoperative recovery, and elderly care. By automating image analysis, these algorithms help streamline hospital workflows, allowing radiologists to prioritize urgent cases and reducing their manual workload. DL algorithms facilitate comprehensive analysis by integrating data from multiple imaging modalities, thus improving the overall diagnostic process. By addressing these issues, ML and computer vision algorithms in medical imaging improve diagnostic accuracy, accelerate clinical decision-making, and ultimately improve patient outcomes.</p>
</sec>
<sec id="s8c"><label>8.3</label><title>Performance comparison of ML and computer vision algorithms</title>
<p>As powerful tools for data analysis and image processing, ML and computer vision algorithms have become increasingly important in various fields. Performance insights by some tasks are summarized in <xref ref-type="table" rid="T2">Table&#x00A0;2</xref> (<xref ref-type="bibr" rid="B187">187</xref>).</p>
<table-wrap id="T2" position="float"><label>Table&#x00A0;2</label>
<caption><p>Performance insights by task.</p></caption>
<table>
<colgroup>
<col align="left"/>
<col align="left"/>
<col align="left"/>
</colgroup>
<thead>
<tr>
<th valign="top" align="left">Tasks</th>
<th valign="top" align="center">Models</th>
<th valign="top" align="center">Performances</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left" rowspan="4">Classification tasks</td>
<td valign="top" align="left">ResNet-50</td>
<td valign="top" align="left">98.37&#x0025; accuracy (chest x-ray)</td>
</tr>
<tr>
<td valign="top" align="left">DeiT-Small</td>
<td valign="top" align="left">92.16&#x0025; accuracy (brain tumor)</td>
</tr>
<tr>
<td valign="top" align="left">EfficientNet-B0</td>
<td valign="top" align="left">81.84&#x0025; accuracy (skin cancer)</td>
</tr>
<tr>
<td valign="top" align="left">Fine-Tuned ResNet50</td>
<td valign="top" align="left">98.20&#x0025; accuracy, 99.00&#x0025; precision, 98.82&#x0025; recall, 98.91&#x0025; F1-score (COVID-19)</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="3">Segmentation tasks</td>
<td valign="top" align="left">DenseNet-121 U-Net</td>
<td valign="top" align="left">0.79&#x2013;0.87 precision, 0.92&#x2013;0.97 recall</td>
</tr>
<tr>
<td valign="top" align="left">Diffusion-CSPAM-U-Net</td>
<td valign="top" align="left">84.4&#x0025; DSC, 73.1&#x0025; IoU</td>
</tr>
<tr>
<td valign="top" align="left">Attention UNet</td>
<td valign="top" align="left">85.36&#x0025; IoU, 91.49&#x0025; Dice score</td>
</tr>
<tr>
<td valign="top" align="left" rowspan="3">Detection tasks</td>
<td valign="top" align="left">YOLOv5-v8</td>
<td valign="top" align="left">95&#x0025;&#x2013;99.17&#x0025; precision, 97.5&#x0025; sensitivity, &#x003E;95&#x0025; mAP</td>
</tr>
<tr>
<td valign="top" align="left">YOLO-NeuroBoost</td>
<td valign="top" align="left">99.48&#x0025; mAP (brain tumors)</td>
</tr>
<tr>
<td valign="top" align="left">YOLOv10</td>
<td valign="top" align="left">20&#x2005;ms inference time (kidney stones)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In addition, the object detection algorithms perform differently. LinearSVC performs the best across all metrics, followed by SGDC and logistic regression (<xref ref-type="table" rid="T2">Table&#x00A0;2</xref>).</p>
<p>Image classification performance:<disp-formula><mml:math xmlns:mml="http://www.w3.org/1998/Math/MathML" id="UDM1"><mml:mrow><mml:mi mathvariant="normal">Accuracy</mml:mi></mml:mrow><mml:mspace width=".1em"/><mml:mo>=</mml:mo><mml:mspace width=".1em"/><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">True</mml:mi><mml:mspace width=".1em"/><mml:mi mathvariant="normal">Positives</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mspace width=".1em"/><mml:mo>+</mml:mo><mml:mspace width=".1em"/><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">True</mml:mi><mml:mspace width=".1em"/><mml:mi mathvariant="normal">Negatives</mml:mi></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mrow><mml:mo>/</mml:mo><mml:mi mathvariant="normal">Total</mml:mi><mml:mspace width=".1em"/><mml:mi mathvariant="normal">Samples</mml:mi></mml:mrow></mml:math></disp-formula>CNNs typically outperform traditional ML algorithms in image classification. A study comparing different architectures showed accuracies of 95.2&#x0025;, 93.7&#x0025;, and 94.5&#x0025; for ResNet-50, VGG-16, and Inception-v3, respectively.</p>
<p>Researchers found that adjacency-matrix (AM) diagrams consistently outperformed node&#x2013;link (NL) diagrams in graph visualization tasks, specifically counting tasks, connectivity tasks, and performance degradation. The performance superiority of AM was more evident for larger graphs. NL accuracy decreased significantly for graphs with more than 30 nodes, but AM remained stable (<xref ref-type="bibr" rid="B188">188</xref>).</p>
<p>Different algorithms excel at varying tasks. CNNs are superior for image-related tasks, whereas support vector machines (SVMs) and random forests perform well on structured data. DL algorithms generally outperform traditional ML algorithms on large datasets but require more computational resources. Simpler models, such as decision trees, offer better interpretability, whereas complex models, such as neural networks, offer higher accuracy at the cost of interpretability. DL models typically require larger datasets to achieve optimal performance compared with traditional ML algorithms. Feature scaling and data preprocessing substantially affect the performance of many algorithms, especially KNNs and SVMs. In graph-related tasks, AM diagrams outperform NL diagrams, especially for larger and denser graphs. Thus, the choice of algorithm depends on the task, dataset size, available computational resources, and interpretability requirement. As the field evolves, hybrid approaches combining traditional ML and DL techniques are emerging, aiming to leverage the strengths of both paradigms (<xref ref-type="bibr" rid="B189">189</xref>).</p>
</sec>
</sec>
<sec id="s9"><label>9</label><title>Critical thoughts and research questions for computer vision in medical imaging</title>
<p>Computer vision in medical imaging has revolutionized disease detection, diagnosis, and image interpretation. Computer vision algorithms can remarkably detect disease in medical images. AI systems can identify lung nodules in chest CT scans at sensitivities comparable to those of experienced radiologists (<xref ref-type="bibr" rid="B190">190</xref>). DL algorithms can detect breast cancer in mammograms with greater accuracy (fewer false positives and false negatives) than human radiologists (<xref ref-type="bibr" rid="B191">191</xref>). AI systems can detect diabetic retinopathy in retinal images with high accuracy (90.3&#x0025; sensitivity and 98.1&#x0025; specificity), enabling early intervention (<xref ref-type="bibr" rid="B192">192</xref>).</p>
<p>Computer vision is improving medical image analysis through the automated analysis of large volumes of medical images, reducing radiologists&#x0027; workload and increasing efficiency; accurate interpretation of x-rays, MRIs, and CT scans, which helps identify abnormalities and accelerates diagnosis; and advanced techniques, such as image segmentation, object detection, disease classification, and image reconstruction (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B141">141</xref>). However, the following research questions remain: (i) Data quality and quantity: How can we ensure the availability of high-quality, diverse, well-annotated medical imaging datasets for training robust AI models? (ii) Interpretability: How can we develop explainable AI models that have a transparent decision-making process, which is critical for clinical confidence and adoption? (iii) Generalization: How can we develop AI models that perform consistently across different patient populations, imaging devices, and healthcare settings? (iv) Integration into clinical workflows: How can computer vision tools be seamlessly integrated into existing clinical workflows to improve efficiency without compromising patient care? (v) Ethical and legal considerations: What are the ethical implications of using AI in medical imaging, and how can we address liability and patient privacy problems? (vi) Multimodal integration: How can we combine computer vision with other data sources (e.g., patient history and genomics) for more comprehensive and accurate diagnoses? (vii) Real-time analysis: How can we develop computer vision algorithms capable of real-time analysis for time-critical applications, such as stroke detection and surgical assistance? (viii) Validation and clinical trials: How can we rigorously validate computer vision algorithms in clinical settings and ensure their safety and efficacy? (ix) Continuous learning: How can we design AI systems that can adapt and improve over time as they encounter new data and clinical scenarios? (x) Resource optimization: How can computer vision technologies be optimized to run efficiently on limited computational resources and thus be accessible in different healthcare settings? (<xref ref-type="bibr" rid="B193">193</xref>, <xref ref-type="bibr" rid="B194">194</xref>). These research questions and critical considerations should be addressed to advance the field of computer vision in medical imaging and realize its full potential to improve patient care and outcomes.</p>
<p>Deep learning has revolutionized medical imaging by enabling automated, faster, and often more accurate analysis of various imaging modalities, such as x-rays, CT scans, MRIs, and ultrasound (<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B195">195</xref>). This success is primarily due to the ability of deep learning models, especially CNNs, to automatically learn complex, hierarchical features directly from raw image data, eliminating the need for manual feature engineering. The most used deep learning architectures and techniques in medical imaging include CNNs, U-Net, Recurrent Neural Networks (RNNs)/Long Short-Term Memory (LSTMs), Generative Adversarial Networks (GANs), and Transformers (<xref ref-type="bibr" rid="B196">196</xref>). The foundational model for most computer vision tasks, CNNs use convolutional layers to extract spatial hierarchies of features (like edges, textures, and shapes) from images. They are the backbone for classification, detection, and segmentation tasks (<xref ref-type="fig" rid="F3">Figure&#x00A0;3</xref>). A highly popular variant of CNNs, U-Net, particularly for image segmentation features a symmetric encoder-decoder architecture with skip connections that allow high-resolution feature maps from the encoder path to be combined with the upsampled output of the decoder (<xref ref-type="bibr" rid="B123">123</xref>). This combination helps produce very precise segmentation masks (<xref ref-type="fig" rid="F4">Figure&#x00A0;4</xref>). While less common than CNNs for 2D images, RNNs are useful for tasks involving sequential data, such as analyzing adjacent slices in 3D scans (like CT or MRI) or tracking disease progression over time (<xref ref-type="bibr" rid="B4">4</xref>). These GANs models consist of a generator and a discriminator network that compete, allowing them to create synthetic, realistic-looking medical images (<xref ref-type="bibr" rid="B197">197</xref>). GANs are used for data augmentation (especially for rare diseases), image denoising, and reconstruction (e.g., low-dose CT reconstruction) (<xref ref-type="bibr" rid="B198">198</xref>). Also, architectures (adapted for vision) of Transformer, initially successful in natural language processing, are increasingly being used in medical imaging for tasks requiring a global understanding of the image content (<xref ref-type="bibr" rid="B199">199</xref>).</p>
<fig id="F3" position="float"><label>Figure&#x00A0;3</label>
<caption><p>A deep learning model processing time-series data. CNN architectures are structured pipelines of layers designed to analyze visual data by using a feature extractor (convolutional, pooling, and activation layers) followed by a classifier (fully connected layers). Key components include convolutional layers for feature extraction, pooling layers to reduce spatial dimensions, and fully connected layers to perform the final classification.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fradi-05-1733003-g003.tif"><alt-text content-type="machine-generated">Diagram illustrating a neural network architecture. On the left is an input chart labeled \"Hydrophobicity\" with values fluctuating over a range. This input is processed through three convolutional layers, each with kernels labeled \\(k_1 \\times k_1\\), \\(k_2 \\times k_2\\), and \\(k_3 \\times k_3\\), maintaining the same padding. Each layer has a set number of channels: \\(n_1\\), \\(n_2\\), \\(n_3\\). The output is then flattened and passed through a fully connected layer with various nodes labeled Alpha, Beta, Delta, Gamma, and Omicron.</alt-text>
</graphic>
</fig>
<fig id="F4" position="float"><label>Figure&#x00A0;4</label>
<caption><p>The architecture of the encoder in a transformer neural network. A transformer is a deep learning model that adopts the mechanism of self-attention, differently weighing the significance of each part of the input data. The encoder&#x0027;s job is to process the input data, such as a sentence in a source language, and convert it into a high-dimensional representation.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fradi-05-1733003-g004.tif"><alt-text content-type="machine-generated">Diagram of an encoder in a transformer neural network. It shows three connected blocks, each containing a self-attention layer in green and a feed-forward layer in blue. Arrows indicate data flow from input to decoder, with a legend explaining the layers' colors.</alt-text>
</graphic>
</fig>
<p>Deep learning methods are applied across the entire medical imaging pipeline, including i) image classification: identifying the presence of a disease (e.g., classifying a mammogram as benign or malignant) or categorizing a condition, ii) object detection and localization: identifying the precise location of abnormalities, lesions, or organs within an image using bounding boxes (e.g., detecting lung nodules on a chest x-ray), iii) image segmentation: assigning a label to every pixel, effectively outlining the boundaries of tumors, organs, or other structures (<xref ref-type="bibr" rid="B200">200</xref>). This is crucial for precise treatment planning, iv) computer-aided diagnosis (CAD): Providing a second, automated opinion to radiologists to help flag potentially missed findings, increasing efficiency and reducing error rates, and v) image reconstruction and enhancemen: low-dose CT/denoising: removing noise from images acquired with lower radiation doses and super-resolution: generating high-resolution images from lower-resolution scans (especially for MRI) (<xref ref-type="bibr" rid="B201">201</xref>). Despite their success, the application of deep learning in a clinical setting still faces several challenges, such as data scarcity and annotation, interpretability (Explainable AI - XAI), and generalization (<xref ref-type="bibr" rid="B202">202</xref>). Medical images are difficult and expensive to label accurately by experts, and datasets for specific rare diseases are often small (<xref ref-type="bibr" rid="B203">203</xref>). Also, the &#x201C;black box&#x201D; nature of deep learning makes it difficult for clinicians to understand why a model made a specific prediction, which is a significant barrier to clinical trust and adoption (<xref ref-type="bibr" rid="B204">204</xref>). Models often perform poorly when tested on data from different hospitals, scanners, or patient populations than those they were trained on. Therefore, future research focuses on addressing these challenges through transfer learning and foundation models; leveraging models pre-trained on large, general datasets and fine-tuning them on smaller medical datasets, self-supervised and semi-supervised learning; training models effectively with limited labeled data, and federated learning; training models across multiple institutions without sharing patient data, helping to improve generalization while maintaining privacy.</p>
</sec>
<sec id="s10"><label>10</label><title>Critical synthesis in the field of computer vision in medical imaging</title>
<p>Computer vision has become a transformative force in medical imaging, with rapid advancements in deep learning reshaping diagnostic workflows and clinical decision-making (<xref ref-type="bibr" rid="B29">29</xref>). This synthesis examines the critical developments, persistent challenges, and emerging paradigms in this rapidly evolving field.</p>
<sec id="s11d"><label>10.1</label><title>The paradigm shift to generative and unified models</title>
<p>A fundamental evolution is occurring in medical imaging AI. Medical Vision Generalist (MVG) represents the first foundation model capable of handling various medical imaging tasks such as cross-modal synthesis, image segmentation, denoising, and inpainting within a unified framework. This marks a departure from task-specific models toward versatile, adaptable systems that can generalize across imaging modalities. Recent developments focus on methods for image-to-image translation, image synthesis, biophysical modeling, super-resolution and image segmentation, demonstrating how synthesis techniques are becoming central to addressing data scarcity and enhancing image quality (<xref ref-type="bibr" rid="B205">205</xref>). The integration of generative AI enables clinicians to overcome traditional limitations like equipment costs and accessibility barriers, as evidenced by work on generating 3D OCT images from 2D fundus photographs (<xref ref-type="bibr" rid="B206">206</xref>).</p>
</sec>
<sec id="s11a"><label>10.2</label><title>The interpretability imperative</title>
<p>Despite impressive accuracy gains, the black-box nature of deep learning models remains the most significant barrier to clinical adoption. State-of-the-art deep learning models have achieved human-level accuracy on classification of different types of medical data, yet these models are hardly adopted in clinical workflows, mainly due to their lack of interpretability (<xref ref-type="bibr" rid="B207">207</xref>). Five core elements define interpretability in medical imaging machine learning: localization, visual recognizability, physical attribution, model transparency, and actionability (<xref ref-type="bibr" rid="B24">24</xref>). The field is increasingly recognizing that inherently interpretable models may be more valuable than <italic>post hoc</italic> explanation methods. Techniques like attention mechanisms, saliency mapping, and gradient-based visualization are evolving to provide clinically meaningful explanations that radiologists can trust and act upon (<xref ref-type="bibr" rid="B208">208</xref>). Key challenges include data availability, interpretability, overfitting, and computational requirements, underscoring that technical performance alone is insufficient for real-world deployment.</p>
</sec>
<sec id="s11b"><label>10.3</label><title>Privacy-preserving collaborative learning</title>
<p>The tension between data-hungry deep learning models and stringent privacy regulations has catalyzed federated learning research (<xref ref-type="bibr" rid="B209">209</xref>). Federated learning offers a privacy-preserving solution by enabling collaborative model training across institutions without sharing sensitive data, with model parameters exchanged between participating sites (<xref ref-type="bibr" rid="B210">210</xref>). However, federated learning faces its own challenges. Sensitive information can still be inferred from shared model parameters, and post deployment data distribution shifts can degrade model performance, making uncertainty quantification essential. Emerging solutions combine differential privacy, secure multi-party computation, and homomorphic encryption to strengthen privacy guarantees while maintaining model performance. Differentially private federated learning has achieved strong privacy bounds with performance comparable to centralized training, demonstrating viability for real-world medical applications (<xref ref-type="bibr" rid="B211">211</xref>).</p>
</sec>
<sec id="s11c"><label>10.4</label><title>The data heterogeneity challenge</title>
<p>Medical imaging presents unique challenges that confront deep learning approaches, with billions of medical imaging studies conducted per year worldwide, and this number is growing (<xref ref-type="bibr" rid="B194">194</xref>). The heterogeneity of medical data&#x2014;across institutions, imaging protocols, patient populations, and equipment&#x2014;creates significant generalization challenges. Challenges such as data variability, interpretability, and generalization across different patient populations and imaging modalities need to be addressed to ensure reliable and effective medical image analysis. Hybrid approaches that integrate traditional computer vision techniques with deep learning, alongside multi-modal learning strategies, are emerging as promising solutions (<xref ref-type="bibr" rid="B212">212</xref>).</p>
</sec>
<sec id="s11h"><label>10.5</label><title>Uncertainty quantification: from prediction to confidence</title>
<p>Clinical decision-making requires not just predictions but confidence estimates. In federated learning contexts, uncertainty quantification becomes particularly challenging due to data heterogeneity across participating sites. Methods including Bayesian approaches, ensemble techniques, and conformal prediction are being developed to provide clinicians with reliable uncertainty estimates that can guide treatment decisions (<xref ref-type="bibr" rid="B213">213</xref>).</p>
</sec>
<sec id="s11e"><label>10.6</label><title>Architectural innovation and efficiency</title>
<p>The field has progressed beyond basic CNNs to sophisticated architectures. U-Net and its variants remain dominant for segmentation tasks, while Vision Transformers (ViTs) are increasingly adopted for capturing long-range dependencies in medical images (<xref ref-type="bibr" rid="B137">137</xref>). Deep learning-based models achieve significant increases in segmentation, classification, and anomaly detection accuracies across various medical imaging modalities (<xref ref-type="bibr" rid="B214">214</xref>). Attention mechanisms enable models to focus on clinically relevant regions, while transfer learning and domain adaptation help overcome limited annotated data (<xref ref-type="bibr" rid="B215">215</xref>). The challenge remains balancing model complexity with computational efficiency, particularly for deployment in resource-constrained clinical settings.</p>
<p>Several critical gaps persist evaluation standards, clinical integration, robustness and fairness, and regulatory frameworks. The field lacks standardized benchmarks for evaluating interpretability, uncertainty quantification, and fairness across diverse patient populations. Despite technical advances, seamless integration with existing clinical workflows and electronic health record systems remains incomplete. Models trained on data from specific institutions often fail to generalize across diverse demographics, raising concerns about algorithmic bias and healthcare equity. The rapid pace of AI development has outpaced regulatory guidance, creating uncertainty around clinical validation and FDA approval pathways.</p>
</sec>
<sec id="s11f"><label>10.7</label><title>Toward trustworthy medical AI</title>
<p>The evolution of computer vision in medical imaging reflects maturation from purely accuracy-focused models toward systems that prioritize trustworthiness, privacy, and clinical utility (<xref ref-type="bibr" rid="B216">216</xref>). The convergence of foundation models, privacy-preserving techniques, and explainable AI represents a holistic approach to addressing the field&#x0027;s fundamental challenges. Success will require continued interdisciplinary collaboration between computer scientists, radiologists, ethicists, and policymakers. The goal is not simply to replicate human performance but to create AI systems that augment clinical expertise, reduce diagnostic errors, and ultimately improve patient outcomes while maintaining the highest standards of privacy and interpretability. The field stands at an inflection point where technical capability increasingly meets clinical need but realizing this potential demands addressing the critical challenges of interpretability, privacy, heterogeneity, and real-world generalization with the same rigor applied to achieving high accuracy metrics.</p>
</sec>
</sec>
<sec id="s12"><label>11</label><title>Quantitative comparisons, performance metrics, and systematic evaluations in computer vision for medical imaging</title>
<sec id="s12a"><label>11.1</label><title>CNN vs. vision transformer performance comparisons</title>
<p>A systematic review analyzing 36 studies indicates that transformer-based models, particularly Vision Transformers, exhibit significant potential in diverse medical imaging tasks, showcasing superior performance when contrasted with conventional CNN models (<xref ref-type="bibr" rid="B217">217</xref>). As for task-specific performance, across three medical imaging tasks - chest x-ray pneumonia detection, brain tumor classification, and skin cancer detection - results demonstrate task-specific model advantages: ResNet-50 achieved 98.37&#x0025; accuracy on chest x-ray classification, DeiT-Small excelled at brain tumor detection with 92.16&#x0025; accuracy, and EfficientNet-B0 led skin cancer classification at 81.84&#x0025; accuracy (<xref ref-type="table" rid="T2">Table&#x00A0;2</xref>) (<xref ref-type="bibr" rid="B218">218</xref>). Further, as for hybrid architecture advantages, a systematic review of 28 articles on hybrid ViT-CNN architecture identified that integrating ViT and CNN can mitigate the limitations of each architecture, offering comprehensive solutions that combine global context understanding with precise local feature extraction (<xref ref-type="bibr" rid="B219">219</xref>). The CSPDarknet53 backbone is the fastest with a decent receptive field, while EfficientNet-B3 has the largest receptive field but the lowest FPS, illustrating the fundamental trade-off between receptive field size and inference speed.</p>
</sec>
<sec id="s12b"><label>11.2</label><title>Benchmark datasets and evaluation frameworks</title>
<p>MedSegBench encompasses 35 datasets with over 60,000 images from ultrasound, MRI, and x-ray modalities, evaluating U-Net architectures with various encoder networks including ResNets, EfficientNet, and DenseNet, with DenseNet-121 consistently demonstrating strong performance across numerous datasets, achieving precision scores such as 0.794 in BusiMSB, 0.870 in ChuahMSB, and 0.801 in Dca1MSB (<xref ref-type="bibr" rid="B220">220</xref>). As for performance across encoders, DenseNet-121 emerged as a top performer for recall, obtaining the highest recall in datasets such as Bbbbc010MSB (0.920) and WbcMSB (0.970), while EfficientNet performed well in recall metrics, particularly in datasets like DynamicNuclearMSB (0.966) and USforKidneyMSB (0.982). Furthermore, MedMNIST v2 provides a collection of standardized biomedical images including 12 datasets for 2D and 6 datasets for 3D, consisting of 708,069 2D images and 9,998 3D images in total, with benchmarking showing that Google AutoML Vision performs well in general, though ResNet-18 and ResNet-50 can outperform it on certain datasets (<xref ref-type="bibr" rid="B221">221</xref>).</p>
</sec>
<sec id="s12c"><label>11.3</label><title>Imagenet performance metrics</title>
<p>Quantitative comparisons provide the essential metrics needed to evaluate and select computer vision algorithms for specific applications. ImageNet remains the gold standard benchmark for image classification. Top-1 error rate measures whether the top predicted label matches the ground truth, while Top-5 error rate checks if the correct label is among the top five predictions. The evolution of architectures shows dramatic improvements such as AlexNet (2012), 15.4&#x0025; top-5 error rate, marking the deep learning revolution, VGG-16/19 (2014), 7.3&#x0025; top-5 error rate, 138 million parameters, ResNet (2015), ResNet achieved a top-5 error rate as low as 3.57&#x0025;, which refreshed the record of precision of CNNs on ImageNet, EfficientNet-B7, Achieved a top-one accuracy of 84.3 percent on ImageNet, and Modern models (2024&#x2013;2025), CoCa achieves 91.0&#x0025; top-1 accuracy on ImageNet after fine-tuning, but requires 2.1B parameters.</p>
<p>EfficientNet-B1 is 7.6&#x00D7; smaller and 5.7&#x00D7; faster than ResNet-152, demonstrating how architectural innovation can dramatically improve the accuracy-to-efficiency trade-off (<xref ref-type="bibr" rid="B222">222</xref>). ResNet-50 is faster than VGG-16 and more accurate than VGG-19, while ResNet-101 is about the same speed as VGG-19 but much more accurate than VGG-16 (<xref ref-type="bibr" rid="B223">223</xref>). The computational differences are stark: VGG-16 requires roughly 138 million parameters with ResNet having 25.5 million parameters, though parameter count alone doesn&#x0027;t determine speed. VGG&#x0027;s early convolutional layers on full-resolution images create massive computational costs, while ResNet&#x0027;s early downsampling strategy dramatically reduces FLOPs.</p>
</sec>
<sec id="s12d"><label>11.4</label><title>Current state-of-the-art (2025)</title>
<p>EfficientNet provides the best accuracy-to-parameter ratio, with EfficientNet-B0 achieving 77.1&#x0025; accuracy with only 5.3M parameters, while ConvNeXt V2 offers a strong balance with the Tiny variant achieving 83.0&#x0025; accuracy using 28.6M parameters. For comparative performance, in general, Faster R-CNN is more accurate while R-FCN and SSD are faster, with Faster R-CNN using Inception ResNet with 300 proposals giving the highest accuracy at 1 FPS for all tested cases (<xref ref-type="bibr" rid="B224">224</xref>).</p>
</sec>
<sec id="s12e"><label>11.5</label><title>U-Net vs. mask R-CNN performance</title>
<p>Quantitative comparisons in medical imaging reveal task-specific performance differences, such as panoramic radiograph segmentation, lung segmentation in CT, pancreas segmentation, cardiac MRI segmentation, and brain tumor segmentation (<xref ref-type="bibr" rid="B225">225</xref>). Multi-Label U-Net achieved a dice coefficient of 0.96 and an IoU score of 0.97, while Mask R-CNN attained a dice coefficient of 0.87 and an IoU score of 0.74, with Mask R-CNN showing accuracy, precision, recall, and F1-score values of 95&#x0025;, 85.6&#x0025;, 88.2&#x0025;, and 86.6&#x0025; respectively (<xref ref-type="bibr" rid="B226">226</xref>). Mask R-CNN coupled with a K-means kernel delivered the best segmentation results, achieving an accuracy of 97.68&#x2009;&#x00B1;&#x2009;3.42&#x0025; with an average runtime of 11.2&#x2005;s (<xref ref-type="bibr" rid="B32">32</xref>). The maximum Dice Similarity Coefficients value for automatic pancreas segmentation techniques is only around 85&#x0025; due to the complex structure of the pancreas, demonstrating that certain anatomical structures remain challenging even for state-of-the-art models (<xref ref-type="bibr" rid="B227">227</xref>). U-Net achieved a mean DSC that reached 97.9&#x0025; and a mean Hausdorff Distance that reached 5.318&#x2005;mm for cardiac MRI segmentation (<xref ref-type="bibr" rid="B228">228</xref>). Hybrid CNN models U-SegNet, Res-SegNet, and Seg-U-Net achieved average accuracies of 91.6&#x0025;, 93.3&#x0025;, and 93.1&#x0025; respectively on the BraTS dataset (<xref ref-type="bibr" rid="B32">32</xref>).</p>
</sec>
<sec id="s12f"><label>11.6</label><title>Segmentation metrics - dice coefficient and IoU</title>
<p>As for U-Net performance metrics, for lung area segmentation in medical images using U-Net, the Dice coefficient increased from 0.5 to 0.9, while the IOU value stabilized at 0.9, representing the model&#x0027;s efficiency in proper segmentation with minimal overfitting as shown by the loss metrics&#x0027; consistent reduction (<xref ref-type="bibr" rid="B229">229</xref>). In prostate segmentation using U-Net with nine different loss functions, Focal Tversky loss function achieved the highest average DSC scores for the whole gland at 0.74&#x2009;&#x00B1;&#x2009;0.09, while models using IoU, Dice, Tversky and weighted BCE&#x2009;&#x002B;&#x2009;Dice loss functions obtained similar DSC scores ranging from 0.71 to 0.73 (<xref ref-type="bibr" rid="B134">134</xref>). In Advanced U-Net Variants, the diffusion-CSPAM-U-Net model for brain metastases segmentation achieved internal validation results with DSC of 84.4&#x0025;&#x2009;&#x00B1;&#x2009;12.8&#x0025;, IoU of 73.1&#x0025;&#x2009;&#x00B1;&#x2009;12.5&#x0025;, accuracy of 97.2&#x0025;&#x2009;&#x00B1;&#x2009;9.6&#x0025;, sensitivity of 83.8&#x0025;&#x2009;&#x00B1;&#x2009;11.3&#x0025;, and specificity of 97.2&#x0025;&#x2009;&#x00B1;&#x2009;13.8&#x0025; (<xref ref-type="bibr" rid="B230">230</xref>). F-measure based metrics like Dice Similarity Coefficient and Intersection-over-Union are highly popular and recommended in medical image segmentation, with the difference being that IoU penalizes under- and over-segmentation more than DSC, and these metrics focus on true positive classification without true negative inclusion, providing better performance representation in medical contexts (<xref ref-type="bibr" rid="B231">231</xref>).</p>
</sec>
<sec id="s12g"><label>11.7</label><title>YOLO performance in medical imaging</title>
<p>The review highlights impressive performance of YOLO models, particularly from YOLOv5 to YOLOv8, in achieving high precision up to 99.17&#x0025;, sensitivity up to 97.5&#x0025;, and mAP exceeding 95&#x0025; in tasks such as lung nodule, breast cancer, and polyp detection, demonstrating significant potential for early disease detection and real-time clinical applications (<xref ref-type="bibr" rid="B232">232</xref>). Across 123 peer-reviewed papers published between 2018 and 2024, mAP scores exceeding 85&#x0025; were commonly achieved in breast cancer and pulmonary nodule detection tasks, while lightweight versions such as YOLOv5s maintained detection speed above 50 FPS in surgical environments (<xref ref-type="bibr" rid="B233">233</xref>). The YOLO-NeuroBoost model for brain tumor detection in MRI images achieved mAP scores of 99.48 on the Br35H dataset and 97.71 on the open-source Roboflow dataset, indicating high accuracy and efficiency in detecting brain tumors (<xref ref-type="bibr" rid="B172">172</xref>). In kidney stone detection comparing YOLOv8 and YOLOv10, YOLOv10 demonstrated faster inference at approximately 20 milliseconds compared to YOLOv8&#x0027;s 30 milliseconds, largely due to its NMS-free architecture, making it better suited for real-time detection where both speed and accuracy are critical.</p>
<p>YOLOv4 achieves 43.5&#x0025; mAP at a real-time speed of approximately 65 FPS on the Tesla V100 GPU, running twice as fast as EfficientDet with comparable performance, and improving YOLOv3&#x0027;s mAP and FPS by 10&#x0025; and 12&#x0025; respectively on the MS COCO dataset.</p>
<p>In comparative analysis for skin disease classification, YOLOv5n achieved the highest accuracy of 0.942 with a training time of 1.204&#x2005;h, while ResNet50 showed accuracy of 0.571 in detecting chickenpox, 0.666 in measles, 0.803 in monkeypox, and 0.864 in normal cases. EfficientNet_b2 demonstrated the highest accuracy among EfficientNet models in this application. As critical insights from quantitative comparisons, the quantitative data reveals several important patterns. Modern architectures like EfficientNet and ConvNeXt achieve similar or better accuracy than older models with dramatically fewer parameters. The speed-accuracy trade-off remains fundamental&#x2014;YOLO excels in real-time detection while Faster R-CNN provides superior accuracy at lower speeds (<xref ref-type="bibr" rid="B234">234</xref>). For medical imaging, U-Net consistently outperforms Mask R-CNN for semantic segmentation tasks, while Mask R-CNN excels when instance-level separation is required (<xref ref-type="bibr" rid="B111">111</xref>). Task-specific factors like anatomical complexity significantly impact achievable accuracy regardless of architecture choice (<xref ref-type="bibr" rid="B235">235</xref>). The evolution from AlexNet&#x0027;s 84.6&#x0025; accuracy to modern models exceeding 91&#x0025; demonstrates the field&#x0027;s rapid progress, though improvements are increasingly marginal as models approach theoretical limits on benchmark datasets (<xref ref-type="bibr" rid="B236">236</xref>).</p>
</sec>
<sec id="s12h"><label>11.8</label><title>Few-shot learning performance</title>
<p>A systematic review analyzing relevant articles found that few-shot learning techniques can reduce data scarcity issues and enhance medical image analysis speed and robustness, with meta-learning being a popular choice because it can adapt to new tasks with few labelled samples (<xref ref-type="bibr" rid="B237">237</xref>). Experiments on four few-shot segmentation tasks show that Interactive Few-Shot Learning approach outperforms state-of-the-art methods by more than 20&#x0025; in the DSC metric, with the interactive optimization algorithm contributing approximately 10&#x0025; DSC improvement for few-shot segmentation models (<xref ref-type="bibr" rid="B238">238</xref>).</p>
</sec>
<sec id="s12i"><label>11.9</label><title>Foundation model benchmarking</title>
<p>A novel dataset and benchmark for foundation model adaptation collected five sets of medical imaging data from multiple institutes targeting real-world clinical tasks, examining generalizability across varied data modalities, image sizes, data sample numbers, and classification tasks including multi-class, multi-label, and regression (<xref ref-type="bibr" rid="B239">239</xref>). Also, evaluation of foundation models on patch-level pathology tasks revealed that pathology image pre-trained foundation models consistently outperformed those based on common images across all datasets, with no consistent winner across all benchmark datasets, emphasizing the importance of measuring performance over a diverse set of downstream tasks (<xref ref-type="bibr" rid="B240">240</xref>).</p>
</sec>
<sec id="s12j"><label>11.10</label><title>Standard evaluation metrics definitions</title>
<p>Standard metrics reported in medical imaging studies include accuracy, precision (ranging from 0.86&#x2013;0.91 in validation studies), recall (ranging from 0.83 in studies), and F1 scores (ranging from 0.84&#x2013;0.87), though AUROC and AUPRC cannot be calculated without access to the model or confusion matrix entries (<xref ref-type="bibr" rid="B241">241</xref>). Also, as for metric dependencies, accuracy, positive predictive value, negative predictive value, AUCPRC, and F1 score are functions of prevalence, while AUCROC, sensitivity, specificity, false-positive rate, and false-negative rate are not dependent on prevalence, making AUCPRC particularly suitable for scenarios with class imbalance (<xref ref-type="bibr" rid="B242">242</xref>).</p>
</sec>
<sec id="s12k"><label>11.11</label><title>Comparative architecture performance</title>
<p>In hip fracture detection using pelvic radiographs, YOLOv5 achieved 92.66&#x0025; accuracy on regular images and 88.89&#x0025; on CLAHE-enhanced images, while classifier models including MobileNetV2, Xception, and InceptionResNetV2 achieved accuracies between 94.66&#x0025; and 97.67&#x0025;, significantly outperforming clinicians&#x0027; mean accuracy of 84.53&#x0025; (<xref ref-type="bibr" rid="B243">243</xref>). For intracranial aneurysm detection, ResNet reached sensitivity of 91&#x0025; and 93&#x0025; for internal and external test datasets respectively, while CNN classifiers achieved detection of 94.2&#x0025; of aneurysms with 2.9 false positives per case (<xref ref-type="table" rid="T2">Table&#x00A0;2</xref>) (<xref ref-type="bibr" rid="B244">244</xref>).</p>
</sec>
<sec id="s12l"><label>11.12</label><title>Dataset quality and duplication and bias and fairness metrics</title>
<p>The proliferation of duplicate datasets poses significant impediments to reproducibility, with the ISIC skin lesion dataset having 27 versions on HuggingFace and 640 datasets on Kaggle totaling 2.35 TB of data compared to the original 38 GB, with many lacking original sources or license information. Current study reveals significant correlation between each model&#x0027;s accuracy in making demographic predictions and the size of its fairness gap, suggesting models may be using demographic categorizations as shortcuts to make disease predictions (<xref ref-type="bibr" rid="B245">245</xref>). This comprehensive quantitative overview demonstrates that model selection should be task-specific, with transformers generally outperforming CNNs for tasks requiring global context, while hybrid approaches and properly optimized CNNs remain competitive for many applications (<xref ref-type="bibr" rid="B246">246</xref>). Performance evaluation should use multiple complementary metrics appropriate to the clinical context and class balance.</p>
</sec>
</sec>
<sec id="s13"><label>12</label><title>New taxonomy or experimental validation, fresh perspectives, techniques, or frameworks for models and problems in the field of computer vision in medical imaging</title>
<sec id="s13a"><label>12.1</label><title>Vision transformers (ViTs) - beyond traditional CNNs</title>
<p>As for hierarchical multi-scale approaches, recent research introduces hierarchical multi-scale Vision Transformer frameworks that incorporate innovative attention methodologies, using multi-resolution patch embedding strategies (8&#x2009;&#x00D7;&#x2009;8, 16&#x2009;&#x00D7;&#x2009;16, and 32&#x2009;&#x00D7;&#x2009;32 patches) for feature extraction across different spatial scales, achieving 35&#x0025; reduction in training duration compared to conventional ViT implementations (<xref ref-type="bibr" rid="B247">247</xref>). Systematic reviews indicate that transformer-based models, particularly ViTs, exhibit significant potential in diverse medical imaging tasks, showcasing superior performance when contrasted with conventional CNN models, with pre-training being particularly important for transformer applications (<xref ref-type="bibr" rid="B217">217</xref>). There are some important innovations, RanMerFormer implements a randomized approach to combining visual tokens, substantially reducing computational demands while classifying brain tumors. Lesion-Centered Vision Transformers integrate lesion-focused MRI preprocessing with adaptive token merging for stroke outcome. Swin Transformers with linear complexity characteristics for processing MRI data prediction (<xref ref-type="bibr" rid="B248">248</xref>).</p>
</sec>
<sec id="s13b"><label>12.2</label><title>Foundation models - the paradigm shift</title>
<p>In Vision-Language Foundation Models (VLFMs), foundation models in the medical domain address critical challenges by combining information from various medical imaging modalities with textual data from radiology reports and clinical notes, enabling development of tools that streamline diagnostic workflows, enhance accuracy, and enable robust decision-making (<xref ref-type="bibr" rid="B249">249</xref>). Currently, there are some novel capabilities. NVIDIA&#x0027;s Clara NV-Reason-CXR-3B model generates detailed thought processes for chest x-ray analysis by capturing radiologist thought processes through voice recordings, with a two-stage training pipeline combining supervised fine-tuning with gradient reinforcement policy optimization. Zero-shot and few-shot performance across multiple downstream tasks. Foundation models exhibit remarkable contextual understanding and generalization capabilities, with active research focusing on developing versatile artificial intelligence solutions for real-world healthcare applications (<xref ref-type="bibr" rid="B250">250</xref>). Also, as challenges identified, studies investigating algorithmic fairness of vision-language foundation models reveal that compared to board-certified radiologists, these models consistently underdiagnose marginalized groups, with even higher rates in intersectional subgroups such as Black female patients (<xref ref-type="bibr" rid="B251">251</xref>).</p>
</sec>
<sec id="s13c"><label>12.3</label><title>Addressing data scarcity</title>
<p>Multi-task learning enables simultaneous training of a single model that generalizes across multiple tasks, taking advantage of many small- and medium-sized datasets in biomedical imaging by efficiently utilizing different label types and data sources to pretrain image representations applicable to all tasks (<xref ref-type="bibr" rid="B20">20</xref>). Physics-Informed Machine Learning (PIML) incorporates governing physical laws such as partial differential equations, boundary conditions, and conservation principles, guiding the learning process toward physically plausible and interpretable outcomes, reducing the need for large datasets while enhancing interpretability. In diffusion models for synthesis, conditional latent diffusion model-based medical image enhancement networks incorporate multi-attention modules and Rotary Position Embedding to effectively capture positional information, with findings indicating generated images can enhance performance of downstream classification tasks, providing effective solutions to scarcity of medical image training data (<xref ref-type="bibr" rid="B252">252</xref>). There are some critical approaches: GANs for data augmentation (though diffusion models are increasingly preferred), self-supervised learning on large unlabeled datasets, synthetic data generation with anatomical control, and cross-domain transfer learning (<xref ref-type="bibr" rid="B253">253</xref>).</p>
</sec>
<sec id="s13d"><label>12.4</label><title>Explainable AI (XAI) - building trust</title>
<p>As for novel XAI frameworks, recent work proposes explainable AI methods specifically designed for medical image analysis, integrating statistical, visual, and rule-based explanations to improve transparency in deep learning models, using decision trees and RuleFit to extract human-readable rules while providing statistical feature map overlay visualizations (<xref ref-type="bibr" rid="B25">25</xref>). Also, there are some important techniques: i) Grad-CAM and LIME for visual saliency maps (<xref ref-type="bibr" rid="B254">254</xref>), ii) attention-based saliency maps improve interpretability of pneumothorax classification, with studies combining saliency-based heatmaps with clinical decision tasks and validating their plausibility through qualitative feedback from radiologists (<xref ref-type="bibr" rid="B255">255</xref>), iii) concept-based explanations using activation vectors (<xref ref-type="bibr" rid="B256">256</xref>), and iv) counterfactual explanations showing how image changes would alter predictions. As critical insight, from a human-centered design perspective, transparency is not a property of the ML model but an affordance - a relationship between algorithm and users, making prototyping and user evaluations critical to attaining solutions that afford transparency (<xref ref-type="bibr" rid="B257">257</xref>).</p>
<p>In addressing bias and fairness, the PROBAST-AI tool is under development specifically for evaluating ML studies, while reporting guidelines such as FUTURE-AI and TRIPOD-AI assist authors in reporting studies according to the Fairness principle, promoting identification of bias sources (<xref ref-type="bibr" rid="B258">258</xref>). Surprisingly, models with less encoding of demographic attributes are often most &#x201C;globally optimal&#x201D;, exhibiting better fairness during model evaluation in new test environments, though correcting shortcuts algorithmically effectively addresses fairness gaps within the original data distribution (<xref ref-type="bibr" rid="B245">245</xref>). Current study reveals significant correlation between each model&#x0027;s accuracy in making demographic predictions and the size of its fairness gap, suggesting models may be using demographic categorizations as shortcuts to make disease predictions. As mitigation strategies, it suggested i) subgroup robustness optimization, ii) group adversarial approaches to remove demographic information, iii) diverse dataset curation across demographics, and iv) regular fairness audits across deployment contexts (<xref ref-type="bibr" rid="B259">259</xref>).</p>
</sec>
<sec id="s13e"><label>12.5</label><title>Novel architecture and frameworks</title>
<p>In U-Net evolution, U-Net&#x002B;&#x002B; represents a nested U-Net architecture for medical image segmentation, while multi-scale attention U-Net with EfficientNet-B4 encoder enhances MRI analysis (<xref ref-type="bibr" rid="B117">117</xref>). CNN, Deep Learning algorithm that takes as an image or a multivariate time series, can successfully capture the spatial and temporal patterns through the application of trainable filters, and assigns importance to these patterns using trainable weights (<xref ref-type="fig" rid="F5">Figure&#x00A0;5</xref>). The preprocessing required in CNN is much lower than compared to other classification algorithms. While in many methods filters are hand-engineered, CNN can learn these filters. As diffusion models for medical imaging, there are applications in image-to-image translation, reconstruction, registration, and anomaly detection and atomically controllable generation following multi-class segmentation masks. Diffusion probabilistic models synthesize high-quality medical data for MRI and CT, with radiologists evaluating synthetic images on realistic appearance, anatomical correctness, and consistency between slices (<xref ref-type="bibr" rid="B260">260</xref>). Also, as hybrid approaches, there are multi-modal fusion networks integrating imaging with clinical data and cross-attention mechanisms for modality integration. CNN-Transformer hybrids combining local and global feature extraction (<xref ref-type="bibr" rid="B261">261</xref>).</p>
<fig id="F5" position="float"><label>Figure&#x00A0;5</label>
<caption><p>The U-Net&#x002B;&#x002B; and attention U-Net architectures are advanced variants of the original U-Net, designed to improve performance in image segmentation tasks, particularly in medical imaging. U-Net&#x002B;&#x002B; aims to improve the standard U-Net by addressing the semantic gap between the encoder and decoder feature maps connected by skip connections. It achieves this through a nested, densely connected structure. Instead of a single U-shaped path, U-Net&#x002B;&#x002B; incorporates multiple U-Net-like structures nested within each other. This creates pathways of varying depths (Nested U-Net Structure). Attention U-Net enhances the U-Net architecture by integrating attention mechanisms, specifically Attention Gates (AGs), into the skip connections. AGs are placed in the skip connections between the encoder and decoder. They learn to suppress irrelevant regions in the input image and highlight salient features relevant to the target structure.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="fradi-05-1733003-g005.tif"><alt-text content-type="machine-generated">Diagrams of U-Net++ and Attention U-Net architectures. Both use skip connections and layers labeled with colors: pink for MaxPooling, red for SoftMax, yellow for Convolution, and blue for Concatenate/ADD. U-Net++ shows nested skip pathways, while Attention U-Net includes attention mechanisms. Arrows indicate the data flow.</alt-text>
</graphic>
</fig>
<p>Transformer-based models have fundamentally reshaped computer vision, moving beyond the convolutional paradigm that dominated for decades (<xref ref-type="bibr" rid="B262">262</xref>). This analysis explores Vision Transformers (ViTs), Segment Anything Model (SAM), and Swin-U-Net architectures, examining their mechanisms, performance characteristics, and transformative impact on the field (<xref ref-type="bibr" rid="B263">263</xref>).</p>
</sec>
<sec id="s13f"><label>12.6</label><title>Architecture and applications of ViTs</title>
<sec id="s13f1"><label>12.6.1</label><title>Fundamental architecture of ViTs</title>
<p>Vision Transformers revolutionized computer vision by adapting the transformer architecture from NLP to image processing. ViT builds upon the transformer architecture using self-attention mechanisms to model global relationships across image patches (<xref ref-type="bibr" rid="B264">264</xref>). The architecture divides images into fixed-size patches (typically 16&#x2009;&#x00D7;&#x2009;16 pixels), flattens them into vectors, and processes them through transformer encoder blocks. The standard ViT-Base configuration uses 768-dimensional embeddings, 12 transformer layers, 12 attention heads, and approximately 86 million parameters. This design enables ViT to capture global dependencies from the earliest layers&#x2014;a stark contrast to CNNs that build hierarchical features progressively (<xref ref-type="bibr" rid="B265">265</xref>).</p>
</sec>
<sec id="s13f2"><label>12.6.2</label><title>Performance and scaling characteristics of ViTs</title>
<p>ViT architecture designs have considerable impact on out-of-distribution generalization, though in-distribution accuracy might not be a very good indicator of OoD accuracy. These finding challenges conventional wisdom about model evaluation and highlight the importance of robust metrics beyond standard benchmarks. When properly scaled, ViTs achieve remarkable results. ViT-22B, currently the largest vision transformer at 22 billion parameters, advances state-of-the-art on several benchmarks and shows increased similarities to human visual object recognition. This massive scale was achieved through architectural innovations including parallel layers (reducing training time by 15&#x0025;) and omitting biases in certain projections (increasing utilization by 3&#x0025;).</p>
</sec>
<sec id="s13f3"><label>12.6.3</label><title>Data efficiency and training requirements of ViTs</title>
<p>ViTs typically require large amounts of training data to achieve their best performance, while CNNs can perform well even with smaller datasets (<xref ref-type="bibr" rid="B217">217</xref>). This data hunger stems from ViT&#x0027;s lack of inductive biases&#x2014;the very property that gives it flexibility also means it must learn spatial relationships from scratch. While ViT shows excellent potential in learning high-quality image features, it is inferior in performance vs. accuracy gains, with the little gain in accuracy not justifying the poor run time of ViT when trained from scratch on mid-sized datasets. However, transfer learning and pre-training on large datasets largely mitigates this limitation.</p>
</sec>
<sec id="s13f4"><label>12.6.4</label><title>Architectural variants and improvements of ViTs</title>
<p>Recent years have seen efficient architectures like Swin Transformer, which introduced a hierarchical structure and shifted windows to reduce computational cost while maintaining strong performance on dense prediction tasks (<xref ref-type="bibr" rid="B266">266</xref>). CSWin Transformer achieved 85.4&#x0025; Top-1 accuracy on ImageNet-1K, 53.9 box AP and 46.4 mask AP on COCO detection, and 52.2 mIOU on ADE20K semantic segmentation. Hybrid approaches are also emerging. ViT-CoMer injects spatial pyramid multi-receptive field convolutional features into the ViT architecture, which effectively alleviates problems of limited local information interaction and single-feature representation in ViT.</p>
</sec>
<sec id="s13f5"><label>12.6.5</label><title>ViT vs. CNN: fundamental differences</title>
<sec id="s13f5a"><label>12.6.5.1</label><title>Information processing paradigms</title>
<p>While CNNs start with very low-level local information and gradually build up more global structures in deeper layers, ViTs already have global information at the earliest layer thanks to global self-attention (<xref ref-type="bibr" rid="B267">267</xref>). This difference is profound: A CNN&#x0027;s approach is like starting at a single pixel and zooming out, while a transformer slowly brings the whole fuzzy image into focus. ViTs utilize a self-attention mechanism that enables the model to have a whole field of view even at the lowest layer, obtaining global representations from the beginning, while CNNs need to propagate layers to obtain global representations.</p>
</sec>
<sec id="s13f5b"><label>12.6.5.2</label><title>Inductive biases and generalization</title>
<p>CNNs have an inductive spatial bias baked into them with convolutional kernels, whereas vision transformers are based on a much more general architecture, with first vision transformers beating CNNs by overcoming the spatial bias given enough data. The main advantage of ViTs is their ability to effectively capture global contextual information through the self-attention mechanism, enabling them to model long-range dependencies and contextual relationships, which can improve robustness in tasks requiring understanding global context.</p>
</sec>
<sec id="s13f5c"><label>12.6.5.3</label><title>Computational complexity</title>
<p>The self-attention mechanism&#x0027;s quadratic complexity with respect to input size creates computational challenges. ViTs are computationally intensive, especially due to the self-attention mechanism, which has quadratic complexity with respect to the number of patches (<xref ref-type="bibr" rid="B268">268</xref>). However, self-attention-based models are highly parallelizable and require substantially fewer parameters, making them much more computationally efficient, less prone to overfitting, and easier to fine-tune for domain-specific tasks.</p>
</sec>
<sec id="s13f5d"><label>12.6.5.4</label><title>Practical performance comparisons</title>
<p>ViTs consistently outperform CNNs when trained on large datasets due to their ability to model long-range dependencies via self-attention mechanisms (<xref ref-type="bibr" rid="B269">269</xref>). In medical imaging specifically, transformer-based models, particularly ViTs, exhibit significant potential in diverse medical imaging tasks, showcasing superior performance when contrasted with conventional CNN models across 36 reviewed studies (<xref ref-type="bibr" rid="B217">217</xref>).</p>
</sec>
</sec>
</sec>
<sec id="s13g"><label>12.7</label><title>Segment anything model (SAM): universal segmentation</title>
<sec id="s13g1"><label>12.7.1</label><title>Segment anything model (SAM): universal segmentation</title>
<p>SAM represents a paradigm shift toward foundation models for computer vision (<xref ref-type="bibr" rid="B270">270</xref>). SAM&#x0027;s architecture comprises three components that work together to return a valid segmentation mask: an image encoder to generate one-time image embedding, a prompt encoder that embeds the prompts, and a mask decoder. SAM is a foundation model for segmentation trained on 11 million images and over 1 billion masks, composed of three primary modules: an image encoder, a prompt encoder, and a mask decoder (<xref ref-type="bibr" rid="B271">271</xref>). The lightweight mask decoder enables rapid inference once image embeddings are computed.</p>
</sec>
<sec id="s13g2"><label>12.7.2</label><title>Training data and zero-shot performance</title>
<p>Trained on the expansive SA-1B dataset, SAM excels in zero-shot performance, adapting to new image distributions and tasks without prior knowledge (<xref ref-type="bibr" rid="B272">272</xref>). SAM has been trained on a dataset of 11 million images and 1.1 billion masks and has strong zero-shot performance on a variety of segmentation tasks (<xref ref-type="bibr" rid="B271">271</xref>). Utilizing the extensive SA-1B dataset, comprising over 11 million meticulously curated images with more than 1 billion masks, SAM has demonstrated remarkable zero-shot performance, often surpassing previous fully supervised results (<xref ref-type="bibr" rid="B272">272</xref>).</p>
</sec>
<sec id="s13g3"><label>12.7.3</label><title>Evolution: SAM 2 and SAM 3</title>
<p>The SAM family has rapidly evolved. SAM 2 processes approximately 44 frames per second, making it suitable for applications requiring immediate feedback like video editing and augmented reality. SAM 2 extends the original model to video by treating images as single-frame videos and incorporating streaming memory for temporal consistency (<xref ref-type="bibr" rid="B273">273</xref>). The most recent iteration introduces revolutionary capabilities. SAM 3 delivers strong rare and unseen object generalization and high prompt reliability, in addition to stronger performance with thin, small, low contrast, and occluded objects compared to SAM 2 and SAM 1. SAM 3 can detect, segment, and track objects using text or visual prompts, introducing the ability to exhaustively segment all instances of an open-vocabulary concept specified by a short text phrase or exemplars. Unlike prior work, SAM 3 can handle a vastly larger set of open-vocabulary prompts, achieving 75&#x0025;&#x2013;80&#x0025; of human performance on the SA-CO benchmark which contains 270K unique concepts, over 50 times more than existing benchmarks.</p>
</sec>
<sec id="s13g4"><label>12.7.4</label><title>Applications and limitations in SAM</title>
<p>SAM&#x0027;s promotable segmentation enables diverse applications. With the adaptation of SAM to medical imaging, MedSAM emerges as the first foundation model for universal medical image segmentation, consistently doing better than the best segmentation foundation models when tested on a wide range of validation tasks (<xref ref-type="bibr" rid="B274">274</xref>). However, SAM&#x0027;s inference speed is quite fast, generating a segmentation result in 50 milliseconds for any prompt in the web browser with CPU, though this assumes the image embedding is already precomputed. The quality may be insufficient for high-precision applications requiring near-pixel-perfect predictions.</p>
</sec>
</sec>
<sec id="s13h"><label>12.8</label><title>Swin-U-Net: transformers for medical image segmentation</title>
<sec id="s13h1"><label>12.8.1</label><title>Architecture design in Swin-U-Net</title>
<p>Swin-U-Net adapts the Swin Transformer&#x0027;s hierarchical architecture to medical segmentation (<xref ref-type="bibr" rid="B275">275</xref>). It uses hierarchical Swin Transformer with shifted windows as the encoder to extract context features, and a symmetric Swin Transformer-based decoder with patch expanding layer is designed to perform the up-sampling operation to restore the spatial resolution of the feature maps (<xref ref-type="bibr" rid="B276">276</xref>). The model uses 224&#x2009;&#x00D7;&#x2009;224 input images with 4&#x2009;&#x00D7;&#x2009;4 patches, creating tokenized inputs at H/4&#x00D7;W/4 resolution (<xref ref-type="bibr" rid="B277">277</xref>). The encoder applies patch merging layers for 2&#x00D7; downsampling while doubling feature dimensions, repeated three times (<xref ref-type="bibr" rid="B278">278</xref>). The decoder mirrors this structure with patch expanding layers for upsampling (<xref ref-type="bibr" rid="B279">279</xref>).</p>
</sec>
<sec id="s13h2"><label>12.8.2.</label><title>Performance in medical imaging in Swin-U-Net</title>
<p>Swin-U-Net achieves the best performance with segmentation accuracy, with pure Transformer approaches without convolution better learning both global and long-range semantic information interactions, resulting in better segmentation results (<xref ref-type="bibr" rid="B275">275</xref>). Swin-Unet achieves excellent performance with an accuracy of 90.00&#x0025; on the ACDC dataset, showing good generalization ability and robustness (<xref ref-type="bibr" rid="B280">280</xref>). Experiments on multi-organ and cardiac segmentation tasks demonstrate that the pure Transformer-based U-shaped Encoder-Decoder network outperforms those methods with full-convolution or the combination of transformer and convolution (<xref ref-type="bibr" rid="B125">125</xref>).</p>
</sec>
<sec id="s13h3"><label>12.8.3</label><title>Hybrid approaches and extensions in Swin-U-Net</title>
<p>The debate between pure transformers and hybrid architecture continues. Swin Pure U-Net3D vs. Swin Unet3D performance differences indicate that the convolutional module can compensate for ViT&#x0027;s inability to fit the image detail information well. Advanced variants show further improvements. FE-SwinUper achieves excellent performance with a Dice of 90.15&#x0025; on the ACDC dataset, showing good generalization ability and robustness (<xref ref-type="bibr" rid="B280">280</xref>). A multi-transformer U-Net surpasses standalone Swin Transformer&#x0027;s Swin Unet and converges more rapidly, yielding accuracy improvements of 0.7&#x0025; (resulting in 88.18&#x0025;) and 2.7&#x0025; (resulting in 98.01&#x0025;) on COVID-19 CT scan and Chest x-ray datasets respectively (<xref ref-type="bibr" rid="B281">281</xref>).</p>
</sec>
<sec id="s13h4"><label>12.8.4</label><title>3D extensions</title>
<p>Swin UNETR employs MONAI and has achieved state-of-the-art benchmarks for various medical image segmentation tasks, demonstrating effectiveness even with a small amount of labeled data (<xref ref-type="bibr" rid="B282">282</xref>). The model was pretrained on 5,050 publicly available CT images and achieved top rankings on the BTCV Segmentation Challenge and Medical Segmentation Decathlon dataset. Swin UNETR has shown better segmentation performance using significantly fewer training GPU hours compared to DiNTS&#x2014;a powerful AutoML methodology for medical image segmentation, making it practical for resource-constrained medical imaging applications (<xref ref-type="bibr" rid="B283">283</xref>).</p>
</sec>
</sec>
<sec id="s13i"><label>12.9</label><title>Critical insights and future directions</title>
<sec id="s13i1"><label>12.9.1</label><title>Architectural evolution</title>
<p>The transformer revolution in computer vision demonstrates several key principles. First, removing inductive biases increases model capacity and generalization at the cost of data efficiency (<xref ref-type="bibr" rid="B284">284</xref>). Second, hierarchical architectures like Swin Transformer successfully balance global and local feature extraction (<xref ref-type="bibr" rid="B285">285</xref>). Third, hybrid approaches combining transformers with convolutions often achieve optimal performance-efficiency trade-offs (<xref ref-type="bibr" rid="B268">268</xref>).</p>
</sec>
<sec id="s13i2"><label>12.9.2</label><title>The foundation model paradigm</title>
<p>SAM exemplifies the shift toward foundation models trained on massive datasets for zero-shot generalization (<xref ref-type="bibr" rid="B286">286</xref>). This approach democratizes computer vision by enabling practitioners to apply powerful models without task-specific training (<xref ref-type="bibr" rid="B287">287</xref>). However, questions remain about performance on specialized domains and the computational costs of inference.</p>
</sec>
<sec id="s13i3"><label>12.9.3</label><title>Medical imaging considerations</title>
<p>In medical imaging, pure transformer architectures like Swin-U-Net demonstrate that global context modeling significantly benefits segmentation tasks (<xref ref-type="bibr" rid="B288">288</xref>). The ability to capture long-range dependencies helps identify anatomical structures and pathologies that span large image regions. Yet hybrid architecture often performs better when fine detail preservation is critical.</p>
</sec>
<sec id="s134"><label>12.9.4.</label><title>Computational reality</title>
<p>Despite theoretical advantages, practical deployment requires balancing accuracy against computational constraints. The quadratic complexity of self-attention remains a bottleneck for high-resolution images, driving research into efficient attention mechanisms and hybrid architectures that selectively apply global attention (<xref ref-type="bibr" rid="B289">289</xref>). The transformer-based revolution in computer vision continues to accelerate, with each generation of models&#x2014;from ViT to Swin Transformers to SAM 3&#x2014;pushing boundaries in performance, efficiency, and generalization. The field is converging toward flexible architectures that adaptively combine local and global processing, learned through self-supervised pretraining on vast datasets, enabling unprecedented capabilities across diverse vision tasks.</p>
</sec>
</sec>
<sec id="s13k"><label>12.10</label><title>Specialized applications</title>
<p>In 3D medical imaging, recent advances in AI, especially vision-language foundation models, show promise in automating radiology report generation from complex 3D medical imaging data, with studies analyzing model architecture, capabilities, training datasets, and evaluation metrics (<xref ref-type="bibr" rid="B290">290</xref>). As domain-specific innovations, there are some technical analyses, such as brain tumor classification with explainable AI, stroke outcome prediction using lesion-centered approaches, lung cancer screening and detection, cardiac imaging analysis, and pathology image analysis (<xref ref-type="bibr" rid="B291">291</xref>).</p>
</sec>
<sec id="s13l"><label>12.11</label><title>Practical deployment challenges</title>
<p>As for current limitations, practical implementation of large models in medical imaging faces notable challenges, including scarcity of high-quality medical data, need for optimized perception of imaging phenotypes, safety considerations, and seamless integration with existing clinical workflows and equipment (<xref ref-type="bibr" rid="B292">292</xref>). There are some emerging solutions, including federated learning for privacy-preserving collaborative training, edge deployment for real-time analysis, integration with PACS and clinical workflows, and continuous model monitoring and updating (<xref ref-type="bibr" rid="B293">293</xref>).</p>
</sec>
<sec id="s13m"><label>12.12</label><title>Future directions</title>
<p>AI algorithms will not only highlight abnormalities but also suggest potential diagnoses and probabilities, with natural language processing streamlining reporting processes by automatically generating structured reports, and advanced machine learning techniques comparing patient scans against extensive databases (<xref ref-type="bibr" rid="B29">29</xref>). There are some emerging technologies, such as photon-counting CT, whole-body MRI, AI-enabled point-of-care devices, generative AI for patient communication, and agentic AI systems for complex workflows. As pivotal takeaways, transformers are surpassing CNNs for many medical imaging tasks, particularly when combined with proper pre-training (<xref ref-type="bibr" rid="B217">217</xref>). Foundation models are democratizing AI in healthcare by requiring fewer local data for fine-tuning (<xref ref-type="bibr" rid="B294">294</xref>). Diffusion models are emerging as superior alternatives to GANs for synthetic data generation (<xref ref-type="bibr" rid="B295">295</xref>). XAI is critical but must be designed with human-centered principles and validated with end users. Bias and fairness require continuous monitoring across deployment contexts, not just initial training data. Multi-modal integration (imaging&#x2009;&#x002B;&#x2009;clinical&#x2009;&#x002B;&#x2009;genomic) is becoming standard (<xref ref-type="bibr" rid="B296">296</xref>). Physics-informed approaches reduce data requirements while improving interpretability. The field is rapidly evolving toward more generalizable, interpretable, and fair AI systems that can be deployed safely in diverse clinical settings.</p>
</sec>
</sec>
<sec id="s14" sec-type="conclusions"><label>13</label><title>Conclusions</title>
<p>Computer vision engineers develop algorithms, software, and hardware systems that enable machines to process, analyze, and make decisions based on visual data, such as images and videos. CNNs excel at tasks such as image segmentation, feature extraction, and classification and analyze various medical images well, including x-rays, MRIs, CT scans, and ultrasound images. As technology continues evolving, computer vision in healthcare will further transform patient care, medical research, and healthcare delivery, ultimately improving health outcomes and healthcare system efficiency.</p>
<p>CNNs are powerful DL models that have revolutionized several fields, particularly image and video processing. CNNs automatically learn features from data, making them highly effective for tasks such as image recognition, classification, and segmentation. Nonetheless, CNNs are still evolving, and ongoing research focuses on improving their efficiency, accuracy, and applicability to diverse problem domains. As the DL field advances, CNN models may play an increasingly important role in solving complex real-world problems in various industries. The CNN architecture U-Net has revolutionized biomedical image segmentation. This powerful model is the foundation of numerous advances in medical image analysis because of its ability to segment rapidly and accurately. U-Net is a cornerstone in the development of advanced medical image segmentation techniques, driving advances in computer-aided diagnosis and precision medicine.</p>
<p>Foundation models are set to redefine how radiologists approach diagnostics and patient care, with these advanced systems capable of handling a wide range of tasks. Vision-language models integrate computer vision and natural language processing to address complex tasks such as disease classification, segmentation, cross-modal retrieval, and automated report generation. Transformer-based multimodal predictions consistently find that transformer models outperform typical recurrent or unimodal models.</p>
</sec>
</body>
<back>
<sec id="s15" sec-type="author-contributions"><title>Author contributions</title>
<p>YM: Conceptualization, Validation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. MI: Conceptualization, Supervision, Validation, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<ack><title>Acknowledgments</title>
<p>The authors thank Enago (<ext-link ext-link-type="uri" xlink:href="https://www.enago.jp">https://www.enago.jp</ext-link>) for English language editing.</p>
</ack>
<sec id="s17" sec-type="COI-statement"><title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s18" sec-type="ai-statement"><title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec id="s19" sec-type="disclaimer"><title>Publisher&#x0027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list><title>References</title>
<ref id="B1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Javaid</surname> <given-names>M</given-names></name> <name><surname>Haleem</surname> <given-names>A</given-names></name> <name><surname>Singh</surname> <given-names>RP</given-names></name> <name><surname>Ahmed</surname> <given-names>M</given-names></name></person-group>. <article-title>Computer vision to enhance healthcare domain: an overview of features, implementation, and opportunities</article-title>. <source>Intell Pharm</source>. (<year>2024</year>) <volume>2</volume>(<issue>6</issue>):<fpage>792</fpage>&#x2013;<lpage>803</lpage>. <pub-id pub-id-type="doi">10.1016/j.ipha.2024.05.007</pub-id></mixed-citation></ref>
<ref id="B2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Olveres</surname> <given-names>J</given-names></name> <name><surname>Gonz&#x00E1;lez</surname> <given-names>G</given-names></name> <name><surname>Torres</surname> <given-names>F</given-names></name> <name><surname>Moreno-Tagle</surname> <given-names>JC</given-names></name> <name><surname>Carbajal-Degante</surname> <given-names>E</given-names></name> <name><surname>Valencia-Rodr&#x00ED;guez</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>What is new in computer vision and artificial intelligence in medical image analysis applications</article-title>. <source>Quant Imaging Med Surg</source>. (<year>2021</year>) <volume>11</volume>(<issue>8</issue>):<fpage>3830</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.21037/qims-20-1151</pub-id><pub-id pub-id-type="pmid">34341753</pub-id></mixed-citation></ref>
<ref id="B3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Elyan</surname> <given-names>E</given-names></name> <name><surname>Vuttipittayamongkol</surname> <given-names>P</given-names></name> <name><surname>Johnston</surname> <given-names>P</given-names></name> <name><surname>Martin</surname> <given-names>K</given-names></name> <name><surname>McPherson</surname> <given-names>K</given-names></name> <name><surname>Moreno-Garc&#x00ED;a</surname> <given-names>CF</given-names></name><etal/></person-group> <article-title>Computer vision and machine learning for medical image analysis: recent advances, challenges, and way forward</article-title>. <source>Artif Intell Surg</source>. (<year>2022</year>) <volume>2</volume>:<fpage>22</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.20517/ais.2021.15</pub-id></mixed-citation></ref>
<ref id="B4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>X</given-names></name> <name><surname>Hu</surname> <given-names>Z</given-names></name> <name><surname>Wang</surname> <given-names>S</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name></person-group>. <article-title>Deep learning for medical image-based cancer diagnosis</article-title>. <source>Cancers (Basel)</source>. (<year>2023</year>) <volume>15</volume>(<issue>14</issue>):<fpage>3609</fpage>. <pub-id pub-id-type="doi">10.3390/cancers15143608</pub-id><pub-id pub-id-type="pmid">37509271</pub-id></mixed-citation></ref>
<ref id="B5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Heinrich</surname> <given-names>K</given-names></name> <name><surname>Janiesch</surname> <given-names>C</given-names></name> <name><surname>Krancher</surname> <given-names>O</given-names></name> <name><surname>Stahmann</surname> <given-names>P</given-names></name> <name><surname>Wanner</surname> <given-names>J</given-names></name> <name><surname>Zschech</surname> <given-names>P</given-names></name></person-group>. <article-title>Decision factors for the selection of AI-based decision support systems-the case of task delegation in prognostics</article-title>. <source>PLoS One</source>. (<year>2025</year>) <volume>20</volume>(<issue>7</issue>):<fpage>e0328411</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0328411</pub-id><pub-id pub-id-type="pmid">40705766</pub-id></mixed-citation></ref>
<ref id="B6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Amirgaliyev</surname> <given-names>B</given-names></name> <name><surname>Mussabek</surname> <given-names>M</given-names></name> <name><surname>Rakhimzhanova</surname> <given-names>T</given-names></name> <name><surname>Zhumadillayeva</surname> <given-names>A</given-names></name></person-group>. <article-title>A review of machine learning and deep learning methods for person detection. Tracking and identification, and face recognition with applications</article-title><source>. Sensors (Basel)</source>. (<year>2025</year>) <volume>25</volume>(<issue>5</issue>):<fpage>1410</fpage>. <pub-id pub-id-type="doi">10.3390/s25051410</pub-id><pub-id pub-id-type="pmid">40096196</pub-id></mixed-citation></ref>
<ref id="B7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hilal</surname> <given-names>AM</given-names></name> <name><surname>Albalawneh</surname> <given-names>D</given-names></name> <name><surname>Bouchelligua</surname> <given-names>W</given-names></name> <name><surname>Sharif</surname> <given-names>MM</given-names></name></person-group>. <article-title>Multi-strategy dung beetle optimization for robust indoor object detection and tracking for visually impaired people with hybrid deep learning networks</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>36516</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-08155-3</pub-id><pub-id pub-id-type="pmid">41120381</pub-id></mixed-citation></ref>
<ref id="B8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Wu</surname> <given-names>J</given-names></name> <name><surname>Cai</surname> <given-names>Y</given-names></name> <name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Chen</surname> <given-names>L</given-names></name> <name><surname>Liu</surname> <given-names>Q</given-names></name></person-group>. <article-title>Dual-stage feature specialization network for robust visual object detection in autonomous vehicles</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>15501</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-99363-4</pub-id><pub-id pub-id-type="pmid">40319138</pub-id></mixed-citation></ref>
<ref id="B9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fathy</surname> <given-names>ME</given-names></name> <name><surname>Mohamed</surname> <given-names>SA</given-names></name> <name><surname>Awad</surname> <given-names>MI</given-names></name></person-group>. <article-title>Abd el munim HE. A vision transformer based CNN for underwater image enhancement ViTClarityNet</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>16768</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-91212-8</pub-id><pub-id pub-id-type="pmid">40369132</pub-id></mixed-citation></ref>
<ref id="B10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sarker</surname> <given-names>IH</given-names></name></person-group>. <article-title>Deep learning: a comprehensive overview on techniques, taxonomy. Applications and Research Directions</article-title>. <source>SN Comput Sci</source>. (<year>2021</year>) <volume>2</volume>(<issue>6</issue>):<fpage>420</fpage>. <pub-id pub-id-type="doi">10.1007/s42979-021-00815-1</pub-id><pub-id pub-id-type="pmid">34426802</pub-id></mixed-citation></ref>
<ref id="B11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hadi</surname> <given-names>SJ</given-names></name> <name><surname>Ahmed</surname> <given-names>I</given-names></name> <name><surname>Iqbal</surname> <given-names>A</given-names></name> <name><surname>Alzahrani</surname> <given-names>AS</given-names></name></person-group>. <article-title>CATR: CNN augmented transformer for object detection in remote sensing imagery</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>42281</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-27872-3</pub-id><pub-id pub-id-type="pmid">41309916</pub-id></mixed-citation></ref>
<ref id="B12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>C</given-names></name> <name><surname>Liu</surname> <given-names>L</given-names></name></person-group>. <article-title>Machine learning prediction model for medical environment comfort based on SHAP and LIME interpretability analysis</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>39269</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-22972-6</pub-id><pub-id pub-id-type="pmid">41214038</pub-id></mixed-citation></ref>
<ref id="B13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fanariotis</surname> <given-names>A</given-names></name> <name><surname>Orphanoudakis</surname> <given-names>T</given-names></name> <name><surname>Kotrotsios</surname> <given-names>K</given-names></name> <name><surname>Fotopoulos</surname> <given-names>V</given-names></name> <name><surname>Keramidas</surname> <given-names>G</given-names></name> <name><surname>Karkazis</surname> <given-names>P</given-names></name></person-group>. <article-title>Power efficient machine learning models deployment on edge IoT devices</article-title>. <source>Sensors (Basel)</source>. (<year>2023</year>) <volume>23</volume>(<issue>3</issue>):<fpage>1595</fpage>. <pub-id pub-id-type="doi">10.3390/s23031595</pub-id><pub-id pub-id-type="pmid">36772635</pub-id></mixed-citation></ref>
<ref id="B14"><label>14.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kadeethum</surname> <given-names>T</given-names></name> <name><surname>O&#x0027;Malley</surname> <given-names>D</given-names></name> <name><surname>Choi</surname> <given-names>Y</given-names></name> <name><surname>Viswanathan</surname> <given-names>HS</given-names></name> <name><surname>Yoon</surname> <given-names>H</given-names></name></person-group>. <article-title>Progressive transfer learning for advancing machine learning-based reduced-order modeling</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>15731</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-64778-y</pub-id><pub-id pub-id-type="pmid">38977759</pub-id></mixed-citation></ref>
<ref id="B15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>S AR</surname> <given-names>EE</given-names></name> <name><surname>Balakrishnan</surname> <given-names>A</given-names></name> <name><surname>Sanisetty</surname> <given-names>B</given-names></name> <name><surname>Bandaru</surname> <given-names>RB</given-names></name></person-group>. <article-title>Stacked hybrid model for load forecasting: integrating transformers, ANN, and fuzzy logic</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>19688</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-04210-1</pub-id><pub-id pub-id-type="pmid">40467760</pub-id></mixed-citation></ref>
<ref id="B16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thakur</surname> <given-names>GK</given-names></name> <name><surname>Thakur</surname> <given-names>A</given-names></name> <name><surname>Kulkarni</surname> <given-names>S</given-names></name> <name><surname>Khan</surname> <given-names>N</given-names></name> <name><surname>Khan</surname> <given-names>S</given-names></name></person-group>. <article-title>Deep learning approaches for medical image analysis and diagnosis</article-title>. <source>Cureus</source>. (<year>2024</year>) <volume>16</volume>(<issue>5</issue>):<fpage>e59507</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.59507</pub-id><pub-id pub-id-type="pmid">38826977</pub-id></mixed-citation></ref>
<ref id="B17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>M</given-names></name> <name><surname>Jiang</surname> <given-names>Y</given-names></name> <name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Zhu</surname> <given-names>H</given-names></name></person-group>. <article-title>Medical image analysis using deep learning algorithms</article-title>. <source>Front Public Health</source>. (<year>2023</year>) <volume>11</volume>:<fpage>1273253</fpage>. <pub-id pub-id-type="doi">10.3389/fpubh.2023.1273253</pub-id><pub-id pub-id-type="pmid">38026291</pub-id></mixed-citation></ref>
<ref id="B18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Goo</surname> <given-names>HW</given-names></name> <name><surname>Park</surname> <given-names>SJ</given-names></name> <name><surname>Yoo</surname> <given-names>SJ</given-names></name></person-group>. <article-title>Advanced medical use of three-dimensional imaging in congenital heart disease: augmented reality, mixed reality, virtual reality, and three-dimensional printing</article-title>. <source>Korean J Radiol</source>. (<year>2020</year>) <volume>21</volume>(<issue>2</issue>):<fpage>133</fpage>&#x2013;<lpage>45</lpage>. <pub-id pub-id-type="doi">10.3348/kjr.2019.0625</pub-id><pub-id pub-id-type="pmid">31997589</pub-id></mixed-citation></ref>
<ref id="B19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Salvestrini</surname> <given-names>V</given-names></name> <name><surname>Lastrucci</surname> <given-names>A</given-names></name> <name><surname>Banini</surname> <given-names>M</given-names></name> <name><surname>Loi</surname> <given-names>M</given-names></name> <name><surname>Carnevale</surname> <given-names>MG</given-names></name> <name><surname>Olmetto</surname> <given-names>E</given-names></name><etal/></person-group> <article-title>Recent advances and current challenges in stereotactic body radiotherapy for ultra-central lung tumors</article-title>. <source>Cancers (Basel)</source>. (<year>2024</year>) <volume>16</volume>(<issue>24</issue>):<fpage>4135</fpage>. <pub-id pub-id-type="doi">10.3390/cancers16244135</pub-id><pub-id pub-id-type="pmid">39766035</pub-id></mixed-citation></ref>
<ref id="B20"><label>20.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sch&#x00E4;fer</surname> <given-names>R</given-names></name> <name><surname>Nicke</surname> <given-names>T</given-names></name> <name><surname>H&#x00F6;fener</surname> <given-names>H</given-names></name> <name><surname>Lange</surname> <given-names>A</given-names></name> <name><surname>Merhof</surname> <given-names>D</given-names></name> <name><surname>Feuerhake</surname> <given-names>F</given-names></name><etal/></person-group> <article-title>Overcoming data scarcity in biomedical imaging with a foundational multi-task model</article-title>. <source>Nat Comput Sci</source>. (<year>2024</year>) <volume>4</volume>(<issue>7</issue>):<fpage>495</fpage>&#x2013;<lpage>509</lpage>. <pub-id pub-id-type="doi">10.1038/s43588-024-00662-z</pub-id></mixed-citation></ref>
<ref id="B21"><label>21.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Herath</surname> <given-names>H</given-names></name> <name><surname>Herath</surname> <given-names>H</given-names></name> <name><surname>Madusanka</surname> <given-names>N</given-names></name> <name><surname>Lee</surname> <given-names>BI</given-names></name></person-group>. <article-title>A systematic review of medical image quality assessment</article-title>. <source>J Imaging</source>. (<year>2025</year>) <volume>11</volume>(<issue>4</issue>):<fpage>100</fpage>. <pub-id pub-id-type="doi">10.3390/jimaging11040100</pub-id><pub-id pub-id-type="pmid">40278016</pub-id></mixed-citation></ref>
<ref id="B22"><label>22.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Guan</surname> <given-names>H</given-names></name> <name><surname>Liu</surname> <given-names>M</given-names></name></person-group>. <article-title>Domain adaptation for medical image analysis: a survey</article-title>. <source>IEEE Trans Biomed Eng</source>. (<year>2022</year>) <volume>69</volume>(<issue>3</issue>):<fpage>1173</fpage>&#x2013;<lpage>85</lpage>. <pub-id pub-id-type="doi">10.1109/tbme.2021.3117407</pub-id><pub-id pub-id-type="pmid">34606445</pub-id></mixed-citation></ref>
<ref id="B23"><label>23.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Byrne</surname> <given-names>R</given-names></name> <name><surname>Schneider</surname> <given-names>G</given-names></name> <name><surname>Yang</surname> <given-names>S</given-names></name></person-group>. <article-title>Concepts of artificial intelligence for computer-assisted drug discovery</article-title>. <source>Chem Rev</source>. (<year>2019</year>) <volume>119</volume>(<issue>18</issue>):<fpage>10520</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1021/acs.chemrev.8b00728</pub-id><pub-id pub-id-type="pmid">31294972</pub-id></mixed-citation></ref>
<ref id="B24"><label>24.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>AQ</given-names></name> <name><surname>Karaman</surname> <given-names>BK</given-names></name> <name><surname>Kim</surname> <given-names>H</given-names></name> <name><surname>Rosenthal</surname> <given-names>J</given-names></name> <name><surname>Saluja</surname> <given-names>R</given-names></name> <name><surname>Young</surname> <given-names>SI</given-names></name><etal/></person-group> <article-title>A framework for interpretability in machine learning for medical imaging</article-title>. <source>IEEE Access</source>. (<year>2024</year>) <volume>12</volume>:<fpage>53277</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1109/access.2024.3387702</pub-id><pub-id pub-id-type="pmid">39421804</pub-id></mixed-citation></ref>
<ref id="B25"><label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ullah</surname> <given-names>N</given-names></name> <name><surname>Guzm&#x00E1;n-Aroca</surname> <given-names>F</given-names></name> <name><surname>Mart&#x00ED;nez-&#x00C1;lvarez</surname> <given-names>F</given-names></name> <name><surname>De Falco</surname> <given-names>I</given-names></name> <name><surname>Sannino</surname> <given-names>G</given-names></name></person-group>. <article-title>A novel explainable AI framework for medical image classification integrating statistical, visual, and rule-based methods</article-title>. <source>Med Image Anal</source>. (<year>2025</year>) <volume>105</volume>:<fpage>103665</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2025.103665</pub-id><pub-id pub-id-type="pmid">40505210</pub-id></mixed-citation></ref>
<ref id="B26"><label>26.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bortsova</surname> <given-names>G</given-names></name> <name><surname>Gonz&#x00E1;lez-Gonzalo</surname> <given-names>C</given-names></name> <name><surname>Wetstein</surname> <given-names>SC</given-names></name> <name><surname>Dubost</surname> <given-names>F</given-names></name> <name><surname>Katramados</surname> <given-names>I</given-names></name> <name><surname>Hogeweg</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Adversarial attack vulnerability of medical image analysis systems: unexplored factors</article-title>. <source>Med Image Anal</source>. (<year>2021</year>) <volume>73</volume>:<fpage>102141</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.102141</pub-id><pub-id pub-id-type="pmid">34246850</pub-id></mixed-citation></ref>
<ref id="B27"><label>27.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yinusa</surname> <given-names>A</given-names></name> <name><surname>Faezipour</surname> <given-names>M</given-names></name></person-group>. <article-title>A multi-layered defense against adversarial attacks in brain tumor classification using ensemble adversarial training and feature squeezing</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>16804</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-00890-x</pub-id><pub-id pub-id-type="pmid">40369011</pub-id></mixed-citation></ref>
<ref id="B28"><label>28.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Datta</surname> <given-names>SD</given-names></name> <name><surname>Islam</surname> <given-names>M</given-names></name> <name><surname>Rahman Sobuz</surname> <given-names>MH</given-names></name> <name><surname>Ahmed</surname> <given-names>S</given-names></name> <name><surname>Kar</surname> <given-names>M</given-names></name></person-group>. <article-title>Artificial intelligence and machine learning applications in the project lifecycle of the construction industry: a comprehensive review</article-title>. <source>Heliyon</source>. (<year>2024</year>) <volume>10</volume>(<issue>5</issue>):<fpage>e26888</fpage>. <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e26888</pub-id><pub-id pub-id-type="pmid">38444479</pub-id></mixed-citation></ref>
<ref id="B29"><label>29.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pinto-Coelho</surname> <given-names>L</given-names></name></person-group>. <article-title>How artificial intelligence is shaping medical imaging technology: a survey of innovations and applications</article-title>. <source>Bioengineering (Basel)</source>. (<year>2023</year>) <volume>10</volume>(<issue>12</issue>):<fpage>1435</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering10121435</pub-id><pub-id pub-id-type="pmid">38136026</pub-id></mixed-citation></ref>
<ref id="B30"><label>30.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Prasad</surname> <given-names>VK</given-names></name> <name><surname>Verma</surname> <given-names>A</given-names></name> <name><surname>Bhattacharya</surname> <given-names>P</given-names></name> <name><surname>Shah</surname> <given-names>S</given-names></name> <name><surname>Chowdhury</surname> <given-names>S</given-names></name> <name><surname>Bhavsar</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Revolutionizing healthcare: a comparative insight into deep learning&#x2019;s role in medical imaging</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>30273</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-71358-7</pub-id><pub-id pub-id-type="pmid">39632902</pub-id></mixed-citation></ref>
<ref id="B31"><label>31.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rayed</surname> <given-names>ME</given-names></name> <name><surname>Islam</surname> <given-names>SNS</given-names></name> <name><surname>Niha</surname> <given-names>SI</given-names></name> <name><surname>Jim</surname> <given-names>JR</given-names></name> <name><surname>Kabir</surname> <given-names>M</given-names></name> <name><surname>Mridha</surname> <given-names>MF</given-names></name></person-group>. <article-title>Deep learning for medical image segmentation: state-of-the-art advancements and challenges</article-title>. <source>Inform Med Unlocked</source>. (<year>2024</year>) <volume>47</volume>:<fpage>101504</fpage>. <pub-id pub-id-type="doi">10.1016/j.imu.2024.101504</pub-id></mixed-citation></ref>
<ref id="B32"><label>32.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>Y</given-names></name> <name><surname>Quan</surname> <given-names>R</given-names></name> <name><surname>Xu</surname> <given-names>W</given-names></name> <name><surname>Huang</surname> <given-names>Y</given-names></name> <name><surname>Chen</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>F</given-names></name></person-group>. <article-title>Advances in medical image segmentation: a comprehensive review of traditional, deep learning and hybrid approaches</article-title>. <source>Bioengineering (Basel)</source>. (<year>2024</year>) <volume>11</volume>(<issue>10</issue>):<fpage>1034</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering11101034</pub-id><pub-id pub-id-type="pmid">39451409</pub-id></mixed-citation></ref>
<ref id="B33"><label>33.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bhandari</surname> <given-names>A</given-names></name></person-group>. <article-title>Revolutionizing radiology with artificial intelligence</article-title>. <source>Cureus</source>. (<year>2024</year>) <volume>16</volume>(<issue>10</issue>):<fpage>e72646</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.72646</pub-id><pub-id pub-id-type="pmid">39474591</pub-id></mixed-citation></ref>
<ref id="B34"><label>34.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Sun</surname> <given-names>Q</given-names></name> <name><surname>Akman</surname> <given-names>A</given-names></name> <name><surname>Schuller</surname> <given-names>BW</given-names></name></person-group>. <comment>Explainable artificial intelligence for medical applications: a review. (2024). <italic>arXiv PreprintarXiv</italic>:241201829v1</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2412.01829</pub-id>. <comment>(Accessed November 15, 2024)</comment></mixed-citation></ref>
<ref id="B35"><label>35.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kitamura</surname> <given-names>FC</given-names></name> <name><surname>Prevedello</surname> <given-names>LM</given-names></name> <name><surname>Colak</surname> <given-names>E</given-names></name> <name><surname>Halabi</surname> <given-names>SS</given-names></name> <name><surname>Lungren</surname> <given-names>MP</given-names></name> <name><surname>Ball</surname> <given-names>RL</given-names></name><etal/></person-group> <article-title>Lessons learned in building expertly annotated multi-institution datasets and hosting the RSNA AI challenges</article-title>. <source>Radiol Artif Intell</source>. (<year>2024</year>) <volume>6</volume>(<issue>3</issue>):<fpage>e230227</fpage>. <pub-id pub-id-type="doi">10.1148/ryai.230227</pub-id><pub-id pub-id-type="pmid">38477659</pub-id></mixed-citation></ref>
<ref id="B36"><label>36.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Seh</surname> <given-names>AH</given-names></name> <name><surname>Zarour</surname> <given-names>M</given-names></name> <name><surname>Alenezi</surname> <given-names>M</given-names></name> <name><surname>Sarkar</surname> <given-names>AK</given-names></name> <name><surname>Agrawal</surname> <given-names>A</given-names></name> <name><surname>Kumar</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Healthcare data breaches: insights and implications</article-title>. <source>Healthcare (Basel)</source>. (<year>2020</year>) <volume>8</volume>(<issue>2</issue>):<fpage>133</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare8020133</pub-id><pub-id pub-id-type="pmid">32414183</pub-id></mixed-citation></ref>
<ref id="B37"><label>37.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Edemekong</surname> <given-names>PF</given-names></name> <name><surname>Annamaraju</surname> <given-names>P</given-names></name> <name><surname>Afzal</surname> <given-names>M</given-names></name> <name><surname>Haydel</surname> <given-names>MJ</given-names></name></person-group>. <source>Health Insurance Portability and Accountability Act (HIPAA) Compliance</source>. <publisher-loc>Treasure Island (FL)</publisher-loc>: <publisher-name>StatPearls Publishing LLC</publisher-name> (<year>2025</year>).</mixed-citation></ref>
<ref id="B38"><label>38.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aboy</surname> <given-names>M</given-names></name> <name><surname>Crespo</surname> <given-names>C</given-names></name> <name><surname>Stern</surname> <given-names>A</given-names></name></person-group>. <article-title>Beyond the 510(k): the regulation of novel moderate-risk medical devices, intellectual property considerations, and innovation incentives in the FDA&#x2019;s <italic>de novo</italic> pathway</article-title>. <source>NPJ Digit Med</source>. (<year>2024</year>) <volume>7</volume>(<issue>1</issue>):<fpage>29</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-024-01021-y</pub-id><pub-id pub-id-type="pmid">38332182</pub-id></mixed-citation></ref>
<ref id="B39"><label>39.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nair</surname> <given-names>M</given-names></name> <name><surname>Svedberg</surname> <given-names>P</given-names></name> <name><surname>Larsson</surname> <given-names>I</given-names></name> <name><surname>Nygren</surname> <given-names>JM</given-names></name></person-group>. <article-title>A comprehensive overview of barriers and strategies for AI implementation in healthcare: mixed-method design</article-title>. <source>PLoS One</source>. (<year>2024</year>) <volume>19</volume>(<issue>8</issue>):<fpage>e0305949</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0305949</pub-id><pub-id pub-id-type="pmid">39121051</pub-id></mixed-citation></ref>
<ref id="B40"><label>40.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ferrara</surname> <given-names>E</given-names></name></person-group>. <article-title>Fairness and bias in artificial intelligence: a brief survey of sources, impacts, and mitigation strategies</article-title><source>. Science</source>. (<year>2023</year>) <volume>6</volume>(<issue>1</issue>):<fpage>3</fpage>. <pub-id pub-id-type="doi">10.3390/sci6010003</pub-id></mixed-citation></ref>
<ref id="B41"><label>41.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>S</given-names></name> <name><surname>Spiteri</surname> <given-names>AG</given-names></name> <name><surname>Huynh</surname> <given-names>ALH</given-names></name> <name><surname>Chu</surname> <given-names>C</given-names></name> <name><surname>Masters</surname> <given-names>CL</given-names></name><etal/></person-group> <article-title>Understanding machine learning applications in dementia research and clinical practice: a review for biomedical scientists and clinicians</article-title>. <source>Alzheimers Res Ther</source>. (<year>2024</year>) <volume>16</volume>(<issue>1</issue>):<fpage>175</fpage>. <pub-id pub-id-type="doi">10.1186/s13195-024-01540-6</pub-id><pub-id pub-id-type="pmid">39085973</pub-id></mixed-citation></ref>
<ref id="B42"><label>42.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sarkies</surname> <given-names>MN</given-names></name> <name><surname>Bowles</surname> <given-names>KA</given-names></name> <name><surname>Skinner</surname> <given-names>EH</given-names></name> <name><surname>Mitchell</surname> <given-names>D</given-names></name> <name><surname>Haas</surname> <given-names>R</given-names></name> <name><surname>Ho</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Data collection methods in health services research: hospital length of stay and discharge destination</article-title>. <source>Appl Clin Inform</source>. (<year>2015</year>) <volume>6</volume>(<issue>1</issue>):<fpage>96</fpage>&#x2013;<lpage>109</lpage>. <pub-id pub-id-type="doi">10.4338/aci-2014-10-ra-0097</pub-id><pub-id pub-id-type="pmid">25848416</pub-id></mixed-citation></ref>
<ref id="B43"><label>43.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Rahman</surname> <given-names>S</given-names></name> <name><surname>Jiang</surname> <given-names>LY</given-names></name> <name><surname>Gabriel</surname> <given-names>S</given-names></name> <name><surname>Aphinyanaphongs</surname> <given-names>Y</given-names></name> <name><surname>Oermann</surname> <given-names>EK</given-names></name> <name><surname>Chunara</surname> <given-names>R</given-names></name></person-group>. <comment>Generalization in healthcare AI: evaluation of a clinical large language model. <italic>arXiv Preprint; arXiv</italic>:2402.10965</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2402.10965</pub-id>. <comment>(Accessed Febrary 14, 2024)</comment></mixed-citation></ref>
<ref id="B44"><label>44.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hanna</surname> <given-names>MG</given-names></name> <name><surname>Pantanowitz</surname> <given-names>L</given-names></name> <name><surname>Jackson</surname> <given-names>B</given-names></name> <name><surname>Palmer</surname> <given-names>O</given-names></name> <name><surname>Visweswaran</surname> <given-names>S</given-names></name> <name><surname>Pantanowitz</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Ethical and bias considerations in artificial intelligence/machine learning</article-title>. <source>Mod Pathol</source>. (<year>2025</year>) <volume>38</volume>(<issue>3</issue>):<fpage>100686</fpage>. <pub-id pub-id-type="doi">10.1016/j.modpat.2024.100686</pub-id><pub-id pub-id-type="pmid">39694331</pub-id></mixed-citation></ref>
<ref id="B45"><label>45.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Williamson</surname> <given-names>SM</given-names></name> <name><surname>Prybutok</surname> <given-names>V</given-names></name></person-group>. <article-title>Balancing privacy and progress: a review of privacy challenges. Systemic oversight, and patient perceptions in AI-driven healthcare</article-title><source>. Appl Sci</source>. (<year>2024</year>) <volume>14</volume>(<issue>2</issue>):<fpage>675</fpage>. <pub-id pub-id-type="doi">10.3390/app14020675</pub-id></mixed-citation></ref>
<ref id="B46"><label>46.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Aburass</surname> <given-names>S</given-names></name></person-group>. <comment>Quantifying overfitting: introducing the overfitting index. <italic>arXiv Preprint; arXiv</italic>:2308.08682.</comment> <pub-id pub-id-type="doi">10.48550/arXiv.2308.08682</pub-id>. <comment>(Accessed August 16, 2023)</comment></mixed-citation></ref>
<ref id="B47"><label>47.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Goetz</surname> <given-names>L</given-names></name> <name><surname>Seedat</surname> <given-names>N</given-names></name> <name><surname>Vandersluis</surname> <given-names>R</given-names></name> <name><surname>van der Schaar</surname> <given-names>M</given-names></name></person-group>. <article-title>Generalization-a key challenge for responsible AI in patient-facing clinical applications</article-title>. <source>NPJ Digit Med</source>. (<year>2024</year>) <volume>7</volume>(<issue>1</issue>):<fpage>126</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-024-01127-3</pub-id><pub-id pub-id-type="pmid">38773304</pub-id></mixed-citation></ref>
<ref id="B48"><label>48.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nazer</surname> <given-names>LH</given-names></name> <name><surname>Zatarah</surname> <given-names>R</given-names></name> <name><surname>Waldrip</surname> <given-names>S</given-names></name> <name><surname>Ke</surname> <given-names>JXC</given-names></name> <name><surname>Moukheiber</surname> <given-names>M</given-names></name> <name><surname>Khanna</surname> <given-names>AK</given-names></name><etal/></person-group> <article-title>Bias in artificial intelligence algorithms and recommendations for mitigation</article-title>. <source>PLoS Digit Health</source>. (<year>2023</year>) <volume>2</volume>(<issue>6</issue>):<fpage>e0000278</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000278</pub-id><pub-id pub-id-type="pmid">37347721</pub-id></mixed-citation></ref>
<ref id="B49"><label>49.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Y</given-names></name> <name><surname>Clayton</surname> <given-names>EW</given-names></name> <name><surname>Novak</surname> <given-names>LL</given-names></name> <name><surname>Anders</surname> <given-names>S</given-names></name> <name><surname>Malin</surname> <given-names>B</given-names></name></person-group>. <article-title>Human-centered design to address biases in artificial intelligence</article-title>. <source>J Med Internet Res</source>. (<year>2023</year>) <volume>25</volume>:<fpage>e43251</fpage>. <pub-id pub-id-type="doi">10.2196/43251</pub-id><pub-id pub-id-type="pmid">36961506</pub-id></mixed-citation></ref>
<ref id="B50"><label>50.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nadarzynski</surname> <given-names>T</given-names></name> <name><surname>Knights</surname> <given-names>N</given-names></name> <name><surname>Husbands</surname> <given-names>D</given-names></name> <name><surname>Graham</surname> <given-names>CA</given-names></name> <name><surname>Llewellyn</surname> <given-names>CD</given-names></name> <name><surname>Buchanan</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Achieving health equity through conversational AI: a roadmap for design and implementation of inclusive chatbots in healthcare</article-title>. <source>PLoS Digit Health</source>. (<year>2024</year>) <volume>3</volume>(<issue>5</issue>):<fpage>e0000492</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000492</pub-id><pub-id pub-id-type="pmid">38696359</pub-id></mixed-citation></ref>
<ref id="B51"><label>51.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bajwa</surname> <given-names>J</given-names></name> <name><surname>Munir</surname> <given-names>U</given-names></name> <name><surname>Nori</surname> <given-names>A</given-names></name> <name><surname>Williams</surname> <given-names>B</given-names></name></person-group>. <article-title>Artificial intelligence in healthcare: transforming the practice of medicine</article-title>. <source>Future Healthc J</source>. (<year>2021</year>) <volume>8</volume>(<issue>2</issue>):<fpage>e188</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.7861/fhj.2021-0095</pub-id><pub-id pub-id-type="pmid">34286183</pub-id></mixed-citation></ref>
<ref id="B52"><label>52.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pi&#x00F1;a</surname> <given-names>IL</given-names></name> <name><surname>Cohen</surname> <given-names>PD</given-names></name> <name><surname>Larson</surname> <given-names>DB</given-names></name> <name><surname>Marion</surname> <given-names>LN</given-names></name> <name><surname>Sills</surname> <given-names>MR</given-names></name> <name><surname>Solberg</surname> <given-names>LI</given-names></name><etal/></person-group> <article-title>A framework for describing health care delivery organizations and systems</article-title>. <source>Am J Public Health</source>. (<year>2015</year>) <volume>105</volume>(<issue>4</issue>):<fpage>670</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.2105/ajph.2014.301926</pub-id></mixed-citation></ref>
<ref id="B53"><label>53.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>S</given-names></name> <name><surname>Johannesson</surname> <given-names>P</given-names></name></person-group>. <article-title>An ontology to bridge the clinical management of patients and public health responses for strengthening infectious disease surveillance: design science study</article-title>. <source>JMIR Form Res</source>. (<year>2024</year>) <volume>8</volume>:<fpage>e53711</fpage>. <pub-id pub-id-type="doi">10.2196/53711</pub-id><pub-id pub-id-type="pmid">39325530</pub-id></mixed-citation></ref>
<ref id="B54"><label>54.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Esmaeilzadeh</surname> <given-names>P</given-names></name></person-group>. <article-title>Evolution of health information sharing between health care organizations: potential of nonfungible tokens</article-title>. <source>Interact J Med Res</source>. (<year>2023</year>) <volume>12</volume>:<fpage>e42685</fpage>. <pub-id pub-id-type="doi">10.2196/42685</pub-id><pub-id pub-id-type="pmid">37043269</pub-id></mixed-citation></ref>
<ref id="B55"><label>55.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>YH</given-names></name> <name><surname>Li</surname> <given-names>YL</given-names></name> <name><surname>Wei</surname> <given-names>MY</given-names></name> <name><surname>Li</surname> <given-names>GY</given-names></name></person-group>. <article-title>Innovation and challenges of artificial intelligence technology in personalized healthcare</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>18994</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-70073-7</pub-id><pub-id pub-id-type="pmid">39152194</pub-id></mixed-citation></ref>
<ref id="B56"><label>56.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Junaid</surname> <given-names>SB</given-names></name> <name><surname>Imam</surname> <given-names>AA</given-names></name> <name><surname>Balogun</surname> <given-names>AO</given-names></name> <name><surname>De Silva</surname> <given-names>LC</given-names></name> <name><surname>Surakat</surname> <given-names>YA</given-names></name> <name><surname>Kumar</surname> <given-names>G</given-names></name><etal/></person-group> <article-title>Recent advancements in emerging technologies for healthcare management systems: a survey</article-title>. <source>Healthcare (Basel)</source>. (<year>2022</year>) <volume>10</volume>(<issue>10</issue>):<fpage>1940</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare10101940</pub-id><pub-id pub-id-type="pmid">36292387</pub-id></mixed-citation></ref>
<ref id="B57"><label>57.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nan</surname> <given-names>J</given-names></name> <name><surname>Xu</surname> <given-names>LQ</given-names></name></person-group>. <article-title>Designing interoperable health care services based on fast healthcare interoperability resources: literature review</article-title>. <source>JMIR Med Inform</source>. (<year>2023</year>) <volume>11</volume>:<fpage>e44842</fpage>. <pub-id pub-id-type="doi">10.2196/44842</pub-id><pub-id pub-id-type="pmid">37603388</pub-id></mixed-citation></ref>
<ref id="B58"><label>58.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Holmgren</surname> <given-names>AJ</given-names></name> <name><surname>Esdar</surname> <given-names>M</given-names></name> <name><surname>H&#x00FC;sers</surname> <given-names>J</given-names></name> <name><surname>Coutinho-Almeida</surname> <given-names>J</given-names></name></person-group>. <article-title>Health information exchange: understanding the policy landscape and future of data interoperability</article-title>. <source>Yearb Med Inform</source>. (<year>2023</year>) <volume>32</volume>(<issue>1</issue>):<fpage>184</fpage>&#x2013;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1055/s-0043-1768719</pub-id><pub-id pub-id-type="pmid">37414031</pub-id></mixed-citation></ref>
<ref id="B59"><label>59.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vieira</surname> <given-names>R</given-names></name> <name><surname>Silva</surname> <given-names>D</given-names></name> <name><surname>Ribeiro</surname> <given-names>E</given-names></name> <name><surname>Perdigoto</surname> <given-names>L</given-names></name> <name><surname>Coelho</surname> <given-names>PJ</given-names></name></person-group>. <article-title>Performance evaluation of computer vision algorithms in a programmable logic controller: an industrial case study</article-title>. <source>Sensors (Basel)</source>. (<year>2024</year>) <volume>24</volume>(<issue>3</issue>):<fpage>843</fpage>. <pub-id pub-id-type="doi">10.3390/s24030843</pub-id><pub-id pub-id-type="pmid">38339560</pub-id></mixed-citation></ref>
<ref id="B60"><label>60.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>D</given-names></name> <name><surname>Jo</surname> <given-names>Y</given-names></name> <name><surname>Lee</surname> <given-names>M</given-names></name> <name><surname>Kim</surname> <given-names>T</given-names></name></person-group>. <comment>Retaining and enhancing pre-trained knowledge in vision-language models with prompt ensembling. <italic>arXiv Preprint; arXiv</italic>:2412.07077</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2412.07077</pub-id>. <comment>(Accessed December 10, 2024)</comment></mixed-citation></ref>
<ref id="B61"><label>61.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hassija</surname> <given-names>V</given-names></name> <name><surname>Chamola</surname> <given-names>V</given-names></name> <name><surname>Mahapatra</surname> <given-names>A</given-names></name> <name><surname>Singal</surname> <given-names>A</given-names></name> <name><surname>Goel</surname> <given-names>D</given-names></name> <name><surname>Huang</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>Interpreting black-box models: a review on explainable artificial intelligence</article-title>. <source>Cognit Comput</source>. (<year>2023</year>) <volume>16</volume>(<issue>1</issue>):<fpage>45</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1007/s12559-023-10179-8</pub-id></mixed-citation></ref>
<ref id="B62"><label>62.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Marey</surname> <given-names>A</given-names></name> <name><surname>Arjmand</surname> <given-names>P</given-names></name> <name><surname>Alerab</surname> <given-names>ADS</given-names></name> <name><surname>Eslami</surname> <given-names>MJ</given-names></name> <name><surname>Saad</surname> <given-names>AM</given-names></name> <name><surname>Sanchez</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>Explainability, transparency and black box challenges of AI in radiology: impact on patient care in cardiovascular radiology</article-title>. <source>Egypt J Radiol Nucl Med</source>. (<year>2024</year>) <volume>55</volume>:<fpage>183</fpage>. <pub-id pub-id-type="doi">10.1186/s43055-024-01356-2</pub-id></mixed-citation></ref>
<ref id="B63"><label>63.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Giannaros</surname> <given-names>A</given-names></name> <name><surname>Karras</surname> <given-names>A</given-names></name> <name><surname>Theodorakopoulos</surname> <given-names>L</given-names></name> <name><surname>Karras</surname> <given-names>C</given-names></name> <name><surname>Kranias</surname> <given-names>P</given-names></name> <name><surname>Schizas</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>Autonomous vehicles: sophisticated attacks, safety issues, challenges, open topics. Blockchain, and future directions</article-title>. <source>J Cybersecur Priv</source>. (<year>2023</year>) <volume>3</volume>(<issue>3</issue>):<fpage>493</fpage>&#x2013;<lpage>543</lpage>. <pub-id pub-id-type="doi">10.3390/jcp3030025</pub-id></mixed-citation></ref>
<ref id="B64"><label>64.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ali</surname> <given-names>S</given-names></name> <name><surname>Abuhmed</surname> <given-names>T</given-names></name> <name><surname>El-Sappagh</surname> <given-names>S</given-names></name> <name><surname>Muhammad</surname> <given-names>K</given-names></name> <name><surname>Alonso-Moral</surname> <given-names>JM</given-names></name> <name><surname>Confalonieri</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Explainable artificial intelligence (XAI): what we know and what is left to attain trustworthy artificial intelligence</article-title>. <source>Inf Fusion</source>. (<year>2023</year>) <volume>99</volume>:<fpage>101805</fpage>. <pub-id pub-id-type="doi">10.1016/j.inffus.2023.101805</pub-id></mixed-citation></ref>
<ref id="B65"><label>65.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cheong</surname> <given-names>BC</given-names></name></person-group>. <article-title>Transparency and accountability in AI systems: safeguarding wellbeing in the age of algorithmic decision-making</article-title>. <source>Front Hum Dyn</source>. (<year>2024</year>) <volume>6</volume>:<fpage>1421273</fpage>. <pub-id pub-id-type="doi">10.3389/fhumd.2024.1421273</pub-id></mixed-citation></ref>
<ref id="B66"><label>66.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ahmed</surname> <given-names>MI</given-names></name> <name><surname>Spooner</surname> <given-names>B</given-names></name> <name><surname>Isherwood</surname> <given-names>J</given-names></name> <name><surname>Lane</surname> <given-names>M</given-names></name> <name><surname>Orrock</surname> <given-names>E</given-names></name> <name><surname>Dennison</surname> <given-names>A</given-names></name></person-group>. <article-title>A systematic review of the barriers to the implementation of artificial intelligence in healthcare</article-title>. <source>Cureus</source>. (<year>2023</year>) <volume>15</volume>(<issue>10</issue>):<fpage>e46454</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.46454</pub-id><pub-id pub-id-type="pmid">37927664</pub-id></mixed-citation></ref>
<ref id="B67"><label>67.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Farhud</surname> <given-names>DD</given-names></name> <name><surname>Zokaei</surname> <given-names>S</given-names></name></person-group>. <article-title>Ethical issues of artificial intelligence in medicine and healthcare</article-title>. <source>Iran J Public Health</source>. (<year>2021</year>) <volume>50</volume>(<issue>11</issue>):<fpage>i</fpage>&#x2013;<lpage>V</lpage>. <pub-id pub-id-type="doi">10.18502/ijph.v50i11.7600</pub-id></mixed-citation></ref>
<ref id="B68"><label>68.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cross</surname> <given-names>JL</given-names></name> <name><surname>Choma</surname> <given-names>MA</given-names></name> <name><surname>Onofrey</surname> <given-names>JA</given-names></name></person-group>. <article-title>Bias in medical AI: implications for clinical decision-making</article-title>. <source>PLoS Digit Health</source>. (<year>2024</year>) <volume>3</volume>(<issue>11</issue>):<fpage>e0000651</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pdig.0000651</pub-id><pub-id pub-id-type="pmid">39509461</pub-id></mixed-citation></ref>
<ref id="B69"><label>69.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Elendu</surname> <given-names>C</given-names></name> <name><surname>Amaechi</surname> <given-names>DC</given-names></name> <name><surname>Elendu</surname> <given-names>TC</given-names></name> <name><surname>Jingwa</surname> <given-names>KA</given-names></name> <name><surname>Okoye</surname> <given-names>OK</given-names></name> <name><surname>John Okah</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Ethical implications of AI and robotics in healthcare: a review</article-title>. <source>Medicine (Baltimore)</source>. (<year>2023</year>) <volume>102</volume>(<issue>50</issue>):<fpage>e36671</fpage>. <pub-id pub-id-type="doi">10.1097/md.0000000000036671</pub-id><pub-id pub-id-type="pmid">38115340</pub-id></mixed-citation></ref>
<ref id="B70"><label>70.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khera</surname> <given-names>R</given-names></name> <name><surname>Simon</surname> <given-names>MA</given-names></name> <name><surname>Ross</surname> <given-names>JS</given-names></name></person-group>. <article-title>Automation bias and assistive AI: risk of harm from AI-driven clinical decision support</article-title>. <source>JAMA</source>. (<year>2023</year>) <volume>330</volume>(<issue>23</issue>):<fpage>2255</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1001/jama.2023.22557</pub-id><pub-id pub-id-type="pmid">38112824</pub-id></mixed-citation></ref>
<ref id="B71"><label>71.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khan</surname> <given-names>B</given-names></name> <name><surname>Fatima</surname> <given-names>H</given-names></name> <name><surname>Qureshi</surname> <given-names>A</given-names></name> <name><surname>Kumar</surname> <given-names>S</given-names></name> <name><surname>Hanan</surname> <given-names>A</given-names></name> <name><surname>Hussain</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Drawbacks of artificial intelligence and their potential solutions in the healthcare sector</article-title>. <source>Biomed Mater Devices</source>. (<year>2023</year>) <volume>1</volume>:<fpage>731</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1007/s44174-023-00063-2</pub-id></mixed-citation></ref>
<ref id="B72"><label>72.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Grote</surname> <given-names>T</given-names></name> <name><surname>Berens</surname> <given-names>P</given-names></name></person-group>. <article-title>On the ethics of algorithmic decision-making in healthcare</article-title>. <source>J Med Ethics</source>. (<year>2020</year>) <volume>46</volume>(<issue>3</issue>):<fpage>205</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1136/medethics-2019-105586</pub-id><pub-id pub-id-type="pmid">31748206</pub-id></mixed-citation></ref>
<ref id="B73"><label>73.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tilala</surname> <given-names>MH</given-names></name> <name><surname>Chenchala</surname> <given-names>PK</given-names></name> <name><surname>Choppadandi</surname> <given-names>A</given-names></name> <name><surname>Kaur</surname> <given-names>J</given-names></name> <name><surname>Naguri</surname> <given-names>S</given-names></name> <name><surname>Saoji</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Ethical considerations in the use of artificial intelligence and machine learning in health care: a comprehensive review</article-title>. <source>Cureus</source>. (<year>2024</year>) <volume>16</volume>(<issue>6</issue>):<fpage>e62443</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.62443</pub-id><pub-id pub-id-type="pmid">39011215</pub-id></mixed-citation></ref>
<ref id="B74"><label>74.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khosravi</surname> <given-names>M</given-names></name> <name><surname>Zare</surname> <given-names>Z</given-names></name> <name><surname>Mojtabaeian</surname> <given-names>SM</given-names></name> <name><surname>Izadi</surname> <given-names>R</given-names></name></person-group>. <article-title>Artificial intelligence and decision-making in healthcare: a thematic analysis of a systematic review of reviews</article-title>. <source>Health Serv Res Manag Epidemiol</source>. (<year>2024</year>) <volume>11</volume>:<fpage>23333928241234863</fpage>. <pub-id pub-id-type="doi">10.1177/23333928241234863</pub-id><pub-id pub-id-type="pmid">38449840</pub-id></mixed-citation></ref>
<ref id="B75"><label>75.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sauerbrei</surname> <given-names>A</given-names></name> <name><surname>Kerasidou</surname> <given-names>A</given-names></name> <name><surname>Lucivero</surname> <given-names>F</given-names></name> <name><surname>Hallowell</surname> <given-names>N</given-names></name></person-group>. <article-title>The impact of artificial intelligence on the person-centred, doctor-patient relationship: some problems and solutions</article-title>. <source>BMC Med Inform Decis Mak</source>. (<year>2023</year>) <volume>23</volume>(<issue>1</issue>):<fpage>73</fpage>. <pub-id pub-id-type="doi">10.1186/s12911-023-02162-y</pub-id><pub-id pub-id-type="pmid">37081503</pub-id></mixed-citation></ref>
<ref id="B76"><label>76.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>HJ</given-names></name></person-group>. <article-title>Patient perspectives on informed consent for medical AI: a web-based experiment</article-title>. <source>Digit Health</source>. (<year>2024</year>) <volume>10</volume>:<fpage>20552076241247938</fpage>. <pub-id pub-id-type="doi">10.1177/20552076241247938</pub-id><pub-id pub-id-type="pmid">38698829</pub-id></mixed-citation></ref>
<ref id="B77"><label>77.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mennella</surname> <given-names>C</given-names></name> <name><surname>Maniscalco</surname> <given-names>U</given-names></name> <name><surname>De Pietro</surname> <given-names>G</given-names></name> <name><surname>Esposito</surname> <given-names>M</given-names></name></person-group>. <article-title>Ethical and regulatory challenges of AI technologies in healthcare: a narrative review</article-title>. <source>Heliyon</source>. (<year>2024</year>) <volume>10</volume>(<issue>4</issue>):<fpage>e26297</fpage>. <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e26297</pub-id><pub-id pub-id-type="pmid">38384518</pub-id></mixed-citation></ref>
<ref id="B78"><label>78.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>SH</given-names></name> <name><surname>Choi</surname> <given-names>J</given-names></name> <name><surname>Byeon</surname> <given-names>JS</given-names></name></person-group>. <article-title>Key principles of clinical validation, device approval, and insurance coverage decisions of artificial intelligence</article-title>. <source>Korean J Radiol</source>. (<year>2021</year>) <volume>22</volume>(<issue>3</issue>):<fpage>442</fpage>&#x2013;<lpage>53</lpage>. <pub-id pub-id-type="doi">10.3348/kjr.2021.0048</pub-id><pub-id pub-id-type="pmid">33629545</pub-id></mixed-citation></ref>
<ref id="B79"><label>79.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rockenschaub</surname> <given-names>P</given-names></name> <name><surname>Akay</surname> <given-names>EM</given-names></name> <name><surname>Carlisle</surname> <given-names>BG</given-names></name> <name><surname>Hilbert</surname> <given-names>A</given-names></name> <name><surname>Wendland</surname> <given-names>J</given-names></name> <name><surname>Meyer-Eschenbach</surname> <given-names>F</given-names></name><etal/></person-group> <article-title>External validation of AI-based scoring systems in the ICU: a systematic review and meta-analysis</article-title>. <source>BMC Med Inform Decis Mak</source>. (<year>2025</year>) <volume>25</volume>(<issue>1</issue>):<fpage>5</fpage>. <pub-id pub-id-type="doi">10.1186/s12911-024-02830-7</pub-id><pub-id pub-id-type="pmid">39762808</pub-id></mixed-citation></ref>
<ref id="B80"><label>80.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alsadoun</surname> <given-names>L</given-names></name> <name><surname>Ali</surname> <given-names>H</given-names></name> <name><surname>Mushtaq</surname> <given-names>MM</given-names></name> <name><surname>Mushtaq</surname> <given-names>M</given-names></name> <name><surname>Burhanuddin</surname> <given-names>M</given-names></name> <name><surname>Anwar</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Artificial intelligence (AI)-enhanced detection of diabetic retinopathy from fundus images: the current landscape and future directions</article-title>. <source>Cureus</source>. (<year>2024</year>) <volume>16</volume>(<issue>8</issue>):<fpage>e67844</fpage>. <pub-id pub-id-type="doi">10.7759/cureus.67844</pub-id><pub-id pub-id-type="pmid">39323686</pub-id></mixed-citation></ref>
<ref id="B81"><label>81.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gala</surname> <given-names>D</given-names></name> <name><surname>Behl</surname> <given-names>H</given-names></name> <name><surname>Shah</surname> <given-names>M</given-names></name> <name><surname>Makaryus</surname> <given-names>AN</given-names></name></person-group>. <article-title>The role of artificial intelligence in improving patient outcomes and future of healthcare delivery in cardiology: a narrative review of the literature</article-title>. <source>Healthcare (Basel)</source>. (<year>2024</year>) <volume>12</volume>(<issue>4</issue>):<fpage>481</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare12040481</pub-id><pub-id pub-id-type="pmid">38391856</pub-id></mixed-citation></ref>
<ref id="B82"><label>82.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Esteva</surname> <given-names>A</given-names></name> <name><surname>Chou</surname> <given-names>K</given-names></name> <name><surname>Yeung</surname> <given-names>S</given-names></name> <name><surname>Naik</surname> <given-names>N</given-names></name> <name><surname>Madani</surname> <given-names>A</given-names></name> <name><surname>Mottaghi</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Deep learning-enabled medical computer vision</article-title>. <source>NPJ Digit Med</source>. (<year>2021</year>) <volume>4</volume>(<issue>1</issue>):<fpage>5</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-020-00376-2</pub-id><pub-id pub-id-type="pmid">33420381</pub-id></mixed-citation></ref>
<ref id="B83"><label>83.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Palaniappan</surname> <given-names>K</given-names></name> <name><surname>Lin</surname> <given-names>EYT</given-names></name> <name><surname>Vogel</surname> <given-names>S</given-names></name></person-group>. <article-title>Global regulatory frameworks for the use of artificial intelligence (AI) in the healthcare services sector</article-title>. <source>Healthcare (Basel)</source>. (<year>2024</year>) <volume>12</volume>(<issue>5</issue>):<fpage>562</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare12050562</pub-id><pub-id pub-id-type="pmid">38470673</pub-id></mixed-citation></ref>
<ref id="B84"><label>84.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tajbakhsh</surname> <given-names>N</given-names></name> <name><surname>Roth</surname> <given-names>H</given-names></name> <name><surname>Terzopoulos</surname> <given-names>D</given-names></name> <name><surname>Liang</surname> <given-names>J</given-names></name></person-group>. <article-title>Guest editorial annotation-efficient deep learning: the holy grail of medical imaging</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2021</year>) <volume>40</volume>(<issue>10</issue>):<fpage>2526</fpage>&#x2013;<lpage>33</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2021.3089292</pub-id><pub-id pub-id-type="pmid">34795461</pub-id></mixed-citation></ref>
<ref id="B85"><label>85.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Elendu</surname> <given-names>C</given-names></name> <name><surname>Amaechi</surname> <given-names>DC</given-names></name> <name><surname>Okatta</surname> <given-names>AU</given-names></name> <name><surname>Amaechi</surname> <given-names>EC</given-names></name> <name><surname>Elendu</surname> <given-names>TC</given-names></name> <name><surname>Ezeh</surname> <given-names>CP</given-names></name><etal/></person-group> <article-title>The impact of simulation-based training in medical education: a review</article-title>. <source>Medicine (Baltimore)</source>. (<year>2024</year>) <volume>103</volume>(<issue>27</issue>):<fpage>e38813</fpage>. <pub-id pub-id-type="doi">10.1097/md.0000000000038813</pub-id><pub-id pub-id-type="pmid">38968472</pub-id></mixed-citation></ref>
<ref id="B86"><label>86.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>S</given-names></name> <name><surname>Li</surname> <given-names>C</given-names></name> <name><surname>Wang</surname> <given-names>R</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Wang</surname> <given-names>M</given-names></name> <name><surname>Tan</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Annotation-efficient deep learning for automatic medical image segmentation</article-title>. <source>Nat Commun</source>. (<year>2021</year>) <volume>12</volume>(<issue>1</issue>):<fpage>5915</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-021-26216-9</pub-id><pub-id pub-id-type="pmid">34625565</pub-id></mixed-citation></ref>
<ref id="B87"><label>87.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>M&#x00FC;ller</surname> <given-names>D</given-names></name> <name><surname>Kramer</surname> <given-names>F</given-names></name></person-group>. <article-title>MIScnn: a framework for medical image segmentation with convolutional neural networks and deep learning</article-title>. <source>BMC Med Imaging</source>. (<year>2021</year>) <volume>21</volume>(<issue>1</issue>):<fpage>12</fpage>. <pub-id pub-id-type="doi">10.1186/s12880-020-00543-7</pub-id></mixed-citation></ref>
<ref id="B88"><label>88.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Miao</surname> <given-names>R</given-names></name> <name><surname>Toth</surname> <given-names>R</given-names></name> <name><surname>Zhou</surname> <given-names>Y</given-names></name> <name><surname>Madabhushi</surname> <given-names>A</given-names></name> <name><surname>Janowczyk</surname> <given-names>A</given-names></name></person-group>. <article-title>Quick annotator: an open-source digital pathology based rapid image annotation tool</article-title>. <source>J Pathol Clin Res</source>. (<year>2021</year>) <volume>7</volume>(<issue>6</issue>):<fpage>542</fpage>&#x2013;<lpage>7</lpage>. <pub-id pub-id-type="doi">10.1002/cjp2.229</pub-id><pub-id pub-id-type="pmid">34288586</pub-id></mixed-citation></ref>
<ref id="B89"><label>89.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Davenport</surname> <given-names>T</given-names></name> <name><surname>Kalakota</surname> <given-names>R</given-names></name></person-group>. <article-title>The potential for artificial intelligence in healthcare</article-title>. <source>Future Healthc J</source>. (<year>2019</year>) <volume>6</volume>(<issue>2</issue>):<fpage>94</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.7861/futurehosp.6-2-94</pub-id><pub-id pub-id-type="pmid">31363513</pub-id></mixed-citation></ref>
<ref id="B90"><label>90.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chan</surname> <given-names>HP</given-names></name> <name><surname>Samala</surname> <given-names>RK</given-names></name> <name><surname>Hadjiiski</surname> <given-names>LM</given-names></name> <name><surname>Zhou</surname> <given-names>C</given-names></name></person-group>. <article-title>Deep learning in medical image analysis</article-title>. <source>Adv Exp Med Biol</source>. (<year>2020</year>) <volume>1213</volume>:<fpage>3</fpage>&#x2013;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-33128-3_1</pub-id><pub-id pub-id-type="pmid">32030660</pub-id></mixed-citation></ref>
<ref id="B91"><label>91.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Z</given-names></name> <name><surname>Yang</surname> <given-names>E</given-names></name> <name><surname>Shen</surname> <given-names>L</given-names></name> <name><surname>Huang</surname> <given-names>H</given-names></name></person-group>. <article-title>A comprehensive survey of forgetting in deep learning beyond continual learning</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source>. (<year>2025</year>) <volume>47</volume>:<fpage>1464</fpage>&#x2013;<lpage>83</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2024.3498346</pub-id></mixed-citation></ref>
<ref id="B92"><label>92.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vela</surname> <given-names>D</given-names></name> <name><surname>Sharp</surname> <given-names>A</given-names></name> <name><surname>Zhang</surname> <given-names>R</given-names></name> <name><surname>Nguyen</surname> <given-names>T</given-names></name> <name><surname>Hoang</surname> <given-names>A</given-names></name> <name><surname>Pianykh</surname> <given-names>OS</given-names></name></person-group>. <article-title>Temporal quality degradation in AI models</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>(<issue>1</issue>):<fpage>11654</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-15245-z</pub-id><pub-id pub-id-type="pmid">35803963</pub-id></mixed-citation></ref>
<ref id="B93"><label>93.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Krishnan</surname> <given-names>G</given-names></name> <name><surname>Singh</surname> <given-names>S</given-names></name> <name><surname>Pathania</surname> <given-names>M</given-names></name> <name><surname>Gosavi</surname> <given-names>S</given-names></name> <name><surname>Abhishek</surname> <given-names>S</given-names></name> <name><surname>Parchani</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Artificial intelligence in clinical medicine: catalyzing a sustainable global healthcare paradigm</article-title>. <source>Front Artif Intell</source>. (<year>2023</year>) <volume>6</volume>:<fpage>1227091</fpage>. <pub-id pub-id-type="doi">10.3389/frai.2023.1227091</pub-id><pub-id pub-id-type="pmid">37705603</pub-id></mixed-citation></ref>
<ref id="B94"><label>94.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kabir</surname> <given-names>MM</given-names></name> <name><surname>Rahman</surname> <given-names>A</given-names></name> <name><surname>Hasan</surname> <given-names>MN</given-names></name> <name><surname>Mridha</surname> <given-names>MF</given-names></name></person-group>. <article-title>Computer vision algorithms in healthcare: recent advancements and future challenges</article-title>. <source>Comput Biol Med</source>. (<year>2025</year>) <volume>185</volume>:<fpage>109531</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2024.109531</pub-id><pub-id pub-id-type="pmid">39675214</pub-id></mixed-citation></ref>
<ref id="B95"><label>95.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ray</surname> <given-names>PP</given-names></name> <name><surname>Majumder</surname> <given-names>P</given-names></name></person-group>. <article-title>The potential of ChatGPT to transform healthcare and address ethical challenges in artificial intelligence-driven medicine</article-title>. <source>J Clin Neurol</source>. (<year>2023</year>) <volume>19</volume>(<issue>5</issue>):<fpage>509</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.3988/jcn.2023.0158</pub-id><pub-id pub-id-type="pmid">37635433</pub-id></mixed-citation></ref>
<ref id="B96"><label>96.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Varnosfaderani</surname> <given-names>SM</given-names></name> <name><surname>Forouzanfar</surname> <given-names>M</given-names></name></person-group>. <article-title>The role of AI in hospitals and clinics: transforming healthcare in the 21st century</article-title>. <source>Bioengineering (Basel)</source>. (<year>2024</year>) <volume>11</volume>(<issue>4</issue>):<fpage>337</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering11040337</pub-id><pub-id pub-id-type="pmid">38671759</pub-id></mixed-citation></ref>
<ref id="B97"><label>97.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mir</surname> <given-names>MM</given-names></name> <name><surname>Mir</surname> <given-names>GM</given-names></name> <name><surname>Raina</surname> <given-names>NT</given-names></name> <name><surname>Mir</surname> <given-names>SM</given-names></name> <name><surname>Mir</surname> <given-names>SM</given-names></name> <name><surname>Miskeen</surname> <given-names>E</given-names></name><etal/></person-group> <article-title>Application of artificial intelligence in medical education: current scenario and future perspectives</article-title>. <source>J Adv Med Educ Prof</source>. (<year>2023</year>) <volume>11</volume>(<issue>3</issue>):<fpage>133</fpage>&#x2013;<lpage>40</lpage>. <pub-id pub-id-type="doi">10.30476/jamp.2023.98655.1803</pub-id><pub-id pub-id-type="pmid">37469385</pub-id></mixed-citation></ref>
<ref id="B98"><label>98.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>Z</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Xu</surname> <given-names>L</given-names></name></person-group>. <article-title>DRA-Net: medical image segmentation based on adaptive feature extraction and region-level information fusion</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>9714</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-60475-y</pub-id><pub-id pub-id-type="pmid">38678063</pub-id></mixed-citation></ref>
<ref id="B99"><label>99.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Bertels</surname> <given-names>J</given-names></name> <name><surname>Robben</surname> <given-names>D</given-names></name> <name><surname>Lemmens</surname> <given-names>R</given-names></name> <name><surname>Vandermeulen</surname> <given-names>D</given-names></name></person-group>. <comment>Convolutional neural networks for medical image segmentation. <italic>arXiv Preprint; arXiv</italic>:2211.09562</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2211.09562</pub-id>. <comment>(Accessed November 17, 2022)</comment></mixed-citation></ref>
<ref id="B100"><label>100.</label><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ronneberger</surname> <given-names>O</given-names></name> <name><surname>Fischer</surname> <given-names>P</given-names></name> <name><surname>Brox</surname> <given-names>T</given-names></name></person-group>. <source>U-Net: Convolutional Networks for Biomedical Image Segmentation</source>. Medical Image Computing and Computer-Assisted Intervention &#x2013; MICCAI. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2015</year>). pp. <fpage>234</fpage>&#x2013;<lpage>41</lpage>.</mixed-citation></ref>
<ref id="B101"><label>101.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Jiangtao</surname> <given-names>W</given-names></name> <name><surname>Ruhaiyem</surname> <given-names>NIR</given-names></name> <name><surname>Panpan</surname> <given-names>F</given-names></name></person-group>. <comment>A comprehensive review of U-Net and its variants: advances and applications in medical image segmentation. <italic>arXiv Preprint; arXiv</italic>:2502.06895</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2502.06895</pub-id>. <comment>(Accessed February 9, 2025)</comment></mixed-citation></ref>
<ref id="B102"><label>102.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>R</given-names></name> <name><surname>Liu</surname> <given-names>W</given-names></name> <name><surname>Yang</surname> <given-names>L</given-names></name> <name><surname>Sun</surname> <given-names>S</given-names></name> <name><surname>Hu</surname> <given-names>W</given-names></name> <name><surname>Zhang</surname> <given-names>F</given-names></name><etal/></person-group> <comment>Convolutional network for pixel-level sea-land segmentation. <italic>arXiv Preprint; arXiv</italic>:1709.00201</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.1709.00201</pub-id>. <comment>(Accessed September 1, 2017)</comment></mixed-citation></ref>
<ref id="B103"><label>103.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Azad</surname> <given-names>R</given-names></name> <name><surname>Aghdam</surname> <given-names>EK</given-names></name> <name><surname>Rauland</surname> <given-names>A</given-names></name> <name><surname>Jia</surname> <given-names>Y</given-names></name> <name><surname>Avval</surname> <given-names>AH</given-names></name> <name><surname>Bozorgpour</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Medical image segmentation review: the success of U-Net</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source>. (<year>2024</year>) <volume>46</volume>(<issue>12</issue>):<fpage>10076</fpage>&#x2013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2024.3435571</pub-id><pub-id pub-id-type="pmid">39167505</pub-id></mixed-citation></ref>
<ref id="B104"><label>104.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sahragard</surname> <given-names>E</given-names></name> <name><surname>Farsi</surname> <given-names>H</given-names></name> <name><surname>Mohamadzadeh</surname> <given-names>S</given-names></name></person-group>. <article-title>Advancing semantic segmentation: enhanced U-Net algorithm with attention mechanism and deformable convolution</article-title>. <source>PLoS One</source>. (<year>2025</year>) <volume>20</volume>(<issue>1</issue>):<fpage>e0305561</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0305561</pub-id><pub-id pub-id-type="pmid">39820812</pub-id></mixed-citation></ref>
<ref id="B105"><label>105.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Siddique</surname> <given-names>N</given-names></name> <name><surname>Paheding</surname> <given-names>S</given-names></name> <name><surname>Elkin</surname> <given-names>CP</given-names></name> <name><surname>Devabhaktuni</surname> <given-names>V</given-names></name></person-group>. <article-title>U-Net and its variants for medical image segmentation: a review of theory and applications</article-title>. <source>IEEE Access</source>. (<year>2021</year>) <volume>9</volume>:<fpage>82031</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1109/access.2021.3086020</pub-id></mixed-citation></ref>
<ref id="B106"><label>106.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>C</given-names></name> <name><surname>Deng</surname> <given-names>X</given-names></name> <name><surname>Ling</surname> <given-names>SH</given-names></name></person-group>. <article-title>Next-gen medical imaging: U-Net evolution and the rise of transformers</article-title>. <source>Sensors (Basel)</source>. (<year>2024</year>) <volume>24</volume>(<issue>14</issue>):<fpage>4668</fpage>. <pub-id pub-id-type="doi">10.3390/s24144668</pub-id><pub-id pub-id-type="pmid">39066065</pub-id></mixed-citation></ref>
<ref id="B107"><label>107.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ghorpade</surname> <given-names>H</given-names></name> <name><surname>Kolhar</surname> <given-names>S</given-names></name> <name><surname>Jagtap</surname> <given-names>J</given-names></name> <name><surname>Chakraborty</surname> <given-names>J</given-names></name></person-group>. <article-title>An optimized two stage U-Net approach for segmentation of pancreas and pancreatic tumor</article-title>. <source>MethodsX</source>. (<year>2024</year>) <volume>13</volume>:<fpage>102995</fpage>. <pub-id pub-id-type="doi">10.1016/j.mex.2024.102995</pub-id><pub-id pub-id-type="pmid">39435045</pub-id></mixed-citation></ref>
<ref id="B108"><label>108.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Srinivasan</surname> <given-names>S</given-names></name> <name><surname>Durairaju</surname> <given-names>K</given-names></name> <name><surname>Deeba</surname> <given-names>K</given-names></name> <name><surname>Mathivanan</surname> <given-names>SK</given-names></name> <name><surname>Karthikeyan</surname> <given-names>P</given-names></name> <name><surname>Shah</surname> <given-names>MA</given-names></name></person-group>. <article-title>Multimodal biomedical image segmentation using multi-dimensional U-convolutional neural network</article-title>. <source>BMC Med Imaging</source>. (<year>2024</year>) <volume>24</volume>(<issue>1</issue>):<fpage>38</fpage>. <pub-id pub-id-type="doi">10.1186/s12880-024-01197-5</pub-id><pub-id pub-id-type="pmid">38331800</pub-id></mixed-citation></ref>
<ref id="B109"><label>109.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Punn</surname> <given-names>NS</given-names></name> <name><surname>Agarwal</surname> <given-names>S</given-names></name></person-group>. <article-title>Modality specific U-net variants for biomedical image segmentation: a survey</article-title>. <source>Artif Intell Rev</source>. (<year>2022</year>) <volume>55</volume>(<issue>7</issue>):<fpage>5845</fpage>&#x2013;<lpage>89</lpage>. <pub-id pub-id-type="doi">10.1007/s10462-022-10152-1</pub-id><pub-id pub-id-type="pmid">35250146</pub-id></mixed-citation></ref>
<ref id="B110"><label>110.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dimitrovski</surname> <given-names>I</given-names></name> <name><surname>Spasev</surname> <given-names>V</given-names></name> <name><surname>Loshkovska</surname> <given-names>S</given-names></name> <name><surname>Kitanovski</surname> <given-names>I</given-names></name></person-group>. <article-title>U-Net ensemble for enhanced semantic segmentation in remote sensing imagery</article-title>. <source>Remote Sens (Basel)</source>. (<year>2024</year>) <volume>16</volume>(<issue>12</issue>):<fpage>2077</fpage>. <pub-id pub-id-type="doi">10.3390/rs16122077</pub-id></mixed-citation></ref>
<ref id="B111"><label>111.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>Y</given-names></name> <name><surname>Jiang</surname> <given-names>Y</given-names></name> <name><surname>Peng</surname> <given-names>Y</given-names></name> <name><surname>Yuan</surname> <given-names>F</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>J</given-names></name></person-group>. <article-title>Medical image segmentation: a comprehensive review of deep learning-based methods</article-title>. <source>Tomography</source>. (<year>2025</year>) <volume>11</volume>(<issue>5</issue>):<fpage>52</fpage>. <pub-id pub-id-type="doi">10.3390/tomography11050052</pub-id><pub-id pub-id-type="pmid">40423254</pub-id></mixed-citation></ref>
<ref id="B112"><label>112.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>R</given-names></name> <name><surname>Jiang</surname> <given-names>G</given-names></name></person-group>. <article-title>Exploring a multi-path U-Net with probability distribution attention and cascade dilated convolution for precise retinal vessel segmentation in fundus images</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>13428</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-98021-z</pub-id><pub-id pub-id-type="pmid">40251298</pub-id></mixed-citation></ref>
<ref id="B113"><label>113.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xiang</surname> <given-names>T</given-names></name> <name><surname>Zhang</surname> <given-names>C</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Song</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>D</given-names></name> <name><surname>Huang</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Towards bi-directional skip connections in encoder-decoder architectures and beyond</article-title>. <source>Med Image Anal</source>. (<year>2022</year>) <volume>78</volume>:<fpage>102420</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2022.102420</pub-id><pub-id pub-id-type="pmid">35334445</pub-id></mixed-citation></ref>
<ref id="B114"><label>114.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname> <given-names>M</given-names></name> <name><surname>Oh</surname> <given-names>S</given-names></name> <name><surname>Park</surname> <given-names>J</given-names></name> <name><surname>Jeong</surname> <given-names>T</given-names></name> <name><surname>Yu</surname> <given-names>S</given-names></name></person-group>. <article-title>ES-UNet: efficient 3D medical image segmentation with enhanced skip connections in 3D UNet</article-title>. <source>BMC Med Imaging</source>. (<year>2025</year>) <volume>25</volume>(<issue>1</issue>):<fpage>327</fpage>. <pub-id pub-id-type="doi">10.1186/s12880-025-01857-0</pub-id><pub-id pub-id-type="pmid">40804359</pub-id></mixed-citation></ref>
<ref id="B115"><label>115.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Z</given-names></name> <name><surname>Siddiquee</surname> <given-names>MMR</given-names></name> <name><surname>Tajbakhsh</surname> <given-names>N</given-names></name> <name><surname>Liang</surname> <given-names>J</given-names></name></person-group>. <article-title>UNet&#x002B;&#x002B;: redesigning skip connections to exploit multiscale features in image segmentation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2020</year>) <volume>39</volume>(<issue>6</issue>):<fpage>1856</fpage>&#x2013;<lpage>67</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2019.2959609</pub-id><pub-id pub-id-type="pmid">31841402</pub-id></mixed-citation></ref>
<ref id="B116"><label>116.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Polattimur</surname> <given-names>R</given-names></name> <name><surname>Y&#x0131;ld&#x0131;r&#x0131;m</surname> <given-names>MS</given-names></name> <name><surname>Dand&#x0131;l</surname> <given-names>E</given-names></name></person-group>. <article-title>Fractal-based architectures with skip connections and attention mechanism for improved segmentation of MS lesions in cervical spinal cord</article-title>. <source>Diagnostics (Basel)</source>. (<year>2025</year>) <volume>15</volume>(<issue>8</issue>):<fpage>1041</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics15081041</pub-id><pub-id pub-id-type="pmid">40310404</pub-id></mixed-citation></ref>
<ref id="B117"><label>117.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>Z</given-names></name> <name><surname>Siddiquee</surname> <given-names>MMR</given-names></name> <name><surname>Tajbakhsh</surname> <given-names>N</given-names></name> <name><surname>Liang</surname> <given-names>J</given-names></name></person-group>. <article-title>UNet&#x002B;&#x002B;: a nested U-net architecture for medical image segmentation</article-title>. <source>Deep Learn Med Image Anal Multimodal Learn Clin Decis Support</source>. (<year>2018</year>) <volume>11045</volume>:<fpage>3</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-030-00889-5_1</pub-id></mixed-citation></ref>
<ref id="B118"><label>118.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Zhu</surname> <given-names>W</given-names></name> <name><surname>Dong</surname> <given-names>X</given-names></name> <name><surname>Dumitrascu</surname> <given-names>OM</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name></person-group>. <article-title>EVIT-UNet: U-net like efficient vision transformer for medical image segmentation on mobile and edge devices</article-title>. <source>arXiv</source> [Preprint]. <volume>arXiv</volume>:<fpage>2410.15036</fpage>. <comment>Available online at:</comment> <ext-link ext-link-type="uri" xlink:href="https://arxiv.org/abs/2410.15036">https://arxiv.org/abs/2410.15036</ext-link> <comment>(Accessed on 19 October 2024)</comment>.</mixed-citation></ref>
<ref id="B119"><label>119.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gadosey</surname> <given-names>PK</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Adjei Agyekum</surname> <given-names>E</given-names></name> <name><surname>Zhang</surname> <given-names>T</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Yamak</surname> <given-names>PT</given-names></name><etal/></person-group> <article-title>SD-UNet: stripping down U-net for segmentation of biomedical images on platforms with low computational budgets</article-title>. <source>Diagnostics (Basel)</source>. (<year>2020</year>) <volume>10</volume>(<issue>2</issue>):<fpage>110</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics10020110</pub-id><pub-id pub-id-type="pmid">32085469</pub-id></mixed-citation></ref>
<ref id="B120"><label>120.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>Z</given-names></name> <name><surname>Ye</surname> <given-names>J</given-names></name> <name><surname>Wang</surname> <given-names>H</given-names></name> <name><surname>Deng</surname> <given-names>Z</given-names></name> <name><surname>Yang</surname> <given-names>Z</given-names></name> <name><surname>Su</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Revisiting model scaling with a U-net benchmark for 3D medical image segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>29795</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-15617-1</pub-id><pub-id pub-id-type="pmid">40813440</pub-id></mixed-citation></ref>
<ref id="B121"><label>121.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ibrahim</surname> <given-names>M</given-names></name> <name><surname>Khalil</surname> <given-names>YA</given-names></name> <name><surname>Amirrajab</surname> <given-names>S</given-names></name> <name><surname>Sun</surname> <given-names>C</given-names></name> <name><surname>Breeuwer</surname> <given-names>M</given-names></name> <name><surname>Pluim</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Generative AI for synthetic data across multiple medical modalities: a systematic review of recent developments and challenges</article-title>. <source>Comput Biol Med</source>. (<year>2025</year>) <volume>189</volume>:<fpage>109834</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2025.109834</pub-id><pub-id pub-id-type="pmid">40023073</pub-id></mixed-citation></ref>
<ref id="B122"><label>122.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alsaleh</surname> <given-names>AM</given-names></name> <name><surname>Albalawi</surname> <given-names>E</given-names></name> <name><surname>Algosaibi</surname> <given-names>A</given-names></name> <name><surname>Albakheet</surname> <given-names>SS</given-names></name> <name><surname>Khan</surname> <given-names>SB</given-names></name></person-group>. <article-title>Few-shot learning for medical image segmentation using 3D U-net and model-agnostic meta-learning (MAML)</article-title>. <source>Diagnostics (Basel)</source>. (<year>2024</year>) <volume>14</volume>(<issue>12</issue>):<fpage>1213</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics14121213</pub-id><pub-id pub-id-type="pmid">38928629</pub-id></mixed-citation></ref>
<ref id="B123"><label>123.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kabil</surname> <given-names>A</given-names></name> <name><surname>Khoriba</surname> <given-names>G</given-names></name> <name><surname>Yousef</surname> <given-names>M</given-names></name> <name><surname>Rashed</surname> <given-names>EA</given-names></name></person-group>. <article-title>Advances in medical image segmentation: a comprehensive survey with a focus on lumbar spine applications</article-title>. <source>Comput Biol Med</source>. (<year>2025</year>) <volume>198</volume>(<issue>Pt A</issue>):<fpage>111171</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2025.111171</pub-id><pub-id pub-id-type="pmid">41056615</pub-id></mixed-citation></ref>
<ref id="B124"><label>124.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dai</surname> <given-names>F</given-names></name></person-group>. <article-title>Deep learning based medical image compression using cross attention learning and wavelet transform</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>40008</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-23582-y</pub-id><pub-id pub-id-type="pmid">41238565</pub-id></mixed-citation></ref>
<ref id="B125"><label>125.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>J</given-names></name> <name><surname>Mei</surname> <given-names>J</given-names></name> <name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Lu</surname> <given-names>Y</given-names></name> <name><surname>Yu</surname> <given-names>Q</given-names></name> <name><surname>Wei</surname> <given-names>Q</given-names></name><etal/></person-group> <article-title>TransUNet: rethinking the U-net architecture design for medical image segmentation through the lens of transformers</article-title>. <source>Med Image Anal</source>. (<year>2024</year>) <volume>97</volume>:<fpage>103280</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2024.103280</pub-id><pub-id pub-id-type="pmid">39096845</pub-id></mixed-citation></ref>
<ref id="B126"><label>126.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kamath</surname> <given-names>A</given-names></name> <name><surname>Willmann</surname> <given-names>J</given-names></name> <name><surname>Andratschke</surname> <given-names>N</given-names></name> <name><surname>Reyes</surname> <given-names>M</given-names></name></person-group>. <article-title>The impact of U-net architecture choices and skip connections on the robustness of segmentation across texture variations</article-title>. <source>Comput Biol Med</source>. (<year>2025</year>) <volume>197</volume>(<issue>Pt B</issue>):<fpage>111056</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2025.111056</pub-id><pub-id pub-id-type="pmid">40945214</pub-id></mixed-citation></ref>
<ref id="B127"><label>127.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bouchekout</surname> <given-names>N</given-names></name> <name><surname>Boukabou</surname> <given-names>A</given-names></name> <name><surname>Grimes</surname> <given-names>M</given-names></name> <name><surname>Habchi</surname> <given-names>Y</given-names></name> <name><surname>Himeur</surname> <given-names>Y</given-names></name> <name><surname>Alkhazaleh</surname> <given-names>HA</given-names></name><etal/></person-group> <article-title>A novel hybrid deep learning and chaotic dynamics approach for thyroid cancer classification</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>40471</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-24334-8</pub-id><pub-id pub-id-type="pmid">41253983</pub-id></mixed-citation></ref>
<ref id="B128"><label>128.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raza</surname> <given-names>A</given-names></name> <name><surname>Hanif</surname> <given-names>F</given-names></name> <name><surname>Mohammed</surname> <given-names>HA</given-names></name></person-group>. <article-title>Efficient crack and surface-type recognition via CNN-block development mechanism and edge profiling</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>40073</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-25956-8</pub-id><pub-id pub-id-type="pmid">41249404</pub-id></mixed-citation></ref>
<ref id="B129"><label>129.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>L</given-names></name></person-group>. <article-title>Deep reinforcement learning-based thermal-visual collaborative optimization control system for multi-sensory art installations</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>38347</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-22173-1</pub-id><pub-id pub-id-type="pmid">41184354</pub-id></mixed-citation></ref>
<ref id="B130"><label>130.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jamshidi</surname> <given-names>M</given-names></name> <name><surname>Kalhor</surname> <given-names>A</given-names></name> <name><surname>Vahabie</surname> <given-names>AH</given-names></name></person-group>. <article-title>Efficient compression of encoder-decoder models for semantic segmentation using the separation index</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>24639</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-10348-9</pub-id><pub-id pub-id-type="pmid">40634505</pub-id></mixed-citation></ref>
<ref id="B131"><label>131.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Trinh</surname> <given-names>MN</given-names></name> <name><surname>Tran</surname> <given-names>TT</given-names></name> <name><surname>Nham</surname> <given-names>DH</given-names></name> <name><surname>Lo</surname> <given-names>MT</given-names></name> <name><surname>Pham</surname> <given-names>VT</given-names></name></person-group>. <article-title>GLAC-Unet: global-local active contour loss with an efficient U-shaped architecture for multiclass medical image segmentation</article-title>. <source>J Imaging Inform Med</source>. (<year>2025</year>) <volume>38</volume>(<issue>5</issue>):<fpage>3198</fpage>&#x2013;<lpage>220</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-025-01387-9</pub-id><pub-id pub-id-type="pmid">39821780</pub-id></mixed-citation></ref>
<ref id="B132"><label>132.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dastgir</surname> <given-names>A</given-names></name> <name><surname>Bin</surname> <given-names>W</given-names></name> <name><surname>Saeed</surname> <given-names>MU</given-names></name> <name><surname>Sheng</surname> <given-names>J</given-names></name> <name><surname>Site</surname> <given-names>L</given-names></name> <name><surname>Hassan</surname> <given-names>H</given-names></name></person-group>. <article-title>Attention LinkNet-152: a novel encoder-decoder based deep learning network for automated spine segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>13102</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-95243-z</pub-id><pub-id pub-id-type="pmid">40240440</pub-id></mixed-citation></ref>
<ref id="B133"><label>133.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gao</surname> <given-names>L</given-names></name> <name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Liu</surname> <given-names>C</given-names></name> <name><surname>Wu</surname> <given-names>S</given-names></name></person-group>. <article-title>Handling imbalanced medical image data: a deep-learning-based one-class classification approach</article-title>. <source>Artif Intell Med</source>. (<year>2020</year>) <volume>108</volume>:<fpage>101935</fpage>. <pub-id pub-id-type="doi">10.1016/j.artmed.2020.101935</pub-id><pub-id pub-id-type="pmid">32972664</pub-id></mixed-citation></ref>
<ref id="B134"><label>134.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Montazerolghaem</surname> <given-names>M</given-names></name> <name><surname>Sun</surname> <given-names>Y</given-names></name> <name><surname>Sasso</surname> <given-names>G</given-names></name> <name><surname>Haworth</surname> <given-names>A</given-names></name></person-group>. <article-title>U-Net architecture for prostate segmentation: the impact of loss function on system performance</article-title>. <source>Bioengineering (Basel)</source>. (<year>2023</year>) <volume>10</volume>(<issue>4</issue>):<fpage>412</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering10040412</pub-id><pub-id pub-id-type="pmid">37106600</pub-id></mixed-citation></ref>
<ref id="B135"><label>135.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mubashar</surname> <given-names>M</given-names></name> <name><surname>Ali</surname> <given-names>H</given-names></name> <name><surname>Gr&#x00F6;nlund</surname> <given-names>C</given-names></name> <name><surname>Azmat</surname> <given-names>S</given-names></name></person-group>. <article-title>R2u&#x002B;&#x002B;: a multiscale recurrent residual U-net with dense skip connections for medical image segmentation</article-title>. <source>Neural Comput Appl</source>. (<year>2022</year>) <volume>34</volume>(<issue>20</issue>):<fpage>17723</fpage>&#x2013;<lpage>39</lpage>. <pub-id pub-id-type="doi">10.1007/s00521-022-07419-7</pub-id><pub-id pub-id-type="pmid">35694048</pub-id></mixed-citation></ref>
<ref id="B136"><label>136.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ali</surname> <given-names>NM</given-names></name> <name><surname>Oyelere</surname> <given-names>SS</given-names></name> <name><surname>Jitani</surname> <given-names>N</given-names></name> <name><surname>Sarmah</surname> <given-names>R</given-names></name> <name><surname>Andrew</surname> <given-names>S</given-names></name></person-group>. <article-title>Hybrid intelligence in medical image segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>41200</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-24990-w</pub-id><pub-id pub-id-type="pmid">41271834</pub-id></mixed-citation></ref>
<ref id="B137"><label>137.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Azad</surname> <given-names>R</given-names></name> <name><surname>Kazerouni</surname> <given-names>A</given-names></name> <name><surname>Heidari</surname> <given-names>M</given-names></name> <name><surname>Aghdam</surname> <given-names>EK</given-names></name> <name><surname>Molaei</surname> <given-names>A</given-names></name> <name><surname>Jia</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Advances in medical image analysis with vision transformers: a comprehensive review</article-title>. <source>Med Image Anal</source>. (<year>2024</year>) <volume>91</volume>:<fpage>103000</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2023.103000</pub-id><pub-id pub-id-type="pmid">37883822</pub-id></mixed-citation></ref>
<ref id="B138"><label>138.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Safarov</surname> <given-names>F</given-names></name> <name><surname>Khojamuratova</surname> <given-names>U</given-names></name> <name><surname>Komoliddin</surname> <given-names>M</given-names></name> <name><surname>Kurbanov</surname> <given-names>Z</given-names></name> <name><surname>Tamara</surname> <given-names>A</given-names></name> <name><surname>Nizamjon</surname> <given-names>I</given-names></name><etal/></person-group> <article-title>Lightweight evolving U-net for next-generation biomedical imaging</article-title>. <source>Diagnostics (Basel)</source>. (<year>2025</year>) <volume>15</volume>(<issue>9</issue>):<fpage>1120</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics15091120</pub-id><pub-id pub-id-type="pmid">40361938</pub-id></mixed-citation></ref>
<ref id="B139"><label>139.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kirimtat</surname> <given-names>A</given-names></name> <name><surname>Krejcar</surname> <given-names>O</given-names></name></person-group>. <article-title>GPU-based parallel processing techniques for enhanced brain magnetic resonance imaging analysis: a review of recent advances</article-title>. <source>Sensors (Basel)</source>. (<year>2024</year>) <volume>24</volume>(<issue>5</issue>):<fpage>1591</fpage>. <pub-id pub-id-type="doi">10.3390/s24051591</pub-id><pub-id pub-id-type="pmid">38475138</pub-id></mixed-citation></ref>
<ref id="B140"><label>140.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>L</given-names></name> <name><surname>Wang</surname> <given-names>C</given-names></name></person-group>. <article-title>Trends in using deep learning algorithms in biomedical prediction systems</article-title>. <source>Front Neurosci</source>. (<year>2023</year>) <volume>17</volume>:<fpage>1256351</fpage>. <pub-id pub-id-type="doi">10.3389/fnins.2023.1256351</pub-id><pub-id pub-id-type="pmid">38027475</pub-id></mixed-citation></ref>
<ref id="B141"><label>141.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hosny</surname> <given-names>A</given-names></name> <name><surname>Parmar</surname> <given-names>C</given-names></name> <name><surname>Quackenbush</surname> <given-names>J</given-names></name> <name><surname>Schwartz</surname> <given-names>LH</given-names></name> <name><surname>Aerts</surname> <given-names>H</given-names></name></person-group>. <article-title>Artificial intelligence in radiology</article-title>. <source>Nat Rev Cancer</source>. (<year>2018</year>) <volume>18</volume>(<issue>8</issue>):<fpage>500</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1038/s41568-018-0016-5</pub-id><pub-id pub-id-type="pmid">29777175</pub-id></mixed-citation></ref>
<ref id="B142"><label>142.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mustafa</surname> <given-names>Z</given-names></name> <name><surname>Nsour</surname> <given-names>H</given-names></name></person-group>. <article-title>Using computer vision techniques to automatically detect abnormalities in chest x-rays</article-title>. <source>Diagnostics (Basel)</source>. (<year>2023</year>) <volume>13</volume>(<issue>18</issue>):<fpage>2927</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics13182979</pub-id><pub-id pub-id-type="pmid">37761294</pub-id></mixed-citation></ref>
<ref id="B143"><label>143.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Amerikanos</surname> <given-names>P</given-names></name> <name><surname>Maglogiannis</surname> <given-names>I</given-names></name></person-group>. <article-title>Image analysis in digital pathology utilizing machine learning and deep neural networks</article-title>. <source>J Pers Med</source>. (<year>2022</year>) <volume>12</volume>(<issue>9</issue>):<fpage>1444</fpage>. <pub-id pub-id-type="doi">10.3390/jpm12091444</pub-id><pub-id pub-id-type="pmid">36143229</pub-id></mixed-citation></ref>
<ref id="B144"><label>144.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tonti</surname> <given-names>E</given-names></name> <name><surname>Tonti</surname> <given-names>S</given-names></name> <name><surname>Mancini</surname> <given-names>F</given-names></name> <name><surname>Bonini</surname> <given-names>C</given-names></name> <name><surname>Spadea</surname> <given-names>L</given-names></name> <name><surname>D&#x0027;Esposito</surname> <given-names>F</given-names></name><etal/></person-group> <article-title>Artificial intelligence and advanced technology in glaucoma: a review</article-title>. <source>J Pers Med</source>. (<year>2024</year>) <volume>14</volume>(<issue>10</issue>):<fpage>1062</fpage>. <pub-id pub-id-type="doi">10.3390/jpm14101062</pub-id><pub-id pub-id-type="pmid">39452568</pub-id></mixed-citation></ref>
<ref id="B145"><label>145.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>JI</given-names></name> <name><surname>Rachitskaya</surname> <given-names>AV</given-names></name> <name><surname>Hallak</surname> <given-names>JA</given-names></name> <name><surname>Gholami</surname> <given-names>S</given-names></name> <name><surname>Alam</surname> <given-names>MN</given-names></name></person-group>. <article-title>Artificial intelligence for retinal diseases</article-title>. <source>Asia Pac J Ophthalmol (Phila)</source>. (<year>2024</year>) <volume>13</volume>(<issue>4</issue>):<fpage>100096</fpage>. <pub-id pub-id-type="doi">10.1016/j.apjo.2024.100096</pub-id><pub-id pub-id-type="pmid">39209215</pub-id></mixed-citation></ref>
<ref id="B146"><label>146.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Parmar</surname> <given-names>UPS</given-names></name> <name><surname>Surico</surname> <given-names>PL</given-names></name> <name><surname>Singh</surname> <given-names>RB</given-names></name> <name><surname>Romano</surname> <given-names>F</given-names></name> <name><surname>Salati</surname> <given-names>C</given-names></name> <name><surname>Spadea</surname> <given-names>L</given-names></name><etal/></person-group> <article-title>Artificial intelligence (AI) for early diagnosis of retinal diseases</article-title>. <source>Medicina (Kaunas)</source>. (<year>2024</year>) <volume>60</volume>(<issue>4</issue>):<fpage>527</fpage>. <pub-id pub-id-type="doi">10.3390/medicina60040527</pub-id><pub-id pub-id-type="pmid">38674173</pub-id></mixed-citation></ref>
<ref id="B147"><label>147.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>Z</given-names></name> <name><surname>Koban</surname> <given-names>KC</given-names></name> <name><surname>Schenck</surname> <given-names>TL</given-names></name> <name><surname>Giunta</surname> <given-names>RE</given-names></name> <name><surname>Li</surname> <given-names>Q</given-names></name> <name><surname>Sun</surname> <given-names>Y</given-names></name></person-group>. <article-title>Artificial intelligence in dermatology image analysis: current developments and future trends</article-title>. <source>J Clin Med</source>. (<year>2022</year>) <volume>11</volume>(<issue>22</issue>):<fpage>6826</fpage>. <pub-id pub-id-type="doi">10.3390/jcm11226826</pub-id><pub-id pub-id-type="pmid">36431301</pub-id></mixed-citation></ref>
<ref id="B148"><label>148.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Meyer-Szary</surname> <given-names>J</given-names></name> <name><surname>Luis</surname> <given-names>MS</given-names></name> <name><surname>Mikulski</surname> <given-names>S</given-names></name> <name><surname>Patel</surname> <given-names>A</given-names></name> <name><surname>Schulz</surname> <given-names>F</given-names></name> <name><surname>Tretiakow</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>The role of 3D printing in planning Complex medical procedures and training of medical professionals-cross-sectional multispecialty review</article-title>. <source>Int J Environ Res Public Health</source>. (<year>2022</year>) <volume>19</volume>(<issue>6</issue>):<fpage>3331</fpage>. <pub-id pub-id-type="doi">10.3390/ijerph19063331</pub-id><pub-id pub-id-type="pmid">35329016</pub-id></mixed-citation></ref>
<ref id="B149"><label>149.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Birlo</surname> <given-names>M</given-names></name> <name><surname>Edwards</surname> <given-names>PJE</given-names></name> <name><surname>Clarkson</surname> <given-names>M</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Utility of optical see-through head mounted displays in augmented reality-assisted surgery: a systematic review</article-title>. <source>Med Image Anal</source>. (<year>2022</year>) <volume>77</volume>:<fpage>102361</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2022.102361</pub-id><pub-id pub-id-type="pmid">35168103</pub-id></mixed-citation></ref>
<ref id="B150"><label>150.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dey</surname> <given-names>D</given-names></name> <name><surname>Slomka</surname> <given-names>PJ</given-names></name> <name><surname>Leeson</surname> <given-names>P</given-names></name> <name><surname>Comaniciu</surname> <given-names>D</given-names></name> <name><surname>Shrestha</surname> <given-names>S</given-names></name> <name><surname>Sengupta</surname> <given-names>PP</given-names></name><etal/></person-group> <article-title>Artificial intelligence in cardiovascular imaging: JACC state-of-the-art review</article-title>. <source>J Am Coll Cardiol</source>. (<year>2019</year>) <volume>73</volume>(<issue>11</issue>):<fpage>1317</fpage>&#x2013;<lpage>35</lpage>. <pub-id pub-id-type="doi">10.1016/j.jacc.2018.12.054</pub-id><pub-id pub-id-type="pmid">30898208</pub-id></mixed-citation></ref>
<ref id="B151"><label>151.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>J</given-names></name> <name><surname>Jeong</surname> <given-names>M</given-names></name> <name><surname>Stiles</surname> <given-names>WR</given-names></name> <name><surname>Choi</surname> <given-names>HS</given-names></name></person-group>. <article-title>Neuroimaging modalities in Alzheimer&#x2019;s disease: diagnosis and clinical features</article-title>. <source>Int J Mol Sci</source>. (<year>2022</year>) <volume>23</volume>(<issue>11</issue>):<fpage>6079</fpage>. <pub-id pub-id-type="doi">10.3390/ijms23116079</pub-id><pub-id pub-id-type="pmid">35682758</pub-id></mixed-citation></ref>
<ref id="B152"><label>152.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>M</given-names></name> <name><surname>Ouyang</surname> <given-names>Y</given-names></name> <name><surname>Yuan</surname> <given-names>Z</given-names></name></person-group>. <article-title>Deep learning aided neuroimaging and brain regulation</article-title>. <source>Sensors (Basel)</source>. (<year>2023</year>) <volume>23</volume>(<issue>11</issue>):<fpage>4993</fpage>. <pub-id pub-id-type="doi">10.3390/s23114993</pub-id><pub-id pub-id-type="pmid">37299724</pub-id></mixed-citation></ref>
<ref id="B153"><label>153.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hejazi</surname> <given-names>S</given-names></name> <name><surname>Karwowski</surname> <given-names>W</given-names></name> <name><surname>Farahani</surname> <given-names>FV</given-names></name> <name><surname>Marek</surname> <given-names>T</given-names></name> <name><surname>Hancock</surname> <given-names>PA</given-names></name></person-group>. <article-title>Graph-based analysis of brain connectivity in multiple sclerosis using functional MRI: a systematic review</article-title>. <source>Brain Sci</source>. (<year>2023</year>) <volume>13</volume>(<issue>2</issue>):<fpage>246</fpage>. <pub-id pub-id-type="doi">10.3390/brainsci13020246</pub-id><pub-id pub-id-type="pmid">36831789</pub-id></mixed-citation></ref>
<ref id="B154"><label>154.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>A</given-names></name> <name><surname>Mo</surname> <given-names>J</given-names></name> <name><surname>Zhong</surname> <given-names>C</given-names></name> <name><surname>Wu</surname> <given-names>S</given-names></name> <name><surname>Wei</surname> <given-names>S</given-names></name> <name><surname>Tu</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Artificial intelligence-assisted detection and classification of colorectal polyps under colonoscopy: a systematic review and meta-analysis</article-title>. <source>Ann Transl Med</source>. (<year>2021</year>) <volume>9</volume>(<issue>22</issue>):<fpage>1662</fpage>. <pub-id pub-id-type="doi">10.21037/atm-21-5081</pub-id><pub-id pub-id-type="pmid">34988171</pub-id></mixed-citation></ref>
<ref id="B155"><label>155.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Loizidou</surname> <given-names>K</given-names></name> <name><surname>Elia</surname> <given-names>R</given-names></name> <name><surname>Pitris</surname> <given-names>C</given-names></name></person-group>. <article-title>Computer-aided breast cancer detection and classification in mammography: a comprehensive review</article-title>. <source>Comput Biol Med</source>. (<year>2023</year>) <volume>153</volume>:<fpage>106554</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2023.106554</pub-id><pub-id pub-id-type="pmid">36646021</pub-id></mixed-citation></ref>
<ref id="B156"><label>156.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>D</given-names></name> <name><surname>Jin</surname> <given-names>R</given-names></name> <name><surname>Shieh</surname> <given-names>CC</given-names></name> <name><surname>Ng</surname> <given-names>AY</given-names></name> <name><surname>Pham</surname> <given-names>H</given-names></name> <name><surname>Dugal</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Real world validation of an AI-based CT hemorrhage detection tool</article-title>. <source>Front Neurol</source>. (<year>2023</year>) <volume>14</volume>:<fpage>1177723</fpage>. <pub-id pub-id-type="doi">10.3389/fneur.2023.1177723</pub-id><pub-id pub-id-type="pmid">37602253</pub-id></mixed-citation></ref>
<ref id="B157"><label>157.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Silva</surname> <given-names>H</given-names></name> <name><surname>Santos</surname> <given-names>GNM</given-names></name> <name><surname>Leite</surname> <given-names>AF</given-names></name> <name><surname>Mesquita</surname> <given-names>CRM</given-names></name> <name><surname>Figueiredo</surname> <given-names>PTS</given-names></name> <name><surname>Stefani</surname> <given-names>CM</given-names></name><etal/></person-group> <article-title>The use of artificial intelligence tools in cancer detection compared to the traditional diagnostic imaging methods: an overview of the systematic reviews</article-title>. <source>PLoS One</source>. (<year>2023</year>) <volume>18</volume>(<issue>10</issue>):<fpage>e0292063</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0292063</pub-id><pub-id pub-id-type="pmid">37796946</pub-id></mixed-citation></ref>
<ref id="B158"><label>158.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>YJ</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Qiu</surname> <given-names>ZX</given-names></name></person-group>. <article-title>Artificial intelligence research advances in discrimination and diagnosis of pulmonary ground-glass nodules</article-title>. <source>Zhonghua Jie He He Hu Xi Za Zhi</source>. (<year>2024</year>) <volume>47</volume>(<issue>6</issue>):<fpage>566</fpage>&#x2013;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.3760/cma.j.cn112147-20231214-00370</pub-id><pub-id pub-id-type="pmid">38858209</pub-id></mixed-citation></ref>
<ref id="B159"><label>159.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname> <given-names>Z</given-names></name> <name><surname>Xie</surname> <given-names>J</given-names></name> <name><surname>Wan</surname> <given-names>Y</given-names></name> <name><surname>Zhang</surname> <given-names>M</given-names></name> <name><surname>Qiao</surname> <given-names>L</given-names></name> <name><surname>Yu</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>A review of the current state of the computer-aided diagnosis (CAD) systems for breast cancer diagnosis</article-title>. <source>Open Life Sci</source>. (<year>2022</year>) <volume>17</volume>(<issue>1</issue>):<fpage>1600</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1515/biol-2022-0517</pub-id><pub-id pub-id-type="pmid">36561500</pub-id></mixed-citation></ref>
<ref id="B160"><label>160.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Patel</surname> <given-names>K</given-names></name> <name><surname>Huang</surname> <given-names>S</given-names></name> <name><surname>Rashid</surname> <given-names>A</given-names></name> <name><surname>Varghese</surname> <given-names>B</given-names></name> <name><surname>Gholamrezanezhad</surname> <given-names>A</given-names></name></person-group>. <article-title>A narrative review of the use of artificial intelligence in breast, lung, and prostate cancer</article-title><source>. Life (Basel)</source>. (<year>2023</year>) <volume>13</volume>(<issue>10</issue>):<fpage>2011</fpage>. <pub-id pub-id-type="doi">10.3390/life13102011</pub-id><pub-id pub-id-type="pmid">37895393</pub-id></mixed-citation></ref>
<ref id="B161"><label>161.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Walsh</surname> <given-names>J</given-names></name> <name><surname>Othmani</surname> <given-names>A</given-names></name> <name><surname>Jain</surname> <given-names>M</given-names></name> <name><surname>Dev</surname> <given-names>S</given-names></name></person-group>. <article-title>Using U-net network for efficient brain tumor segmentation in MRI images</article-title>. <source>Healthc Anal</source>. (<year>2022</year>) <volume>2</volume>:<fpage>100098</fpage>. <pub-id pub-id-type="doi">10.1016/j.health.2022.100098</pub-id></mixed-citation></ref>
<ref id="B162"><label>162.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Reza</surname> <given-names>SMS</given-names></name> <name><surname>Bradley</surname> <given-names>D</given-names></name> <name><surname>Aiosa</surname> <given-names>N</given-names></name> <name><surname>Castro</surname> <given-names>M</given-names></name> <name><surname>Lee</surname> <given-names>JH</given-names></name> <name><surname>Lee</surname> <given-names>BY</given-names></name><etal/></person-group> <article-title>Deep learning for automated liver segmentation to aid in the study of infectious diseases in nonhuman primates</article-title>. <source>Acad Radiol</source>. (<year>2021</year>) <volume>28 Suppl 1</volume>(<issue>Suppl 1</issue>):<fpage>S37</fpage>&#x2013;<lpage>44</lpage>. <pub-id pub-id-type="doi">10.1016/j.acra.2020.08.023</pub-id><pub-id pub-id-type="pmid">32943333</pub-id></mixed-citation></ref>
<ref id="B163"><label>163.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kugelman</surname> <given-names>J</given-names></name> <name><surname>Allman</surname> <given-names>J</given-names></name> <name><surname>Read</surname> <given-names>SA</given-names></name> <name><surname>Vincent</surname> <given-names>SJ</given-names></name> <name><surname>Tong</surname> <given-names>J</given-names></name> <name><surname>Kalloniatis</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>A comparison of deep learning U-net architectures for posterior segment OCT retinal layer segmentation</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>(<issue>1</issue>):<fpage>14888</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-18646-2</pub-id><pub-id pub-id-type="pmid">36050364</pub-id></mixed-citation></ref>
<ref id="B164"><label>164.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dhakal</surname> <given-names>A</given-names></name> <name><surname>McKay</surname> <given-names>C</given-names></name> <name><surname>Tanner</surname> <given-names>JJ</given-names></name> <name><surname>Cheng</surname> <given-names>J</given-names></name></person-group>. <article-title>Artificial intelligence in the prediction of protein-ligand interactions: recent advances and future directions</article-title>. <source>Brief Bioinform</source>. (<year>2022</year>) <volume>23</volume>(<issue>1</issue>):<fpage>bbab476</fpage>. <pub-id pub-id-type="doi">10.1093/bib/bbab476</pub-id><pub-id pub-id-type="pmid">34849575</pub-id></mixed-citation></ref>
<ref id="B165"><label>165.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>R</given-names></name> <name><surname>Butt</surname> <given-names>D</given-names></name> <name><surname>Cross</surname> <given-names>S</given-names></name> <name><surname>Verkade</surname> <given-names>P</given-names></name> <name><surname>Achim</surname> <given-names>A</given-names></name></person-group>. <article-title>Bright-field to fluorescence microscopy image translation for cell nuclei health quantification</article-title>. <source>Biol Imaging</source>. (<year>2023</year>) <volume>3</volume>:<fpage>e12</fpage>. <pub-id pub-id-type="doi">10.1017/s2633903x23000120</pub-id><pub-id pub-id-type="pmid">38510164</pub-id></mixed-citation></ref>
<ref id="B166"><label>166.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hirose</surname> <given-names>I</given-names></name> <name><surname>Tsunomura</surname> <given-names>M</given-names></name> <name><surname>Shishikura</surname> <given-names>M</given-names></name> <name><surname>Ishii</surname> <given-names>T</given-names></name> <name><surname>Yoshimura</surname> <given-names>Y</given-names></name> <name><surname>Ogawa-Ochiai</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>U-Net-based segmentation of microscopic images of colorants and simplification of labeling in the learning process</article-title>. <source>J Imaging</source>. (<year>2022</year>) <volume>8</volume>(<issue>7</issue>):<fpage>177</fpage>. <pub-id pub-id-type="doi">10.3390/jimaging8070177</pub-id><pub-id pub-id-type="pmid">35877621</pub-id></mixed-citation></ref>
<ref id="B167"><label>167.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yao</surname> <given-names>W</given-names></name> <name><surname>Zeng</surname> <given-names>Z</given-names></name> <name><surname>Lian</surname> <given-names>C</given-names></name> <name><surname>Tang</surname> <given-names>H</given-names></name></person-group>. <article-title>Pixel-wise regression using U-net and its application on pansharpening</article-title>. <source>Neurocomputing</source>. (<year>2018</year>) <volume>312</volume>:<fpage>364</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1016/j.neucom.2018.05.103</pub-id></mixed-citation></ref>
<ref id="B168"><label>168.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Waseem</surname> <given-names>F</given-names></name> <name><surname>Shahzad</surname> <given-names>M</given-names></name></person-group>. <comment>Video is worth a thousand images: exploring the latest trends in long video generation. <italic>arXiv Preprint; arXiv</italic>:2412.18688</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2412.18688</pub-id>. <comment>(Accessed December 24, 2024)</comment></mixed-citation></ref>
<ref id="B169"><label>169.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>&#x00C7;i&#x00E7;ek</surname> <given-names>&#x00D6;</given-names></name> <name><surname>Abdulkadir</surname> <given-names>A</given-names></name> <name><surname>Lienkamp</surname> <given-names>SS</given-names></name> <name><surname>Brox</surname> <given-names>T</given-names></name> <name><surname>Ronneberger</surname> <given-names>O</given-names></name></person-group>. <comment>3D U-Net: learning dense volumetric segmentation from sparse annotation</comment>. (<year>2016</year>).</mixed-citation></ref>
<ref id="B170"><label>170.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aldughayfiq</surname> <given-names>B</given-names></name> <name><surname>Ashfaq</surname> <given-names>F</given-names></name> <name><surname>Jhanjhi</surname> <given-names>NZ</given-names></name> <name><surname>Humayun</surname> <given-names>M</given-names></name></person-group>. <article-title>YOLO-based deep learning model for pressure ulcer detection and classification</article-title>. <source>Healthcare (Basel)</source>. (<year>2023</year>) <volume>11</volume>(<issue>9</issue>):<fpage>1222</fpage>. <pub-id pub-id-type="doi">10.3390/healthcare11091222</pub-id><pub-id pub-id-type="pmid">37174764</pub-id></mixed-citation></ref>
<ref id="B171"><label>171.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Mahendrakar</surname> <given-names>T</given-names></name> <name><surname>Ekblad</surname> <given-names>A</given-names></name> <name><surname>Fischer</surname> <given-names>N</given-names></name> <name><surname>White</surname> <given-names>RT</given-names></name> <name><surname>Wilde</surname> <given-names>M</given-names></name> <name><surname>Kish</surname> <given-names>B</given-names></name><etal/></person-group> <comment>Performance study of YOLOv5 and faster R-CNN for autonomous navigation around non-cooperative targets. <italic>arXiv Preprint; arXiv</italic>:2301.09056</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2301.09056</pub-id>. <comment>(Accessed January 22, 2022)</comment></mixed-citation></ref>
<ref id="B172"><label>172.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>A</given-names></name> <name><surname>Lin</surname> <given-names>D</given-names></name> <name><surname>Gao</surname> <given-names>Q</given-names></name></person-group>. <article-title>Enhancing brain tumor detection in MRI images using YOLO-NeuroBoost model</article-title>. <source>Front Neurol</source>. (<year>2024</year>) <volume>15</volume>:<fpage>1445882</fpage>. <pub-id pub-id-type="doi">10.3389/fneur.2024.1445882</pub-id><pub-id pub-id-type="pmid">39239397</pub-id></mixed-citation></ref>
<ref id="B173"><label>173.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sobek</surname> <given-names>J</given-names></name> <name><surname>Medina Inojosa</surname> <given-names>JR</given-names></name> <name><surname>Medina Inojosa</surname> <given-names>BJ</given-names></name> <name><surname>Rassoulinejad-Mousavi</surname> <given-names>SM</given-names></name> <name><surname>Conte</surname> <given-names>GM</given-names></name> <name><surname>Lopez-Jimenez</surname> <given-names>F</given-names></name><etal/></person-group> <article-title>MedYOLO: a medical image object detection framework</article-title>. <source>J Imaging Inform Med</source>. (<year>2024</year>) <volume>37</volume>(<issue>6</issue>):<fpage>3208</fpage>&#x2013;<lpage>16</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-024-01138-2</pub-id><pub-id pub-id-type="pmid">38844717</pub-id></mixed-citation></ref>
<ref id="B174"><label>174.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Khanam</surname> <given-names>R</given-names></name> <name><surname>Hussain</surname> <given-names>M</given-names></name></person-group>. <comment>YOLOv11: an overview of the key architectural enhancements. <italic>arXiv Preprint; arXiv</italic>:2410.17725</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2410.17725</pub-id>. <comment>(Accessed October 23, 2024)</comment></mixed-citation></ref>
<ref id="B175"><label>175.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abdusalomov</surname> <given-names>AB</given-names></name> <name><surname>Mukhiddinov</surname> <given-names>M</given-names></name> <name><surname>Whangbo</surname> <given-names>TK</given-names></name></person-group>. <article-title>Brain tumor detection based on deep learning approaches and magnetic resonance imaging</article-title>. <source>Cancers (Basel)</source>. (<year>2023</year>) <volume>15</volume>(<issue>16</issue>):<fpage>4172</fpage>. <pub-id pub-id-type="doi">10.3390/cancers15164172</pub-id><pub-id pub-id-type="pmid">37627200</pub-id></mixed-citation></ref>
<ref id="B176"><label>176.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bangert</surname> <given-names>P</given-names></name> <name><surname>Moon</surname> <given-names>H</given-names></name> <name><surname>Woo</surname> <given-names>JO</given-names></name> <name><surname>Didari</surname> <given-names>S</given-names></name> <name><surname>Hao</surname> <given-names>H</given-names></name></person-group>. <article-title>Active learning performance in labeling radiology images is 90&#x0025; effective</article-title>. <source>Front Radiol</source>. (<year>2021</year>) <volume>1</volume>:<fpage>748968</fpage>. <pub-id pub-id-type="doi">10.3389/fradi.2021.748968</pub-id><pub-id pub-id-type="pmid">37492167</pub-id></mixed-citation></ref>
<ref id="B177"><label>177.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bani Baker</surname> <given-names>Q</given-names></name> <name><surname>Hammad</surname> <given-names>M</given-names></name> <name><surname>Al-Smadi</surname> <given-names>M</given-names></name> <name><surname>Al-Jarrah</surname> <given-names>H</given-names></name> <name><surname>Al-Hamouri</surname> <given-names>R</given-names></name> <name><surname>Al-Zboon</surname> <given-names>SA</given-names></name></person-group>. <article-title>Enhanced COVID-19 detection from x-ray images with convolutional neural network and transfer learning</article-title>. <source>J Imaging</source>. (<year>2024</year>) <volume>10</volume>(<issue>10</issue>):<fpage>250</fpage>. <pub-id pub-id-type="doi">10.3390/jimaging10100250</pub-id><pub-id pub-id-type="pmid">39452413</pub-id></mixed-citation></ref>
<ref id="B178"><label>178.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>T</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Zhou</surname> <given-names>Y</given-names></name> <name><surname>Lan</surname> <given-names>J</given-names></name> <name><surname>Tan</surname> <given-names>T</given-names></name> <name><surname>Du</surname> <given-names>M</given-names></name><etal/></person-group> <comment>PCDAL: a perturbation consistency-driven active learning approach for medical image segmentation and classification. <italic>arXiv Preprint; arXiv</italic>:2306.16918</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2306.16918</pub-id>. <comment>(Accessed June 29, 2023)</comment></mixed-citation></ref>
<ref id="B179"><label>179.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Anaam</surname> <given-names>A</given-names></name> <name><surname>Al-Antari</surname> <given-names>MA</given-names></name> <name><surname>Hussain</surname> <given-names>J</given-names></name> <name><surname>Abdel Samee</surname> <given-names>N</given-names></name> <name><surname>Alabdulhafith</surname> <given-names>M</given-names></name> <name><surname>Gofuku</surname> <given-names>A</given-names></name></person-group>. <article-title>Deep active learning for automatic mitotic cell detection on HEp-2 specimen medical images</article-title>. <source>Diagnostics (Basel)</source>. (<year>2023</year>) <volume>13</volume>(<issue>8</issue>):<fpage>1416</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics13081416</pub-id><pub-id pub-id-type="pmid">37189517</pub-id></mixed-citation></ref>
<ref id="B180"><label>180.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>G</given-names></name> <name><surname>Wu</surname> <given-names>J</given-names></name> <name><surname>Luo</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Li</surname> <given-names>K</given-names></name> <name><surname>Zhang</surname> <given-names>S</given-names></name></person-group>. <comment>MIS-FM: 3D medical image segmentation using foundation models pretrained on a large-scale unannotated dataset. <italic>arXiv Preprint; arXiv</italic>:2306.16925</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2306.16925</pub-id>. <comment>(Accessed June 29, 2023)</comment></mixed-citation></ref>
<ref id="B181"><label>181.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>SK</given-names></name> <name><surname>Le</surname> <given-names>HN</given-names></name> <name><surname>Luu K</surname> <given-names>HVN</given-names></name> <name><surname>Ayache</surname> <given-names>N</given-names></name></person-group>. <article-title>Deep reinforcement learning in medical imaging: a literature review</article-title>. <source>Med Image Anal</source>. (<year>2021</year>) <volume>73</volume>:<fpage>102193</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2021.102193</pub-id><pub-id pub-id-type="pmid">34371440</pub-id></mixed-citation></ref>
<ref id="B182"><label>182.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rehman</surname> <given-names>MHU</given-names></name> <name><surname>Hugo Lopez Pinaya</surname> <given-names>W</given-names></name> <name><surname>Nachev</surname> <given-names>P</given-names></name> <name><surname>Teo</surname> <given-names>JT</given-names></name> <name><surname>Ourselin</surname> <given-names>S</given-names></name> <name><surname>Cardoso</surname> <given-names>MJ</given-names></name></person-group>. <article-title>Federated learning for medical imaging radiology</article-title>. <source>Br J Radiol</source>. (<year>2023</year>) <volume>96</volume>(<issue>1150</issue>):<fpage>20220890</fpage>. <pub-id pub-id-type="doi">10.1259/bjr.20220890</pub-id><pub-id pub-id-type="pmid">38011227</pub-id></mixed-citation></ref>
<ref id="B183"><label>183.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sandhu</surname> <given-names>SS</given-names></name> <name><surname>Gorji</surname> <given-names>HT</given-names></name> <name><surname>Tavakolian</surname> <given-names>P</given-names></name> <name><surname>Tavakolian</surname> <given-names>K</given-names></name> <name><surname>Akhbardeh</surname> <given-names>A</given-names></name></person-group>. <article-title>Medical imaging applications of federated learning</article-title>. <source>Diagnostics (Basel)</source>. (<year>2023</year>) <volume>13</volume>(<issue>19</issue>):<fpage>3140</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics13193140</pub-id><pub-id pub-id-type="pmid">37835883</pub-id></mixed-citation></ref>
<ref id="B184"><label>184.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Biswas</surname> <given-names>A</given-names></name> <name><surname>Al Nasim</surname> <given-names>MDA</given-names></name> <name><surname>Ali</surname> <given-names>MDS</given-names></name> <name><surname>Hossain</surname> <given-names>I</given-names></name> <name><surname>Ullah</surname> <given-names>MDA</given-names></name> <name><surname>Talukder</surname> <given-names>S</given-names></name></person-group>. <comment>Active Learning on Medical Image. (2023). <italic>arXiv Preprint; arXiv</italic>:2306.01827</comment>. <pub-id pub-id-type="doi">10.48550/arXiv.2306.01827</pub-id>. <comment>(Accessed June 2, 2023)</comment></mixed-citation></ref>
<ref id="B185"><label>185.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shen</surname> <given-names>Y</given-names></name> <name><surname>Shamout</surname> <given-names>FE</given-names></name> <name><surname>Oliver</surname> <given-names>JR</given-names></name> <name><surname>Witowski</surname> <given-names>J</given-names></name> <name><surname>Kannan</surname> <given-names>K</given-names></name> <name><surname>Park</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Artificial intelligence system reduces false-positive findings in the interpretation of breast ultrasound exams</article-title>. <source>Nat Commun</source>. (<year>2021</year>) <volume>12</volume>(<issue>1</issue>):<fpage>5645</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-021-26023-2</pub-id><pub-id pub-id-type="pmid">34561440</pub-id></mixed-citation></ref>
<ref id="B186"><label>186.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chadebecq</surname> <given-names>F</given-names></name> <name><surname>Vasconcelos</surname> <given-names>F</given-names></name> <name><surname>Mazomenos</surname> <given-names>E</given-names></name> <name><surname>Stoyanov</surname> <given-names>D</given-names></name></person-group>. <article-title>Computer vision in the surgical operating room</article-title>. <source>Visc Med</source>. (<year>2020</year>) <volume>36</volume>(<issue>6</issue>):<fpage>456</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1159/000511934</pub-id><pub-id pub-id-type="pmid">33447601</pub-id></mixed-citation></ref>
<ref id="B187"><label>187.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Oliveira</surname> <given-names>DF</given-names></name> <name><surname>Nogueira</surname> <given-names>AS</given-names></name> <name><surname>Brito</surname> <given-names>MA</given-names></name></person-group>. <article-title>Performance comparison of machine learning algorithms in classifying information technologies incident tickets</article-title>. <source>AIi</source>. (<year>2022</year>) <volume>3</volume>(<issue>3</issue>):<fpage>601</fpage>&#x2013;<lpage>22</lpage>. <pub-id pub-id-type="doi">10.3390/ai3030035</pub-id></mixed-citation></ref>
<ref id="B188"><label>188.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Peng</surname> <given-names>H</given-names></name> <name><surname>Ruan</surname> <given-names>Z</given-names></name> <name><surname>Atasoy</surname> <given-names>D</given-names></name> <name><surname>Sternson</surname> <given-names>S</given-names></name></person-group>. <article-title>Automatic reconstruction of 3D neuron structures using a graph-augmented deformable model</article-title>. <source>Bioinformatics</source>. (<year>2010</year>) <volume>26</volume>(<issue>12</issue>):<fpage>i38</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btq212</pub-id><pub-id pub-id-type="pmid">20529931</pub-id></mixed-citation></ref>
<ref id="B189"><label>189.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Manakitsa</surname> <given-names>N</given-names></name> <name><surname>Maraslidis</surname> <given-names>GS</given-names></name> <name><surname>Moysis</surname> <given-names>L</given-names></name> <name><surname>Fragulis</surname> <given-names>GF</given-names></name></person-group>. <article-title>A review of machine learning and deep learning for object detection, semantic segmentation, and human action recognition in machine and robotic vision</article-title>. <source>Technologies</source>. (<year>2024</year>) <volume>12</volume>(<issue>2</issue>):<fpage>15</fpage>. <pub-id pub-id-type="doi">10.3390/technologies12020015</pub-id></mixed-citation></ref>
<ref id="B190"><label>190.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schreuder</surname> <given-names>A</given-names></name> <name><surname>Scholten</surname> <given-names>ET</given-names></name> <name><surname>van Ginneken</surname> <given-names>B</given-names></name> <name><surname>Jacobs</surname> <given-names>C</given-names></name></person-group>. <article-title>Artificial intelligence for detection and characterization of pulmonary nodules in lung cancer CT screening: ready for practice?</article-title> <source>Transl Lung Cancer Res</source>. (<year>2021</year>) <volume>10</volume>(<issue>5</issue>):<fpage>2378</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.21037/tlcr-2020-lcs-06</pub-id><pub-id pub-id-type="pmid">34164285</pub-id></mixed-citation></ref>
<ref id="B191"><label>191.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>L</given-names></name></person-group>. <article-title>Mammography with deep learning for breast cancer detection</article-title>. <source>Front Oncol</source>. (<year>2024</year>) <volume>14</volume>:<fpage>1281922</fpage>. <pub-id pub-id-type="doi">10.3389/fonc.2024.1281922</pub-id><pub-id pub-id-type="pmid">38410114</pub-id></mixed-citation></ref>
<ref id="B192"><label>192.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lim</surname> <given-names>JI</given-names></name> <name><surname>Regillo</surname> <given-names>CD</given-names></name> <name><surname>Sadda</surname> <given-names>SR</given-names></name> <name><surname>Ipp</surname> <given-names>E</given-names></name> <name><surname>Bhaskaranand</surname> <given-names>M</given-names></name> <name><surname>Ramachandra</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>Artificial intelligence detection of diabetic retinopathy: subgroup comparison of the EyeArt system with Ophthalmologists&#x2019; dilated examinations</article-title>. <source>Ophthalmol Sci</source>. (<year>2023</year>) <volume>3</volume>(<issue>1</issue>):<fpage>100228</fpage>. <pub-id pub-id-type="doi">10.1016/j.xops.2022.100228</pub-id><pub-id pub-id-type="pmid">36345378</pub-id></mixed-citation></ref>
<ref id="B193"><label>193.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Varoquaux</surname> <given-names>G</given-names></name> <name><surname>Cheplygina</surname> <given-names>V</given-names></name></person-group>. <article-title>Machine learning for medical imaging: methodological failures and recommendations for the future</article-title>. <source>NPJ Digit Med</source>. (<year>2022</year>) <volume>5</volume>(<issue>1</issue>):<fpage>48</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-022-00592-y</pub-id><pub-id pub-id-type="pmid">35413988</pub-id></mixed-citation></ref>
<ref id="B194"><label>194.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhou</surname> <given-names>SK</given-names></name> <name><surname>Greenspan</surname> <given-names>H</given-names></name> <name><surname>Davatzikos</surname> <given-names>C</given-names></name> <name><surname>Duncan</surname> <given-names>JS</given-names></name> <name><surname>van Ginneken</surname> <given-names>B</given-names></name> <name><surname>Madabhushi</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>A review of deep learning in medical imaging: imaging traits, technology trends, case studies with progress highlights, and future promises</article-title>. <source>Proc IEEE Inst Electr Electron Eng</source>. (<year>2021</year>) <volume>109</volume>(<issue>5</issue>):<fpage>820</fpage>&#x2013;<lpage>38</lpage>. <pub-id pub-id-type="doi">10.1109/jproc.2021.3054390</pub-id><pub-id pub-id-type="pmid">37786449</pub-id></mixed-citation></ref>
<ref id="B195"><label>195.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khalifa</surname> <given-names>M</given-names></name> <name><surname>Albadawy</surname> <given-names>M</given-names></name></person-group>. <article-title>AI in diagnostic imaging: revolutionising accuracy and efficiency</article-title>. <source>Comput Methods Programs Biomed Update</source>. (<year>2024</year>) <volume>5</volume>:<fpage>100146</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpbup.2024.100146</pub-id></mixed-citation></ref>
<ref id="B196"><label>196.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shobayo</surname> <given-names>O</given-names></name> <name><surname>Saatchi</surname> <given-names>R</given-names></name></person-group>. <article-title>Developments in deep learning artificial neural network techniques for medical image analysis and interpretation</article-title>. <source>Diagnostics (Basel)</source>. (<year>2025</year>) <volume>15</volume>(<issue>9</issue>):<fpage>1072</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics15091072</pub-id><pub-id pub-id-type="pmid">40361891</pub-id></mixed-citation></ref>
<ref id="B197"><label>197.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hussain</surname> <given-names>J</given-names></name> <name><surname>B&#x00E5;th</surname> <given-names>M</given-names></name> <name><surname>Ivarsson</surname> <given-names>J</given-names></name></person-group>. <article-title>Generative adversarial networks in medical image reconstruction: a systematic literature review</article-title>. <source>Comput Biol Med</source>. (<year>2025</year>) <volume>191</volume>:<fpage>110094</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiomed.2025.110094</pub-id><pub-id pub-id-type="pmid">40198987</pub-id></mixed-citation></ref>
<ref id="B198"><label>198.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khosravi</surname> <given-names>B</given-names></name> <name><surname>Purkayastha</surname> <given-names>S</given-names></name> <name><surname>Erickson</surname> <given-names>BJ</given-names></name> <name><surname>Trivedi</surname> <given-names>HM</given-names></name> <name><surname>Gichoya</surname> <given-names>JW</given-names></name></person-group>. <article-title>Exploring the potential of generative artificial intelligence in medical image synthesis: opportunities, challenges, and future directions</article-title>. <source>Lancet Digit Health</source>. (<year>2025</year>) <volume>7</volume>(<issue>9</issue>):<fpage>100890</fpage>. <pub-id pub-id-type="doi">10.1016/j.landig.2025.100890</pub-id></mixed-citation></ref>
<ref id="B199"><label>199.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khaliq</surname> <given-names>A</given-names></name> <name><surname>Ahmad</surname> <given-names>F</given-names></name> <name><surname>Rehman</surname> <given-names>HU</given-names></name> <name><surname>Alanazi</surname> <given-names>SA</given-names></name> <name><surname>Haleem</surname> <given-names>H</given-names></name> <name><surname>Junaid</surname> <given-names>K</given-names></name><etal/></person-group> <article-title>Revolutionizing medical imaging: a cutting-edge AI framework with vision transformers and perceiver IO for multi-disease diagnosis</article-title>. <source>Comput Biol Chem</source>. (<year>2025</year>) <volume>119</volume>:<fpage>108586</fpage>. <pub-id pub-id-type="doi">10.1016/j.compbiolchem.2025.108586</pub-id><pub-id pub-id-type="pmid">40633409</pub-id></mixed-citation></ref>
<ref id="B200"><label>200.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Carriero</surname> <given-names>A</given-names></name> <name><surname>Groenhoff</surname> <given-names>L</given-names></name> <name><surname>Vologina</surname> <given-names>E</given-names></name> <name><surname>Basile</surname> <given-names>P</given-names></name> <name><surname>Albera</surname> <given-names>M</given-names></name></person-group>. <article-title>Deep learning in breast cancer imaging: state of the art and recent advancements in early 2024</article-title>. <source>Diagnostics (Basel)</source>. (<year>2024</year>) <volume>14</volume>(<issue>8</issue>):<fpage>848</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics14080848</pub-id><pub-id pub-id-type="pmid">38667493</pub-id></mixed-citation></ref>
<ref id="B201"><label>201.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chang</surname> <given-names>JY</given-names></name> <name><surname>Makary</surname> <given-names>MS</given-names></name></person-group>. <article-title>Evolving and novel applications of artificial intelligence in thoracic imaging</article-title>. <source>Diagnostics (Basel)</source>. (<year>2024</year>) <volume>14</volume>(<issue>13</issue>):<fpage>1456</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics14131456</pub-id><pub-id pub-id-type="pmid">39001346</pub-id></mixed-citation></ref>
<ref id="B202"><label>202.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yao</surname> <given-names>IZ</given-names></name> <name><surname>Dong</surname> <given-names>M</given-names></name> <name><surname>Hwang</surname> <given-names>WYK</given-names></name></person-group>. <article-title>Deep learning applications in clinical cancer detection: a review of implementation challenges and solutions</article-title>. <source>Mayo Clin Proc Digit Health</source>. (<year>2025</year>) <volume>3</volume>(<issue>3</issue>):<fpage>100253</fpage>. <pub-id pub-id-type="doi">10.1016/j.mcpdig.2025.100253</pub-id><pub-id pub-id-type="pmid">40822144</pub-id></mixed-citation></ref>
<ref id="B203"><label>203.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Piffer</surname> <given-names>S</given-names></name> <name><surname>Ubaldi</surname> <given-names>L</given-names></name> <name><surname>Tangaro</surname> <given-names>S</given-names></name> <name><surname>Retico</surname> <given-names>A</given-names></name> <name><surname>Talamonti</surname> <given-names>C</given-names></name></person-group>. <article-title>Tackling the small data problem in medical image classification with artificial intelligence: a systematic review</article-title>. <source>Prog Biomed Eng (Bristol)</source>. (<year>2024</year>) <volume>6</volume>:<fpage>032001</fpage>. <pub-id pub-id-type="doi">10.1088/2516-1091/ad525b</pub-id></mixed-citation></ref>
<ref id="B204"><label>204.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Teng</surname> <given-names>Q</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Song</surname> <given-names>Y</given-names></name> <name><surname>Han</surname> <given-names>K</given-names></name> <name><surname>Lu</surname> <given-names>Y</given-names></name></person-group>. <article-title>A survey on the interpretability of deep learning in medical diagnosis</article-title>. <source>Multimed Syst</source>. (<year>2022</year>) <volume>28</volume>(<issue>6</issue>):<fpage>2335</fpage>&#x2013;<lpage>55</lpage>. <pub-id pub-id-type="doi">10.1007/s00530-022-00960-4</pub-id><pub-id pub-id-type="pmid">35789785</pub-id></mixed-citation></ref>
<ref id="B205"><label>205.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Xiong</surname> <given-names>H</given-names></name> <name><surname>Sun</surname> <given-names>K</given-names></name> <name><surname>Bai</surname> <given-names>S</given-names></name> <name><surname>Dai</surname> <given-names>L</given-names></name> <name><surname>Ding</surname> <given-names>Z</given-names></name><etal/></person-group> <article-title>Toward general text-guided multimodal brain MRI synthesis for diagnosis and medical image analysis</article-title>. <source>Cell Rep Med</source>. (<year>2025</year>) <volume>6</volume>(<issue>6</issue>):<fpage>102182</fpage>. <pub-id pub-id-type="doi">10.1016/j.xcrm.2025.102182</pub-id><pub-id pub-id-type="pmid">40494349</pub-id></mixed-citation></ref>
<ref id="B206"><label>206.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gonz&#x00E1;lez-Gonzalo</surname> <given-names>C</given-names></name> <name><surname>Thee</surname> <given-names>EF</given-names></name> <name><surname>Klaver</surname> <given-names>CCW</given-names></name> <name><surname>Lee</surname> <given-names>AY</given-names></name> <name><surname>Schlingemann</surname> <given-names>RO</given-names></name> <name><surname>Tufail</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Trustworthy AI: closing the gap between development and integration of AI systems in ophthalmic practice</article-title>. <source>Prog Retin Eye Res</source>. (<year>2022</year>) <volume>90</volume>:<fpage>101034</fpage>. <pub-id pub-id-type="doi">10.1016/j.preteyeres.2021.101034</pub-id></mixed-citation></ref>
<ref id="B207"><label>207.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>A</given-names></name> <name><surname>Sengupta</surname> <given-names>S</given-names></name> <name><surname>Lakshminarayanan</surname> <given-names>V</given-names></name></person-group>. <article-title>Explainable deep learning models in medical image analysis</article-title>. <source>J Imaging</source>. (<year>2020</year>) <volume>6</volume>(<issue>6</issue>):<fpage>52</fpage>. <pub-id pub-id-type="doi">10.3390/jimaging6060052</pub-id><pub-id pub-id-type="pmid">34460598</pub-id></mixed-citation></ref>
<ref id="B208"><label>208.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Salehi</surname> <given-names>S</given-names></name> <name><surname>Singh</surname> <given-names>Y</given-names></name> <name><surname>Habibi</surname> <given-names>P</given-names></name> <name><surname>Erickson</surname> <given-names>BJ</given-names></name></person-group>. <article-title>Beyond single systems: how multi-agent AI is reshaping ethics in radiology</article-title>. <source>Bioengineering (Basel)</source>. (<year>2025</year>) <volume>12</volume>(<issue>10</issue>):<fpage>1100</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering12101100</pub-id><pub-id pub-id-type="pmid">41155099</pub-id></mixed-citation></ref>
<ref id="B209"><label>209.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brauneck</surname> <given-names>A</given-names></name> <name><surname>Schmalhorst</surname> <given-names>L</given-names></name> <name><surname>Kazemi Majdabadi</surname> <given-names>MM</given-names></name> <name><surname>Bakhtiari</surname> <given-names>M</given-names></name> <name><surname>V&#x00F6;lker</surname> <given-names>U</given-names></name> <name><surname>Baumbach</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Federated machine learning, privacy-enhancing technologies, and data protection laws in medical research: scoping review</article-title>. <source>J Med Internet Res</source>. (<year>2023</year>) <volume>25</volume>:<fpage>e41588</fpage>. <pub-id pub-id-type="doi">10.2196/41588</pub-id><pub-id pub-id-type="pmid">36995759</pub-id></mixed-citation></ref>
<ref id="B210"><label>210.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Koutsoubis</surname> <given-names>N</given-names></name> <name><surname>Waqas</surname> <given-names>A</given-names></name> <name><surname>Yilmaz</surname> <given-names>Y</given-names></name> <name><surname>Ramachandran</surname> <given-names>RP</given-names></name> <name><surname>Schabath</surname> <given-names>MB</given-names></name> <name><surname>Rasool</surname> <given-names>G</given-names></name></person-group>. <article-title>Privacy-preserving federated learning and uncertainty quantification in medical imaging</article-title>. <source>Radiol Artif Intell</source>. (<year>2025</year>) <volume>7</volume>(<issue>4</issue>):<fpage>e240637</fpage>. <pub-id pub-id-type="doi">10.1148/ryai.240637</pub-id><pub-id pub-id-type="pmid">40366260</pub-id></mixed-citation></ref>
<ref id="B211"><label>211.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Adnan</surname> <given-names>M</given-names></name> <name><surname>Kalra</surname> <given-names>S</given-names></name> <name><surname>Cresswell</surname> <given-names>JC</given-names></name> <name><surname>Taylor</surname> <given-names>GW</given-names></name> <name><surname>Tizhoosh</surname> <given-names>HR</given-names></name></person-group>. <article-title>Federated learning and differential privacy for medical image analysis</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>(<issue>1</issue>):<fpage>1953</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-05539-7</pub-id><pub-id pub-id-type="pmid">35121774</pub-id></mixed-citation></ref>
<ref id="B212"><label>212.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>C</given-names></name> <name><surname>Zhu</surname> <given-names>M</given-names></name> <name><surname>Zakaria</surname> <given-names>SAS</given-names></name></person-group>. <article-title>Cross-modal deep learning enhanced mixed reality accelerates construction skill transfer from experts to students</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>34462</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-17656-0</pub-id><pub-id pub-id-type="pmid">41039046</pub-id></mixed-citation></ref>
<ref id="B213"><label>213.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lambert</surname> <given-names>B</given-names></name> <name><surname>Forbes</surname> <given-names>F</given-names></name> <name><surname>Doyle</surname> <given-names>S</given-names></name> <name><surname>Dehaene</surname> <given-names>H</given-names></name> <name><surname>Dojat</surname> <given-names>M</given-names></name></person-group>. <article-title>Trustworthy clinical AI solutions: a unified review of uncertainty quantification in deep learning models for medical image analysis</article-title>. <source>Artif Intell Med</source>. (<year>2024</year>) <volume>150</volume>:<fpage>102830</fpage>. <pub-id pub-id-type="doi">10.1016/j.artmed.2024.102830</pub-id><pub-id pub-id-type="pmid">38553168</pub-id></mixed-citation></ref>
<ref id="B214"><label>214.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Albuquerque</surname> <given-names>C</given-names></name> <name><surname>Henriques</surname> <given-names>R</given-names></name> <name><surname>Castelli</surname> <given-names>M</given-names></name></person-group>. <article-title>Deep learning-based object detection algorithms in medical imaging: systematic review</article-title>. <source>Heliyon</source>. (<year>2025</year>) <volume>11</volume>(<issue>1</issue>):<fpage>e41137</fpage>. <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e41137</pub-id><pub-id pub-id-type="pmid">39758372</pub-id></mixed-citation></ref>
<ref id="B215"><label>215.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aly</surname> <given-names>M</given-names></name> <name><surname>Alotaibi</surname> <given-names>NS</given-names></name></person-group>. <article-title>A comprehensive deep learning framework for real time emotion detection in online learning using hybrid models</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>42012</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-26381-7</pub-id><pub-id pub-id-type="pmid">41291007</pub-id></mixed-citation></ref>
<ref id="B216"><label>216.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chaddad</surname> <given-names>A</given-names></name> <name><surname>Peng</surname> <given-names>J</given-names></name> <name><surname>Xu</surname> <given-names>J</given-names></name> <name><surname>Bouridane</surname> <given-names>A</given-names></name></person-group>. <article-title>Survey of explainable AI techniques in healthcare</article-title>. <source>Sensors (Basel)</source>. (<year>2023</year>) <volume>23</volume>(<issue>2</issue>):<fpage>634</fpage>. <pub-id pub-id-type="doi">10.3390/s23020634</pub-id><pub-id pub-id-type="pmid">36679430</pub-id></mixed-citation></ref>
<ref id="B217"><label>217.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Takahashi</surname> <given-names>S</given-names></name> <name><surname>Sakaguchi</surname> <given-names>Y</given-names></name> <name><surname>Kouno</surname> <given-names>N</given-names></name> <name><surname>Takasawa</surname> <given-names>K</given-names></name> <name><surname>Ishizu</surname> <given-names>K</given-names></name> <name><surname>Akagi</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>Comparison of vision transformers and convolutional neural networks in medical image analysis: a systematic review</article-title>. <source>J Med Syst</source>. (<year>2024</year>) <volume>48</volume>(<issue>1</issue>):<fpage>84</fpage>. <pub-id pub-id-type="doi">10.1007/s10916-024-02105-8</pub-id><pub-id pub-id-type="pmid">39264388</pub-id></mixed-citation></ref>
<ref id="B218"><label>218.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sebastian</surname> <given-names>N</given-names></name> <name><surname>Ankayarkanni</surname> <given-names>B</given-names></name></person-group>. <article-title>Enhanced ResNet-50 with multi-feature fusion for robust detection of pneumonia in chest x-ray images</article-title>. <source>Diagnostics (Basel)</source>. (<year>2025</year>) <volume>15</volume>(<issue>16</issue>):<fpage>2041</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics15162041</pub-id><pub-id pub-id-type="pmid">40870892</pub-id></mixed-citation></ref>
<ref id="B219"><label>219.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kim</surname> <given-names>JW</given-names></name> <name><surname>Khan</surname> <given-names>AU</given-names></name> <name><surname>Banerjee</surname> <given-names>I</given-names></name></person-group>. <article-title>Systematic review of hybrid vision transformer architectures for radiological image analysis</article-title>. <source>J Imaging Inform Med</source>. (<year>2025</year>) <volume>38</volume>(<issue>5</issue>):<fpage>3248</fpage>&#x2013;<lpage>62</lpage>. <pub-id pub-id-type="doi">10.1007/s10278-024-01322-4</pub-id><pub-id pub-id-type="pmid">39871042</pub-id></mixed-citation></ref>
<ref id="B220"><label>220.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ku&#x015F;</surname> <given-names>Z</given-names></name> <name><surname>Aydin</surname> <given-names>M</given-names></name></person-group>. <article-title>Medsegbench: a comprehensive benchmark for medical image segmentation in diverse data modalities</article-title>. <source>Sci Data</source>. (<year>2024</year>) <volume>11</volume>(<issue>1</issue>):<fpage>1283</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-024-04159-2</pub-id></mixed-citation></ref>
<ref id="B221"><label>221.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>J</given-names></name> <name><surname>Shi</surname> <given-names>R</given-names></name> <name><surname>Wei</surname> <given-names>D</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name> <name><surname>Zhao</surname> <given-names>L</given-names></name> <name><surname>Ke</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>MedMNIST v2 - A large-scale lightweight benchmark for 2D and 3D biomedical image classification</article-title>. <source>Sci Data</source>. (<year>2023</year>) <volume>10</volume>(<issue>1</issue>):<fpage>41</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-022-01721-8</pub-id><pub-id pub-id-type="pmid">36658144</pub-id></mixed-citation></ref>
<ref id="B222"><label>222.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Buric</surname> <given-names>M</given-names></name> <name><surname>Grozdanic</surname> <given-names>S</given-names></name> <name><surname>Ivasic-Kos</surname> <given-names>M</given-names></name></person-group>. <article-title>Diagnosis of ophthalmologic diseases in canines based on images using neural networks for image segmentation</article-title>. <source>Heliyon</source>. (<year>2024</year>) <volume>10</volume>(<issue>19</issue>):<fpage>e38287</fpage>. <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e38287</pub-id><pub-id pub-id-type="pmid">39397908</pub-id></mixed-citation></ref>
<ref id="B223"><label>223.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Deepak</surname> <given-names>GD</given-names></name> <name><surname>Bhat</surname> <given-names>SK</given-names></name></person-group>. <article-title>Optimization of deep learning-based faster R-CNN network for vehicle detection</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>38937</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-22828-z</pub-id><pub-id pub-id-type="pmid">41198791</pub-id></mixed-citation></ref>
<ref id="B224"><label>224.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Berwo</surname> <given-names>MA</given-names></name> <name><surname>Khan</surname> <given-names>A</given-names></name> <name><surname>Fang</surname> <given-names>Y</given-names></name> <name><surname>Fahim</surname> <given-names>H</given-names></name> <name><surname>Javaid</surname> <given-names>S</given-names></name> <name><surname>Mahmood</surname> <given-names>J</given-names></name><etal/></person-group> <article-title>Deep learning techniques for vehicle detection and classification from images/videos: a survey</article-title>. <source>Sensors (Basel)</source>. (<year>2023</year>) <volume>23</volume>(<issue>10</issue>):<fpage>4832</fpage>. <pub-id pub-id-type="doi">10.3390/s23104832</pub-id><pub-id pub-id-type="pmid">37430745</pub-id></mixed-citation></ref>
<ref id="B225"><label>225.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hsia</surname> <given-names>CCW</given-names></name> <name><surname>Bates</surname> <given-names>JHT</given-names></name> <name><surname>Driehuys</surname> <given-names>B</given-names></name> <name><surname>Fain</surname> <given-names>SB</given-names></name> <name><surname>Goldin</surname> <given-names>JG</given-names></name> <name><surname>Hoffman</surname> <given-names>EA</given-names></name><etal/></person-group> <article-title>Quantitative imaging metrics for the assessment of pulmonary pathophysiology: an official American thoracic society and Fleischner society joint workshop report</article-title>. <source>Ann Am Thorac Soc</source>. (<year>2023</year>) <volume>20</volume>(<issue>2</issue>):<fpage>161</fpage>&#x2013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1513/AnnalsATS.202211-915ST</pub-id><pub-id pub-id-type="pmid">36723475</pub-id></mixed-citation></ref>
<ref id="B226"><label>226.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Widyaningrum</surname> <given-names>R</given-names></name> <name><surname>Candradewi</surname> <given-names>I</given-names></name> <name><surname>Aji</surname> <given-names>N</given-names></name> <name><surname>Aulianisa</surname> <given-names>R</given-names></name></person-group>. <article-title>Comparison of multi-label U-net and mask R-CNN for panoramic radiograph segmentation to detect periodontitis</article-title>. <source>Imaging Sci Dent</source>. (<year>2022</year>) <volume>52</volume>(<issue>4</issue>):<fpage>383</fpage>&#x2013;<lpage>91</lpage>. <pub-id pub-id-type="doi">10.5624/isd.20220105</pub-id><pub-id pub-id-type="pmid">36605859</pub-id></mixed-citation></ref>
<ref id="B227"><label>227.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dogan</surname> <given-names>RO</given-names></name> <name><surname>Dogan</surname> <given-names>H</given-names></name> <name><surname>Bayrak</surname> <given-names>C</given-names></name> <name><surname>Kayikcioglu</surname> <given-names>T</given-names></name></person-group>. <article-title>A two-phase approach using mask R-CNN and 3D U-net for high-accuracy automatic segmentation of pancreas in CT imaging</article-title>. <source>Comput Methods Programs Biomed</source>. (<year>2021</year>) <volume>207</volume>:<fpage>106141</fpage>. <pub-id pub-id-type="doi">10.1016/j.cmpb.2021.106141</pub-id><pub-id pub-id-type="pmid">34020373</pub-id></mixed-citation></ref>
<ref id="B228"><label>228.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>ST</given-names></name> <name><surname>Huang</surname> <given-names>J</given-names></name> <name><surname>Lai</surname> <given-names>QQ</given-names></name> <name><surname>Guo</surname> <given-names>YF</given-names></name> <name><surname>Huang</surname> <given-names>YH</given-names></name><etal/></person-group> <article-title>Cardiac MRI segmentation of the atria based on UU-NET</article-title>. <source>Front Cardiovasc Med</source>. (<year>2022</year>) <volume>9</volume>:<fpage>1011916</fpage>. <pub-id pub-id-type="doi">10.3389/fcvm.2022.1011916</pub-id><pub-id pub-id-type="pmid">36505371</pub-id></mixed-citation></ref>
<ref id="B229"><label>229.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Turk</surname> <given-names>F</given-names></name> <name><surname>K&#x0131;l&#x0131;&#x00E7;aslan</surname> <given-names>M</given-names></name></person-group>. <article-title>Lung image segmentation with improved U-net, V-net and seg-net techniques</article-title>. <source>PeerJ Comput Sci</source>. (<year>2025</year>) <volume>11</volume>:<fpage>e2700</fpage>. <pub-id pub-id-type="doi">10.7717/peerj-cs.2700</pub-id><pub-id pub-id-type="pmid">40062241</pub-id></mixed-citation></ref>
<ref id="B230"><label>230.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Wen</surname> <given-names>Z</given-names></name> <name><surname>Bao</surname> <given-names>S</given-names></name> <name><surname>Huang</surname> <given-names>D</given-names></name> <name><surname>Wang</surname> <given-names>Y</given-names></name> <name><surname>Yang</surname> <given-names>B</given-names></name><etal/></person-group> <article-title>Diffusion-CSPAM U-net: a U-net model integrated hybrid attention mechanism and diffusion model for segmentation of computed tomography images of brain metastases</article-title>. <source>Radiat Oncol</source>. (<year>2025</year>) <volume>20</volume>(<issue>1</issue>):<fpage>50</fpage>. <pub-id pub-id-type="doi">10.1186/s13014-025-02622-x</pub-id><pub-id pub-id-type="pmid">40188354</pub-id></mixed-citation></ref>
<ref id="B231"><label>231.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>M&#x00FC;ller</surname> <given-names>D</given-names></name> <name><surname>Soto-Rey</surname> <given-names>I</given-names></name> <name><surname>Kramer</surname> <given-names>F</given-names></name></person-group>. <article-title>Towards a guideline for evaluation metrics in medical image segmentation</article-title>. <source>BMC Res Notes</source>. (<year>2022</year>) <volume>15</volume>(<issue>1</issue>):<fpage>210</fpage>. <pub-id pub-id-type="doi">10.1186/s13104-022-06096-y</pub-id></mixed-citation></ref>
<ref id="B232"><label>232.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname> <given-names>X</given-names></name> <name><surname>Xie</surname> <given-names>H</given-names></name> <name><surname>Gao</surname> <given-names>T</given-names></name> <name><surname>Cheng</surname> <given-names>N</given-names></name> <name><surname>Gou</surname> <given-names>J</given-names></name></person-group>. <article-title>Improved YOLO-based pulmonary nodule detection with spatial-SE attention and an aspect ratio penalty</article-title>. <source>Sensors (Basel)</source>. (<year>2025</year>) <volume>25</volume>(<issue>14</issue>):<fpage>4245</fpage>. <pub-id pub-id-type="doi">10.3390/s25144245</pub-id><pub-id pub-id-type="pmid">40732373</pub-id></mixed-citation></ref>
<ref id="B233"><label>233.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>Z</given-names></name> <name><surname>Shen</surname> <given-names>T</given-names></name> <name><surname>Ye</surname> <given-names>C</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Zhao</surname> <given-names>D</given-names></name> <name><surname>Zhang</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Hierarchical reasoning for lung cancer detection: from multi-scale perception to hypergraph inference with CR-YOLO</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>1</issue>):<fpage>726</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-02106-y</pub-id><pub-id pub-id-type="pmid">41310134</pub-id></mixed-citation></ref>
<ref id="B234"><label>234.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>X</given-names></name> <name><surname>Pu</surname> <given-names>X</given-names></name> <name><surname>Ling</surname> <given-names>W</given-names></name> <name><surname>Song</surname> <given-names>X</given-names></name></person-group>. <article-title>YOLO-SAM an end-to-end framework for efficient real time object detection and segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>40854</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-24576-6</pub-id><pub-id pub-id-type="pmid">41258157</pub-id></mixed-citation></ref>
<ref id="B235"><label>235.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hearne</surname> <given-names>LJ</given-names></name> <name><surname>Cocchi</surname> <given-names>L</given-names></name> <name><surname>Zalesky</surname> <given-names>A</given-names></name> <name><surname>Mattingley</surname> <given-names>JB</given-names></name></person-group>. <article-title>Reconfiguration of brain network architectures between resting-state and complexity-dependent cognitive reasoning</article-title>. <source>J Neurosci</source>. (<year>2017</year>) <volume>37</volume>(<issue>35</issue>):<fpage>8399</fpage>&#x2013;<lpage>411</lpage>. <pub-id pub-id-type="doi">10.1523/jneurosci.0485-17.2017</pub-id><pub-id pub-id-type="pmid">28760864</pub-id></mixed-citation></ref>
<ref id="B236"><label>236.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tun&#x00E7;</surname> <given-names>H</given-names></name> <name><surname>Akkaya</surname> <given-names>N</given-names></name> <name><surname>Aykanat</surname> <given-names>B</given-names></name> <name><surname>&#x00DC;nsal</surname> <given-names>G</given-names></name></person-group>. <article-title>U-Net-based deep learning for simultaneous segmentation and agenesis detection of primary and permanent teeth in panoramic radiographs</article-title>. <source>Diagnostics (Basel)</source>. (<year>2025</year>) <volume>15</volume>(<issue>20</issue>):<fpage>2577</fpage>. <pub-id pub-id-type="doi">10.3390/diagnostics15202577</pub-id></mixed-citation></ref>
<ref id="B237"><label>237.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pachetti</surname> <given-names>E</given-names></name> <name><surname>Colantonio</surname> <given-names>S</given-names></name></person-group>. <article-title>A systematic review of few-shot learning in medical imaging</article-title>. <source>Artif Intell Med</source>. (<year>2024</year>) <volume>156</volume>:<fpage>102949</fpage>. <pub-id pub-id-type="doi">10.1016/j.artmed.2024.102949</pub-id><pub-id pub-id-type="pmid">39178621</pub-id></mixed-citation></ref>
<ref id="B238"><label>238.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Feng</surname> <given-names>R</given-names></name> <name><surname>Zheng</surname> <given-names>X</given-names></name> <name><surname>Gao</surname> <given-names>T</given-names></name> <name><surname>Chen</surname> <given-names>J</given-names></name> <name><surname>Wang</surname> <given-names>W</given-names></name> <name><surname>Chen</surname> <given-names>DZ</given-names></name><etal/></person-group> <article-title>Interactive few-shot learning: limited supervision, better medical image segmentation</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2021</year>) <volume>40</volume>(<issue>10</issue>):<fpage>2575</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2021.3060551</pub-id><pub-id pub-id-type="pmid">33606628</pub-id></mixed-citation></ref>
<ref id="B239"><label>239.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>D</given-names></name> <name><surname>Wang</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>L</given-names></name> <name><surname>Li</surname> <given-names>M</given-names></name> <name><surname>Da</surname> <given-names>Q</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name><etal/></person-group> <article-title>A real-world dataset and benchmark for foundation model adaptation in medical image classification</article-title>. <source>Sci Data</source>. (<year>2023</year>) <volume>10</volume>(<issue>1</issue>):<fpage>574</fpage>. <pub-id pub-id-type="doi">10.1038/s41597-023-02460-0</pub-id><pub-id pub-id-type="pmid">37660106</pub-id></mixed-citation></ref>
<ref id="B240"><label>240.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Bareja</surname> <given-names>R</given-names></name> <name><surname>Carrillo-Perez</surname> <given-names>F</given-names></name> <name><surname>Zheng</surname> <given-names>Y</given-names></name> <name><surname>Pizurica</surname> <given-names>M</given-names></name> <name><surname>Nandi</surname> <given-names>TN</given-names></name> <name><surname>Shen</surname> <given-names>J</given-names></name><etal/></person-group> <comment>Evaluating vision and pathology foundation models for computational pathology: a comprehensive benchmark study. <italic>medRxiv</italic></comment>. (<year>2025</year>). <pub-id pub-id-type="doi">10.1101/2025.05.08.25327250</pub-id></mixed-citation></ref>
<ref id="B241"><label>241.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hicks</surname> <given-names>SA</given-names></name> <name><surname>Str&#x00FC;mke</surname> <given-names>I</given-names></name> <name><surname>Thambawita</surname> <given-names>V</given-names></name> <name><surname>Hammou</surname> <given-names>M</given-names></name> <name><surname>Riegler</surname> <given-names>MA</given-names></name> <name><surname>Halvorsen</surname> <given-names>P</given-names></name><etal/></person-group> <article-title>On evaluation metrics for medical applications of artificial intelligence</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>(<issue>1</issue>):<fpage>5979</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-022-09954-8</pub-id><pub-id pub-id-type="pmid">35395867</pub-id></mixed-citation></ref>
<ref id="B242"><label>242.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Correia</surname> <given-names>V</given-names></name> <name><surname>Mascarenhas</surname> <given-names>T</given-names></name> <name><surname>Mascarenhas</surname> <given-names>M</given-names></name></person-group>. <article-title>Smart pregnancy: AI-driven approaches to personalised maternal and foetal health-A scoping review</article-title>. <source>J Clin Med</source>. (<year>2025</year>) <volume>14</volume>(<issue>19</issue>):<fpage>6974</fpage>. <pub-id pub-id-type="doi">10.3390/jcm14196974</pub-id><pub-id pub-id-type="pmid">41096056</pub-id></mixed-citation></ref>
<ref id="B243"><label>243.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yilmaz</surname> <given-names>A</given-names></name> <name><surname>Gem</surname> <given-names>K</given-names></name> <name><surname>Kalebasi</surname> <given-names>M</given-names></name> <name><surname>Varol</surname> <given-names>R</given-names></name> <name><surname>Gencoglan</surname> <given-names>ZO</given-names></name> <name><surname>Samoylenko</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>An automated hip fracture detection, classification system on pelvic radiographs and comparison with 35 clinicians</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>16001</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-98852-w</pub-id><pub-id pub-id-type="pmid">40341645</pub-id></mixed-citation></ref>
<ref id="B244"><label>244.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Terasaki</surname> <given-names>Y</given-names></name> <name><surname>Yokota</surname> <given-names>H</given-names></name> <name><surname>Tashiro</surname> <given-names>K</given-names></name> <name><surname>Maejima</surname> <given-names>T</given-names></name> <name><surname>Takeuchi</surname> <given-names>T</given-names></name> <name><surname>Kurosawa</surname> <given-names>R</given-names></name><etal/></person-group> <article-title>Multidimensional deep learning reduces false-positives in the automated detection of cerebral aneurysms on time-of-flight magnetic resonance angiography: a multi-center study</article-title>. <source>Front Neurol</source>. (<year>2021</year>) <volume>12</volume>:<fpage>742126</fpage>. <pub-id pub-id-type="doi">10.3389/fneur.2021.742126</pub-id><pub-id pub-id-type="pmid">35115991</pub-id></mixed-citation></ref>
<ref id="B245"><label>245.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Zhang</surname> <given-names>H</given-names></name> <name><surname>Gichoya</surname> <given-names>JW</given-names></name> <name><surname>Katabi</surname> <given-names>D</given-names></name> <name><surname>Ghassemi</surname> <given-names>M</given-names></name></person-group>. <article-title>The limits of fair medical imaging AI in real-world generalization</article-title>. <source>Nat Med</source>. (<year>2024</year>) <volume>30</volume>(<issue>10</issue>):<fpage>2838</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1038/s41591-024-03113-4</pub-id><pub-id pub-id-type="pmid">38942996</pub-id></mixed-citation></ref>
<ref id="B246"><label>246.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raza</surname> <given-names>A</given-names></name> <name><surname>Hanif</surname> <given-names>F</given-names></name> <name><surname>Mohammed</surname> <given-names>HA</given-names></name></person-group>. <article-title>Analyzing the enhancement of CNN-YOLO and transformer based architectures for real-time animal detection in complex ecological environments</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>39142</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-26645-2</pub-id><pub-id pub-id-type="pmid">41203810</pub-id></mixed-citation></ref>
<ref id="B247"><label>247.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sankari</surname> <given-names>C</given-names></name> <name><surname>Jamuna</surname> <given-names>V</given-names></name> <name><surname>Kavitha</surname> <given-names>AR</given-names></name></person-group>. <article-title>Hierarchical multi-scale vision transformer model for accurate detection and classification of brain tumors in MRI-based medical imaging</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>38275</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-23100-0</pub-id><pub-id pub-id-type="pmid">41174141</pub-id></mixed-citation></ref>
<ref id="B248"><label>248.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vamsidhar</surname> <given-names>D</given-names></name> <name><surname>Desai</surname> <given-names>P</given-names></name> <name><surname>Joshi</surname> <given-names>S</given-names></name> <name><surname>Kolhar</surname> <given-names>S</given-names></name> <name><surname>Deshpande</surname> <given-names>N</given-names></name> <name><surname>Gite</surname> <given-names>S</given-names></name></person-group>. <article-title>Hybrid model integration with explainable AI for brain tumor diagnosis: a unified approach to MRI analysis and prediction</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>20542</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-06455-2</pub-id><pub-id pub-id-type="pmid">40596288</pub-id></mixed-citation></ref>
<ref id="B249"><label>249.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hartsock</surname> <given-names>I</given-names></name> <name><surname>Rasool</surname> <given-names>G</given-names></name></person-group>. <article-title>Vision-language models for medical report generation and visual question answering: a review</article-title>. <source>Front Artif Intell</source>. (<year>2024</year>) <volume>7</volume>:<fpage>1430984</fpage>. <pub-id pub-id-type="doi">10.3389/frai.2024.1430984</pub-id><pub-id pub-id-type="pmid">39628839</pub-id></mixed-citation></ref>
<ref id="B250"><label>250.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Noh</surname> <given-names>S</given-names></name> <name><surname>Lee</surname> <given-names>BD</given-names></name></person-group>. <article-title>A narrative review of foundation models for medical image segmentation: zero-shot performance evaluation on diverse modalities</article-title>. <source>Quant Imaging Med Surg</source>. (<year>2025</year>) <volume>15</volume>(<issue>6</issue>):<fpage>5825</fpage>&#x2013;<lpage>58</lpage>. <pub-id pub-id-type="doi">10.21037/qims-2024-2826</pub-id><pub-id pub-id-type="pmid">40606341</pub-id></mixed-citation></ref>
<ref id="B251"><label>251.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Gulhane</surname> <given-names>A</given-names></name> <name><surname>Mastrodicasa</surname> <given-names>D</given-names></name> <name><surname>Wu</surname> <given-names>W</given-names></name><etal/></person-group> <article-title>Demographic bias of expert-level vision-language foundation models in medical imaging</article-title>. <source>Sci Adv</source>. (<year>2025</year>) <volume>11</volume>(<issue>13</issue>):<fpage>eadq0305</fpage>. <pub-id pub-id-type="doi">10.1126/sciadv.adq0305</pub-id><pub-id pub-id-type="pmid">40138420</pub-id></mixed-citation></ref>
<ref id="B252"><label>252.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yuan</surname> <given-names>W</given-names></name> <name><surname>Feng</surname> <given-names>Y</given-names></name> <name><surname>Wen</surname> <given-names>T</given-names></name> <name><surname>Luo</surname> <given-names>G</given-names></name> <name><surname>Liang</surname> <given-names>J</given-names></name> <name><surname>Sun</surname> <given-names>Q</given-names></name><etal/></person-group> <article-title>MedIENet: medical image enhancement network based on conditional latent diffusion model</article-title>. <source>BMC Med Imaging</source>. (<year>2025</year>) <volume>25</volume>(<issue>1</issue>):<fpage>372</fpage>. <pub-id pub-id-type="doi">10.1186/s12880-025-01909-5</pub-id><pub-id pub-id-type="pmid">41013286</pub-id></mixed-citation></ref>
<ref id="B253"><label>253.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Akpinar</surname> <given-names>MH</given-names></name> <name><surname>Sengur</surname> <given-names>A</given-names></name> <name><surname>Salvi</surname> <given-names>M</given-names></name> <name><surname>Seoni</surname> <given-names>S</given-names></name> <name><surname>Faust</surname> <given-names>O</given-names></name> <name><surname>Mir</surname> <given-names>H</given-names></name><etal/></person-group> <article-title>Synthetic data generation via generative adversarial networks in healthcare: a systematic review of image- and signal-based studies</article-title>. <source>IEEE Open J Eng Med Biol</source>. (<year>2025</year>) <volume>6</volume>:<fpage>183</fpage>&#x2013;<lpage>92</lpage>. <pub-id pub-id-type="doi">10.1109/ojemb.2024.3508472</pub-id><pub-id pub-id-type="pmid">39698120</pub-id></mixed-citation></ref>
<ref id="B254"><label>254.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Skliarov</surname> <given-names>M</given-names></name> <name><surname>Shawi</surname> <given-names>RE</given-names></name> <name><surname>Dhaoui</surname> <given-names>C</given-names></name> <name><surname>Ahmed</surname> <given-names>N</given-names></name></person-group>. <article-title>A comparative evaluation of explainability techniques for image data</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>41898</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-25839-y</pub-id><pub-id pub-id-type="pmid">41290895</pub-id></mixed-citation></ref>
<ref id="B255"><label>255.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wollek</surname> <given-names>A</given-names></name> <name><surname>Graf</surname> <given-names>R</given-names></name> <name><surname>&#x010C;e&#x010D;atka</surname> <given-names>S</given-names></name> <name><surname>Fink</surname> <given-names>N</given-names></name> <name><surname>Willem</surname> <given-names>T</given-names></name> <name><surname>Sabel</surname> <given-names>BO</given-names></name><etal/></person-group> <article-title>Attention-based saliency maps improve interpretability of pneumothorax classification</article-title>. <source>Radiol Artif Intell</source>. (<year>2023</year>) <volume>5</volume>(<issue>2</issue>):<fpage>e220187</fpage>. <pub-id pub-id-type="doi">10.1148/ryai.220187</pub-id><pub-id pub-id-type="pmid">37035429</pub-id></mixed-citation></ref>
<ref id="B256"><label>256.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hossain</surname> <given-names>I</given-names></name> <name><surname>Zamzmi</surname> <given-names>G</given-names></name> <name><surname>Mouton</surname> <given-names>P</given-names></name> <name><surname>Sun</surname> <given-names>Y</given-names></name> <name><surname>Goldgof</surname> <given-names>D</given-names></name></person-group>. <article-title>Enhancing concept-based explanation with vision-language models</article-title>. <source>Proc IEEE Int Symp Comput Based Med Syst</source>. (<year>2024</year>) <volume>2024</volume>:<fpage>219</fpage>&#x2013;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1109/cbms61543.2024.00044</pub-id><pub-id pub-id-type="pmid">41000095</pub-id></mixed-citation></ref>
<ref id="B257"><label>257.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Gomez</surname> <given-names>C</given-names></name> <name><surname>Huang</surname> <given-names>CM</given-names></name> <name><surname>Unberath</surname> <given-names>M</given-names></name></person-group>. <article-title>Explainable medical imaging AI needs human-centered design: guidelines and evidence from a systematic review</article-title>. <source>NPJ Digit Med</source>. (<year>2022</year>) <volume>5</volume>(<issue>1</issue>):<fpage>156</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-022-00699-2</pub-id><pub-id pub-id-type="pmid">36261476</pub-id></mixed-citation></ref>
<ref id="B258"><label>258.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname> <given-names>Y</given-names></name> <name><surname>Lin</surname> <given-names>M</given-names></name> <name><surname>Zhao</surname> <given-names>H</given-names></name> <name><surname>Peng</surname> <given-names>Y</given-names></name> <name><surname>Huang</surname> <given-names>F</given-names></name> <name><surname>Lu</surname> <given-names>Z</given-names></name></person-group>. <article-title>A survey of recent methods for addressing AI fairness and bias in biomedicine</article-title>. <source>J Biomed Inform</source>. (<year>2024</year>) <volume>154</volume>:<fpage>104646</fpage>. <pub-id pub-id-type="doi">10.1016/j.jbi.2024.104646</pub-id><pub-id pub-id-type="pmid">38677633</pub-id></mixed-citation></ref>
<ref id="B259"><label>259.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xiang</surname> <given-names>A</given-names></name> <name><surname>Andrews</surname> <given-names>JTA</given-names></name> <name><surname>Bourke</surname> <given-names>RL</given-names></name> <name><surname>Thong</surname> <given-names>W</given-names></name> <name><surname>LaChance</surname> <given-names>JM</given-names></name> <name><surname>Georgievski</surname> <given-names>T</given-names></name><etal/></person-group> <article-title>Fair human-centric image dataset for ethical AI benchmarking</article-title>. <source>Nature</source>. (<year>2025</year>) <volume>648</volume>(<issue>8092</issue>):<fpage>97</fpage>&#x2013;<lpage>108</lpage>. <pub-id pub-id-type="doi">10.1038/s41586-025-09716-2</pub-id><pub-id pub-id-type="pmid">41193813</pub-id></mixed-citation></ref>
<ref id="B260"><label>260.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khader</surname> <given-names>F</given-names></name> <name><surname>M&#x00FC;ller-Franzes</surname> <given-names>G</given-names></name> <name><surname>Tayebi Arasteh</surname> <given-names>S</given-names></name> <name><surname>Han</surname> <given-names>T</given-names></name> <name><surname>Haarburger</surname> <given-names>C</given-names></name> <name><surname>Schulze-Hagen</surname> <given-names>M</given-names></name><etal/></person-group> <article-title>Denoising diffusion probabilistic models for 3D medical image generation</article-title>. <source>Sci Rep</source>. (<year>2023</year>) <volume>13</volume>(<issue>1</issue>):<fpage>7303</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-023-34341-2</pub-id><pub-id pub-id-type="pmid">37147413</pub-id></mixed-citation></ref>
<ref id="B261"><label>261.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>J</given-names></name> <name><surname>Liang</surname> <given-names>Z</given-names></name> <name><surname>Lu</surname> <given-names>X</given-names></name></person-group>. <article-title>A dual attention and cross layer fusion network with a hybrid CNN and transformer architecture for medical image segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>35707</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-19563-w</pub-id><pub-id pub-id-type="pmid">41083525</pub-id></mixed-citation></ref>
<ref id="B262"><label>262.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>Y</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Cao</surname> <given-names>X</given-names></name> <name><surname>Huang</surname> <given-names>C</given-names></name> <name><surname>Liu</surname> <given-names>E</given-names></name> <name><surname>Qian</surname> <given-names>S</given-names></name><etal/></person-group> <article-title>Artificial intelligence: a powerful paradigm for scientific research</article-title>. <source>Innovation (Camb)</source>. (<year>2021</year>) <volume>2</volume>(<issue>4</issue>):<fpage>100179</fpage>. <pub-id pub-id-type="doi">10.1016/j.xinn.2021.100179</pub-id><pub-id pub-id-type="pmid">34877560</pub-id></mixed-citation></ref>
<ref id="B263"><label>263.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ge</surname> <given-names>C</given-names></name> <name><surname>Pan</surname> <given-names>H</given-names></name> <name><surname>Song</surname> <given-names>Y</given-names></name> <name><surname>Zhang</surname> <given-names>X</given-names></name> <name><surname>Zhou</surname> <given-names>Z</given-names></name></person-group>. <article-title>SEFormer for medical image segmentation with integrated global and local features</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>41530</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-25450-1</pub-id><pub-id pub-id-type="pmid">41285947</pub-id></mixed-citation></ref>
<ref id="B264"><label>264.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Boulila</surname> <given-names>W</given-names></name> <name><surname>Ghandorh</surname> <given-names>H</given-names></name> <name><surname>Masood</surname> <given-names>S</given-names></name> <name><surname>Alzahem</surname> <given-names>A</given-names></name> <name><surname>Koubaa</surname> <given-names>A</given-names></name> <name><surname>Ahmed</surname> <given-names>F</given-names></name><etal/></person-group> <article-title>A transformer-based approach empowered by a self-attention technique for semantic segmentation in remote sensing</article-title>. <source>Heliyon</source>. (<year>2024</year>) <volume>10</volume>(<issue>8</issue>):<fpage>e29396</fpage>. <pub-id pub-id-type="doi">10.1016/j.heliyon.2024.e29396</pub-id><pub-id pub-id-type="pmid">38665569</pub-id></mixed-citation></ref>
<ref id="B265"><label>265.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ozdemir</surname> <given-names>B</given-names></name> <name><surname>Pacal</surname> <given-names>I</given-names></name></person-group>. <article-title>A robust deep learning framework for multiclass skin cancer classification</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>4938</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-89230-7</pub-id><pub-id pub-id-type="pmid">39930026</pub-id></mixed-citation></ref>
<ref id="B266"><label>266.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ahmed</surname> <given-names>MR</given-names></name> <name><surname>Rahman</surname> <given-names>H</given-names></name> <name><surname>Limon</surname> <given-names>ZH</given-names></name> <name><surname>Siddiqui</surname> <given-names>MIH</given-names></name> <name><surname>Khan</surname> <given-names>MA</given-names></name> <name><surname>Pranta</surname> <given-names>A</given-names></name><etal/></person-group> <article-title>Hierarchical swin transformer ensemble with explainable AI for robust and decentralized breast cancer diagnosis</article-title>. <source>Bioengineering (Basel)</source>. (<year>2025</year>) <volume>12</volume>(<issue>6</issue>):<fpage>651</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering12060651</pub-id><pub-id pub-id-type="pmid">40564466</pub-id></mixed-citation></ref>
<ref id="B267"><label>267.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>T</given-names></name> <name><surname>Cui</surname> <given-names>Z</given-names></name> <name><surname>Zhang</surname> <given-names>H</given-names></name></person-group>. <article-title>Semantic segmentation feature fusion network based on transformer</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>6110</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-90518-x</pub-id><pub-id pub-id-type="pmid">39971961</pub-id></mixed-citation></ref>
<ref id="B268"><label>268.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schiavella</surname> <given-names>C</given-names></name> <name><surname>Cirillo</surname> <given-names>L</given-names></name> <name><surname>Papa</surname> <given-names>L</given-names></name> <name><surname>Russo</surname> <given-names>P</given-names></name> <name><surname>Amerini</surname> <given-names>I</given-names></name></person-group>. <article-title>Efficient attention vision transformers for monocular depth estimation on resource-limited hardware</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>24001</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-06112-8</pub-id><pub-id pub-id-type="pmid">40615443</pub-id></mixed-citation></ref>
<ref id="B269"><label>269.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rodrigo</surname> <given-names>M</given-names></name> <name><surname>Cuevas</surname> <given-names>C</given-names></name> <name><surname>Garc&#x00ED;a</surname> <given-names>N</given-names></name></person-group>. <article-title>Comprehensive comparison between vision transformers and convolutional neural networks for face recognition tasks</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>21392</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-72254-w</pub-id><pub-id pub-id-type="pmid">39271805</pub-id></mixed-citation></ref>
<ref id="B270"><label>270.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ali</surname> <given-names>L</given-names></name> <name><surname>Alnajjar</surname> <given-names>F</given-names></name> <name><surname>Swavaf</surname> <given-names>M</given-names></name> <name><surname>Elharrouss</surname> <given-names>O</given-names></name> <name><surname>Abd-Alrazaq</surname> <given-names>A</given-names></name> <name><surname>Damseh</surname> <given-names>R</given-names></name></person-group>. <article-title>Evaluating segment anything model (SAM) on MRI scans of brain tumors</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>21659</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-72342-x</pub-id><pub-id pub-id-type="pmid">39289395</pub-id></mixed-citation></ref>
<ref id="B271"><label>271.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname> <given-names>K</given-names></name> <name><surname>Liang</surname> <given-names>L</given-names></name> <name><surname>Li</surname> <given-names>H</given-names></name> <name><surname>Situ</surname> <given-names>W</given-names></name> <name><surname>Zhao</surname> <given-names>W</given-names></name> <name><surname>Li</surname> <given-names>G</given-names></name></person-group>. <article-title>Research on medical image segmentation based on SAM and its future prospects</article-title>. <source>Bioengineering (Basel)</source>. (<year>2025</year>) <volume>12</volume>(<issue>6</issue>):<fpage>608</fpage>. <pub-id pub-id-type="doi">10.3390/bioengineering12060608</pub-id><pub-id pub-id-type="pmid">40564424</pub-id></mixed-citation></ref>
<ref id="B272"><label>272.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nanni</surname> <given-names>L</given-names></name> <name><surname>Fusaro</surname> <given-names>D</given-names></name> <name><surname>Fantozzi</surname> <given-names>C</given-names></name> <name><surname>Pretto</surname> <given-names>A</given-names></name></person-group>. <article-title>Improving existing segmentators performance with zero-shot segmentators</article-title>. <source>Entropy (Basel)</source>. (<year>2023</year>) <volume>25</volume>(<issue>11</issue>):<fpage>1502</fpage>. <pub-id pub-id-type="doi">10.3390/e25111502</pub-id><pub-id pub-id-type="pmid">37998194</pub-id></mixed-citation></ref>
<ref id="B273"><label>273.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yin</surname> <given-names>J</given-names></name> <name><surname>Wu</surname> <given-names>F</given-names></name> <name><surname>Su</surname> <given-names>H</given-names></name> <name><surname>Huang</surname> <given-names>P</given-names></name> <name><surname>Qixuan</surname> <given-names>Y</given-names></name></person-group>. <article-title>Improvement of SAM2 algorithm based on Kalman filtering for long-term video object segmentation</article-title>. <source>Sensors (Basel)</source>. (<year>2025</year>) <volume>25</volume>(<issue>13</issue>):<fpage>4199</fpage>. <pub-id pub-id-type="doi">10.3390/s25134199</pub-id><pub-id pub-id-type="pmid">40648454</pub-id></mixed-citation></ref>
<ref id="B274"><label>274.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>J</given-names></name> <name><surname>He</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>F</given-names></name> <name><surname>Han</surname> <given-names>L</given-names></name> <name><surname>You</surname> <given-names>C</given-names></name> <name><surname>Wang</surname> <given-names>B</given-names></name></person-group>. <article-title>Segment anything in medical images</article-title>. <source>Nat Commun</source>. (<year>2024</year>) <volume>15</volume>(<issue>1</issue>):<fpage>654</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-024-44824-z</pub-id><pub-id pub-id-type="pmid">38253604</pub-id></mixed-citation></ref>
<ref id="B275"><label>275.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>L</given-names></name> <name><surname>Hu</surname> <given-names>J</given-names></name> <name><surname>Huang</surname> <given-names>T</given-names></name></person-group>. <article-title>Improved SwinUNet with fusion transformer and large kernel convolutional attention for liver and tumor segmentation in CT images</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>14286</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-98938-5</pub-id><pub-id pub-id-type="pmid">40274913</pub-id></mixed-citation></ref>
<ref id="B276"><label>276.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jin</surname> <given-names>J</given-names></name> <name><surname>Xu</surname> <given-names>Y</given-names></name> <name><surname>He</surname> <given-names>H</given-names></name> <name><surname>Gao</surname> <given-names>F</given-names></name> <name><surname>Zeng</surname> <given-names>W</given-names></name> <name><surname>Wang</surname> <given-names>W</given-names></name><etal/></person-group> <article-title>A swin transformer-based hybrid reconstruction discriminative network for image anomaly detection</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>33929</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-10303-8</pub-id><pub-id pub-id-type="pmid">41028069</pub-id></mixed-citation></ref>
<ref id="B277"><label>277.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rajaguru H</surname> <given-names>SRS</given-names></name></person-group>. <article-title>MobileDANet integrating transfer learning and dynamic attention for classifying multi target histopathology images with explainable AI</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>37293</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-21360-4</pub-id><pub-id pub-id-type="pmid">41136678</pub-id></mixed-citation></ref>
<ref id="B278"><label>278.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>J</given-names></name> <name><surname>Cheang</surname> <given-names>CF</given-names></name> <name><surname>Yu</surname> <given-names>X</given-names></name> <name><surname>Tang</surname> <given-names>S</given-names></name> <name><surname>Du</surname> <given-names>Z</given-names></name> <name><surname>Cheng</surname> <given-names>Q</given-names></name></person-group>. <article-title>A segmentation network for enhancing autonomous driving scene understanding using skip connection and adaptive weighting</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>36692</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-20592-8</pub-id><pub-id pub-id-type="pmid">41120514</pub-id></mixed-citation></ref>
<ref id="B279"><label>279.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhong</surname> <given-names>X</given-names></name> <name><surname>Lu</surname> <given-names>G</given-names></name> <name><surname>Li</surname> <given-names>H</given-names></name></person-group>. <article-title>Vision Mamba and xLSTM-UNet for medical image segmentation</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>8163</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-88967-5</pub-id><pub-id pub-id-type="pmid">40059111</pub-id></mixed-citation></ref>
<ref id="B280"><label>280.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>L</given-names></name> <name><surname>Yin</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>X</given-names></name> <name><surname>Liu</surname> <given-names>Z</given-names></name></person-group>. <article-title>Medical image segmentation by combining feature enhancement swin transformer and UperNet</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>14565</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-97779-6</pub-id><pub-id pub-id-type="pmid">40281077</pub-id></mixed-citation></ref>
<ref id="B281"><label>281.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dan</surname> <given-names>Y</given-names></name> <name><surname>Jin</surname> <given-names>W</given-names></name> <name><surname>Yue</surname> <given-names>X</given-names></name> <name><surname>Wang</surname> <given-names>Z</given-names></name></person-group>. <article-title>Enhancing medical image segmentation with a multi-transformer U-net</article-title>. <source>PeerJ</source>. (<year>2024</year>) <volume>12</volume>:<fpage>e17005</fpage>. <pub-id pub-id-type="doi">10.7717/peerj.17005</pub-id><pub-id pub-id-type="pmid">38435997</pub-id></mixed-citation></ref>
<ref id="B282"><label>282.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>T</given-names></name> <name><surname>Hosseini</surname> <given-names>S</given-names></name> <name><surname>Anderson</surname> <given-names>C</given-names></name> <name><surname>Rinaldi</surname> <given-names>A</given-names></name> <name><surname>Krishnan</surname> <given-names>RG</given-names></name> <name><surname>Martel</surname> <given-names>AL</given-names></name><etal/></person-group> <article-title>A generalizable 3D framework and model for self-supervised learning in medical imaging</article-title>. <source>NPJ Digit Med</source>. (<year>2025</year>) <volume>8</volume>(<issue>1</issue>):<fpage>639</fpage>. <pub-id pub-id-type="doi">10.1038/s41746-025-02035-w</pub-id><pub-id pub-id-type="pmid">41203891</pub-id></mixed-citation></ref>
<ref id="B283"><label>283.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pang</surname> <given-names>Y</given-names></name> <name><surname>Liang</surname> <given-names>J</given-names></name> <name><surname>Huang</surname> <given-names>T</given-names></name> <name><surname>Chen</surname> <given-names>H</given-names></name> <name><surname>Li</surname> <given-names>Y</given-names></name> <name><surname>Li</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Slim UNETR: scale hybrid transformers to efficient 3D medical image segmentation under limited computational resources</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2024</year>) <volume>43</volume>(<issue>3</issue>):<fpage>994</fpage>&#x2013;<lpage>1005</lpage>. <pub-id pub-id-type="doi">10.1109/tmi.2023.3326188</pub-id><pub-id pub-id-type="pmid">37862274</pub-id></mixed-citation></ref>
<ref id="B284"><label>284.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Scabini</surname> <given-names>L</given-names></name> <name><surname>Sacilotti</surname> <given-names>A</given-names></name> <name><surname>Zielinski</surname> <given-names>KM</given-names></name> <name><surname>Ribas</surname> <given-names>LC</given-names></name> <name><surname>De Baets</surname> <given-names>B</given-names></name> <name><surname>Bruno</surname> <given-names>OM</given-names></name></person-group>. <article-title>A comparative survey of vision transformers for feature extraction in texture analysis</article-title>. <source>J Imaging</source>. (<year>2025</year>) <volume>11</volume>(<issue>9</issue>):<fpage>304</fpage>. <pub-id pub-id-type="doi">10.3390/jimaging11090304</pub-id><pub-id pub-id-type="pmid">41003354</pub-id></mixed-citation></ref>
<ref id="B285"><label>285.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Guodong</surname> <given-names>S</given-names></name> <name><surname>Huiyu</surname> <given-names>W</given-names></name></person-group>. <article-title>Global attention and local features using deep perceptron ensemble with vision transformers for landscape design detection</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>40900</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-24844-5</pub-id><pub-id pub-id-type="pmid">41258431</pub-id></mixed-citation></ref>
<ref id="B286"><label>286.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>Y</given-names></name> <name><surname>Lv</surname> <given-names>B</given-names></name> <name><surname>Xue</surname> <given-names>L</given-names></name> <name><surname>Zhang</surname> <given-names>W</given-names></name> <name><surname>Liu</surname> <given-names>Y</given-names></name> <name><surname>Fu</surname> <given-names>Y</given-names></name><etal/></person-group> <article-title>SemiSAM&#x002B;: rethinking semi-supervised medical image segmentation in the era of foundation models</article-title>. <source>Med Image Anal</source>. (<year>2025</year>) <volume>106</volume>:<fpage>103733</fpage>. <pub-id pub-id-type="doi">10.1016/j.media.2025.103733</pub-id><pub-id pub-id-type="pmid">40769095</pub-id></mixed-citation></ref>
<ref id="B287"><label>287.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ferber</surname> <given-names>D</given-names></name> <name><surname>W&#x00F6;lflein</surname> <given-names>G</given-names></name> <name><surname>Wiest</surname> <given-names>IC</given-names></name> <name><surname>Ligero</surname> <given-names>M</given-names></name> <name><surname>Sainath</surname> <given-names>S</given-names></name> <name><surname>Ghaffari Laleh</surname> <given-names>N</given-names></name><etal/></person-group> <article-title>In-context learning enables multimodal large language models to classify cancer pathology images</article-title>. <source>Nat Commun</source>. (<year>2024</year>) <volume>15</volume>(<issue>1</issue>):<fpage>10104</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-024-51465-9</pub-id><pub-id pub-id-type="pmid">39572531</pub-id></mixed-citation></ref>
<ref id="B288"><label>288.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Renugadevi</surname> <given-names>M</given-names></name> <name><surname>Narasimhan</surname> <given-names>K</given-names></name> <name><surname>Ramkumar</surname> <given-names>K</given-names></name> <name><surname>Raju</surname> <given-names>N</given-names></name></person-group>. <article-title>A novel hybrid vision UNet architecture for brain tumor segmentation and classification</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>(<issue>1</issue>):<fpage>23742</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-025-09833-y</pub-id><pub-id pub-id-type="pmid">40610748</pub-id></mixed-citation></ref>
<ref id="B289"><label>289.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>Z</given-names></name> <name><surname>Qin</surname> <given-names>P</given-names></name> <name><surname>Zeng</surname> <given-names>J</given-names></name> <name><surname>Song</surname> <given-names>Q</given-names></name> <name><surname>Zhao</surname> <given-names>P</given-names></name> <name><surname>Chai</surname> <given-names>R</given-names></name></person-group>. <article-title>LGIT: local-global interaction transformer for low-light image denoising</article-title>. <source>Sci Rep</source>. (<year>2024</year>) <volume>14</volume>(<issue>1</issue>):<fpage>21760</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-024-72912-z</pub-id><pub-id pub-id-type="pmid">39294345</pub-id></mixed-citation></ref>
<ref id="B290"><label>290.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ryu</surname> <given-names>JS</given-names></name> <name><surname>Kang</surname> <given-names>H</given-names></name> <name><surname>Chu</surname> <given-names>Y</given-names></name> <name><surname>Yang</surname> <given-names>S</given-names></name></person-group>. <article-title>Vision-language foundation models for medical imaging: a review of current practices and innovations</article-title>. <source>Biomed Eng Lett</source>. (<year>2025</year>) <volume>15</volume>(<issue>5</issue>):<fpage>809</fpage>&#x2013;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.1007/s13534-025-00484-6</pub-id><pub-id pub-id-type="pmid">40917147</pub-id></mixed-citation></ref>
<ref id="B291"><label>291.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Huang</surname> <given-names>J</given-names></name> <name><surname>Xiang</surname> <given-names>Y</given-names></name> <name><surname>Gan</surname> <given-names>S</given-names></name> <name><surname>Wu</surname> <given-names>L</given-names></name> <name><surname>Yan</surname> <given-names>J</given-names></name> <name><surname>Ye</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Application of artificial intelligence in medical imaging for tumor diagnosis and treatment: a comprehensive approach</article-title>. <source>Discov Oncol</source>. (<year>2025</year>) <volume>16</volume>(<issue>1</issue>):<fpage>1625</fpage>. <pub-id pub-id-type="doi">10.1007/s12672-025-03307-3</pub-id><pub-id pub-id-type="pmid">40856916</pub-id></mixed-citation></ref>
<ref id="B292"><label>292.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fang</surname> <given-names>M</given-names></name> <name><surname>Wang</surname> <given-names>Z</given-names></name> <name><surname>Pan</surname> <given-names>S</given-names></name> <name><surname>Feng</surname> <given-names>X</given-names></name> <name><surname>Zhao</surname> <given-names>Y</given-names></name> <name><surname>Hou</surname> <given-names>D</given-names></name><etal/></person-group> <article-title>Large models in medical imaging: advances and prospects</article-title>. <source>Chin Med J (Engl)</source>. (<year>2025</year>) <volume>138</volume>(<issue>14</issue>):<fpage>1647</fpage>&#x2013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1097/cm9.0000000000003699</pub-id><pub-id pub-id-type="pmid">40539307</pub-id></mixed-citation></ref>
<ref id="B293"><label>293.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>de Almeida</surname> <given-names>JG</given-names></name> <name><surname>Messiou</surname> <given-names>C</given-names></name> <name><surname>Withey</surname> <given-names>SJ</given-names></name> <name><surname>Matos</surname> <given-names>C</given-names></name> <name><surname>Koh</surname> <given-names>DM</given-names></name> <name><surname>Papanikolaou</surname> <given-names>N</given-names></name></person-group>. <article-title>Medical machine learning operations: a framework to facilitate clinical AI development and deployment in radiology</article-title>. <source>Eur Radiol</source>. (<year>2025</year>) <volume>35</volume>(<issue>11</issue>):<fpage>6828</fpage>&#x2013;<lpage>41</lpage>. <pub-id pub-id-type="doi">10.1007/s00330-025-11654-6</pub-id><pub-id pub-id-type="pmid">40341975</pub-id></mixed-citation></ref>
<ref id="B294"><label>294.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Reddy</surname> <given-names>S</given-names></name></person-group>. <article-title>Generative AI in healthcare: an implementation science informed translational path on application, integration and governance</article-title>. <source>Implement Sci</source>. (<year>2024</year>) <volume>19</volume>(<issue>1</issue>):<fpage>27</fpage>. <pub-id pub-id-type="doi">10.1186/s13012-024-01357-9</pub-id><pub-id pub-id-type="pmid">38491544</pub-id></mixed-citation></ref>
<ref id="B295"><label>295.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>M&#x00FC;ller-Franzes</surname> <given-names>G</given-names></name> <name><surname>Niehues</surname> <given-names>JM</given-names></name> <name><surname>Khader</surname> <given-names>F</given-names></name> <name><surname>Arasteh</surname> <given-names>ST</given-names></name> <name><surname>Haarburger</surname> <given-names>C</given-names></name> <name><surname>Kuhl</surname> <given-names>C</given-names></name><etal/></person-group> <article-title>A multimodal comparison of latent denoising diffusion probabilistic models and generative adversarial networks for medical image synthesis</article-title>. <source>Sci Rep</source>. (<year>2023</year>) <volume>13</volume>(<issue>1</issue>):<fpage>12098</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-023-39278-0</pub-id></mixed-citation></ref>
<ref id="B296"><label>296.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Foran</surname> <given-names>DJ</given-names></name> <name><surname>Chen</surname> <given-names>W</given-names></name> <name><surname>Kurc</surname> <given-names>T</given-names></name> <name><surname>Gupta</surname> <given-names>R</given-names></name> <name><surname>Kaczmarzyk</surname> <given-names>JR</given-names></name> <name><surname>Torre-Healy</surname> <given-names>LA</given-names></name><etal/></person-group> <article-title>An intelligent search &#x0026; retrieval system (IRIS) and clinical and research repository for decision support based on machine learning and joint kernel-based supervised hashing</article-title>. <source>Cancer Inform</source>. (<year>2024</year>) <volume>23</volume>:<fpage>11769351231223806</fpage>. <pub-id pub-id-type="doi">10.1177/11769351231223806</pub-id><pub-id pub-id-type="pmid">38322427</pub-id></mixed-citation></ref></ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by"><p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3116191/overview">Ziyang Wang</ext-link>, Aston University, United Kingdom</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by"><p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2246850/overview">Sibusiso Mdletshe</ext-link>, The University of Auckland, Auckland, New Zealand</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3115194/overview">Fangyijie Wang</ext-link>, University College Dublin, Ireland</p></fn>
</fn-group>
</back>
</article>