<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Public Health</journal-id>
<journal-title-group>
<journal-title>Frontiers in Public Health</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Public Health</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-2565</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpubh.2026.1772514</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Scalable and objective wound infection screening from clinical images using deep learning</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Wang</surname>
<given-names>Chao</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3263917"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wang</surname>
<given-names>Hongyu</given-names>
</name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Hu</surname>
<given-names>Jianhong</given-names>
</name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Huang</surname>
<given-names>Zhiyong</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Yang</surname>
<given-names>Yan</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tan</surname>
<given-names>Ziming</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Li</surname>
<given-names>Dan</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wu</surname>
<given-names>Li</given-names>
</name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Burn and Plastic Surgery, West China School of Medicine, Sichuan University, Sichuan University Affiliated Chengdu Second People's Hospital, Chengdu Second People's Hospital</institution>, <city>Chengdu</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Burns and Plastic Department, PLA No.983 Hospital</institution>, <city>Tianjin</city>, <country country="cn">China</country></aff>
<aff id="aff3"><label>3</label><institution>State Key Laboratory of Trauma, Burns and Combined Injury, Institute of Burn Research, Southwest Hospital, Third Military Medical University (Army Medical University)</institution>, <city>Chongqing</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Chao Wang, <email xlink:href="mailto:wangchaoxy@126.com">wangchaoxy@126.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-16">
<day>16</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>14</volume>
<elocation-id>1772514</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>31</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>04</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Wang, Wang, Hu, Huang, Yang, Tan, Li and Wu.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Wang, Wang, Hu, Huang, Yang, Tan, Li and Wu</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-16">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Background</title>
<p>Wound infection is a common and clinically significant complication that can delay healing, increase healthcare costs, and contribute to inappropriate antimicrobial use. Rapid, objective, and scalable screening tools are urgently needed, particularly in resource-limited or non-specialist clinical settings. This study aimed to develop and evaluate a deep learning&#x2013;based framework for automated wound infection detection using clinical wound images, with a focus on improving diagnostic consistency and supporting public health&#x2013;oriented wound management.</p>
</sec>
<sec>
<title>Methods</title>
<p>A dataset of 4,000 diverse clinical wound images was used to train and evaluate multiple deep learning models. The Swin Transformer architecture was compared with conventional convolutional neural networks. Model performance was assessed using accuracy, area under the receiver operating characteristic curve, and F1-score. To evaluate real-world applicability, model predictions were further compared with assessments made by non-specialist clinicians.</p>
</sec>
<sec>
<title>Results</title>
<p>The Swin Transformer outperformed conventional convolutional neural networks, achieving an accuracy of 0.9025 (95% CI: 0.8695&#x2013;0.9279), an area under the receiver operating characteristic curve of 0.9546, and an F1-score of 0.9042. Compared with non-specialist clinicians, the model reduced diagnostic variability and enabled earlier and more consistent recognition of wound infections.</p>
</sec>
<sec>
<title>Conclusion</title>
<p>Deep learning applied to clinical wound images provides a scalable and objective approach for wound infection screening. Such tools have the potential to support earlier detection, reduce diagnostic variability, and improve wound management and antimicrobial stewardship, particularly in public health and resource-limited settings.</p>
</sec>
</abstract>
<kwd-group>
<kwd>clinical images</kwd>
<kwd>deep learning</kwd>
<kwd>public health</kwd>
<kwd>Swin Transformer</kwd>
<kwd>wound infection detection</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="7"/>
<table-count count="4"/>
<equation-count count="7"/>
<ref-count count="25"/>
<page-count count="12"/>
<word-count count="5297"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Digital Public Health</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<title>Introduction</title>
<p>Wound infection is a prevalent and multifactorial clinical complication that frequently arises from trauma, surgical procedures, or chronic conditions such as diabetes and vascular diseases. These infections can delay tissue repair and significantly increase the risks of systemic sepsis, prolonged hospitalization, and mortality, highlighting the urgent need for effective wound management strategies (<xref ref-type="bibr" rid="ref1 ref2 ref3">1&#x2013;3</xref>).</p>
<p>Timely, accurate, and comprehensive assessment of patients and their wounds is essential to guide appropriate therapeutic interventions. Although conventional diagnostic modalities, such as microbial culture and histopathological examination, remain the gold standard, their clinical use is often limited by long turnaround times and technical constraints, which may delay urgent treatment decisions. Furthermore, positive microbial cultures can reflect mere colonization rather than true infection (<xref ref-type="bibr" rid="ref3">3</xref>, <xref ref-type="bibr" rid="ref4">4</xref>). Clinicians often rely on frameworks such as the International Wound Infection Continuum (IWII-WIC) to interpret clinical signs and symptoms. However, substantial inter-observer variability between specialists and non-specialists can compromise diagnostic consistency, potentially leading to delayed recognition or mismanagement of wound infections (<xref ref-type="bibr" rid="ref5">5</xref>). This underscores the critical need for rapid, objective, and scalable diagnostic tools that can support non-specialist clinicians, patients, and caregivers in early identification and timely referral for specialist evaluation.</p>
<p>Recent advances in deep learning (DL), particularly convolutional neural networks (CNNs) and Transformer-based architectures, have revolutionized medical image analysis by achieving state-of-the-art performance in image classification and pattern recognition (<xref ref-type="bibr" rid="ref6">6</xref>). DL models trained on large, annotated medical image datasets can autonomously extract discriminative features and identify pathological patterns, providing real-time diagnostic support. Such approaches have shown transformative potential in applications including cancer screening (<xref ref-type="bibr" rid="ref7 ref8 ref9">7&#x2013;9</xref>), neurological disorder detection (<xref ref-type="bibr" rid="ref10 ref11 ref12">10&#x2013;12</xref>), hematological disorder classification (<xref ref-type="bibr" rid="ref13">13</xref>, <xref ref-type="bibr" rid="ref14">14</xref>) and radiological diagnostics (<xref ref-type="bibr" rid="ref15">15</xref>). In the specific domain of wound care, recent studies such as SwinDFU-Net have demonstrated the efficacy of hybrid architectures for identifying infections in diabetic foot ulcers (<xref ref-type="bibr" rid="ref16">16</xref>). However, most existing models are specialized for a single wound etiology, potentially limiting their generalizability in diverse clinical environments.</p>
<p>In this study, we aimed to evaluate the feasibility and effectiveness of DL-based approaches for automated wound infection screening. By developing and optimizing models on a diverse clinical wound image dataset encompassing various types of wounds&#x2014;including burn wounds, traumatic wounds, surgical incisions, diabetic foot ulcers, pressure injuries, and vascular ulcers&#x2014;at different stages of the healing process, we sought to establish a robust, scalable, and objective pipeline. The proposed framework is designed to improve diagnostic accuracy, reduce observer-dependent variability, and streamline clinical workflows, ultimately supporting timely, precise, and equitable wound care, particularly in resource-limited or non-specialist settings.</p>
</sec>
<sec sec-type="methods" id="sec2">
<title>Methods</title>
<p>In this study, clinical wound images were classified as infected or non-infected according to the pipeline illustrated in <xref ref-type="fig" rid="fig1">Figure 1</xref>.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Overview of the deep learning pipeline designed to classify clinical wound images to determine whether the wound is infected.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart illustrating a machine learning workflow for wound classification, starting with data collection and preprocessing, including annotation protocol, data augmentation, and dataset splitting. Model development involves various architectures and hyperparameter tuning. Evaluation metrics assessed by non-specialist clinicians include accuracy, sensitivity, specificity, PPV, NPV, F1 score, and AUC. Final statistical analysis uses DeLong&#x2019;s method and McNemar&#x2019;s test.</alt-text>
</graphic>
</fig>
<sec id="sec3">
<title>Data collection and preprocessing</title>
<sec id="sec4">
<title>Data sources</title>
<p>This study established a curated dataset comprising 4,000 wound images. All images were sourced from a single clinical institution (Sichuan University affiliated Chengdu Second People&#x2019;s Hospital). These images encompass various types of wounds, including burn wounds, traumatic wounds, surgical incisions, diabetic foot ulcers, pressure injuries, and vascular ulcers, among other conditions, covering different stages of the healing process.</p>
</sec>
<sec id="sec5">
<title>Data annotation</title>
<p>A balanced dataset of 4,000 wound images (2,000 infected and 2,000 non-infected) was utilized. To establish a high-quality ground truth, the annotation process was independently conducted by three senior wound specialists. The annotation criteria were strictly aligned with the established International Wound Infection Institute (IWII-WIC) framework (<xref ref-type="bibr" rid="ref4">4</xref>). A multi-round cross-validation protocol was adopted. Each image was initially evaluated in a double-blind manner. Discrepancies were resolved through collective discussion until a 100% consensus was achieved. To quantify the reliability of the annotation framework, inter-observer agreement was assessed using the Fleiss&#x2019; Kappa coefficient, which yielded a value of 0.84.</p>
</sec>
<sec id="sec6">
<title>Data augmentation</title>
<p>Apply data augmentation techniques such as rotation, flipping, scaling, color adjustment and Mixup to increase the diversity of the training data and enhance the robustness of the deep learning models (<xref ref-type="bibr" rid="ref17">17</xref>, <xref ref-type="bibr" rid="ref18">18</xref>).</p>
</sec>
<sec id="sec7">
<title>Data splitting</title>
<p>To ensure unbiased evaluation, the 4,000 images were divided into training (80%, <italic>n</italic> =&#x202F;3,200), validation (10%, <italic>n</italic> =&#x202F;400), and testing (10%, <italic>n</italic> =&#x202F;400) sets using patient-level separation. This strategy ensures that all images from a single patient belong exclusively to one subset, preventing the model from learning patient-specific features. All subsets were carefully balanced to remain representative of the overall dataset and mitigate potential sampling bias.</p>
</sec>
</sec>
<sec id="sec8">
<title>Model development</title>
<sec id="sec9">
<title>Model selection</title>
<p>Choose appropriate deep learning architectures for the task. AlexNet, ResNet34, ResNet152, InceptionV3, and the Swin Transformer (Swin-T) are commonly used for image classification tasks (<xref ref-type="bibr" rid="ref19 ref20 ref21 ref22 ref23 ref24">19&#x2013;24</xref>). Consider using pre-trained models (transfer learning) to leverage existing knowledge and accelerate training.</p>
</sec>
<sec id="sec10">
<title>Model architecture</title>
<p>Design or modify the chosen deep learning architecture to suit the specific requirements of wound infection detection. This may involve adjusting layers, activation functions, and regularization techniques.</p>
</sec>
<sec id="sec11">
<title>Training</title>
<p>Train the model using the training dataset. Implement techniques such as early stopping and learning rate scheduling to optimize performance. Use cross-entropy loss or other suitable loss functions for classification tasks.</p>
</sec>
<sec id="sec12">
<title>Hyperparameter tuning</title>
<p>Optimize hyperparameters, including learning rate, batch size, and number of epochs, to achieve the best model performance. Utilize grid search or random search methods to find the optimal hyperparameters.</p>
</sec>
</sec>
<sec id="sec13">
<title>Model evaluation</title>
<p>The performance of deep learning models is assessed using standard evaluation metrics derived from the confusion matrix. In this matrix, correct predictions are categorized as True Positives (TP) for correctly identified positive cases, and True Negatives (TN) for correctly identified negative cases. Incorrect predictions are categorized as False Positives (FP), where a negative case is incorrectly classified as positive, and False Negatives (FN), where a positive case is incorrectly classified as negative.</p>
<p>This study evaluates model performance using key metrics including accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1-score, and the area under the receiver operating characteristic curve (AUC-ROC). These metrics collectively help to assess the model&#x2019;s effectiveness in correctly distinguishing between infected and non-infected wounds.</p>
<p>The evaluation metrics are defined as follows:</p>
<disp-formula id="E1">
<mml:math id="M1">
<mml:mtext>Accuracy</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>TN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<disp-formula id="E2">
<mml:math id="M2">
<mml:mtext>Sensitivity</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi>TP</mml:mi>
<mml:mrow>
<mml:mi>TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<disp-formula id="E3">
<mml:math id="M3">
<mml:mtext>Specificity</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mspace width="0.25em"/>
<mml:mi>TN</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>TN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<disp-formula id="E4">
<mml:math id="M4">
<mml:mi>PPV</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi>TP</mml:mi>
<mml:mrow>
<mml:mi>TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<disp-formula id="E5">
<mml:math id="M5">
<mml:mi>NPV</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi>TN</mml:mi>
<mml:mrow>
<mml:mi>TN</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<disp-formula id="E6">
<mml:math id="M6">
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mn>1</mml:mn>
<mml:mspace width="0.25em"/>
<mml:mtext>Score</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x00D7;</mml:mo>
<mml:mi>TP</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
</disp-formula>
<p>AUC-ROC: Area under the receiver operating characteristic curve, computed using the trapezoidal rule (<xref ref-type="bibr" rid="ref25">25</xref>).</p>
<p>All metrics were reported with 95% confidence intervals (CI) calculated via 1,000 bootstrap resamples to account for sampling variability.</p>
<p>Validation: Assess the final model on the test dataset to measure its performance on unseen data. Analyze results to determine the model&#x2019;s effectiveness in real-world scenarios.</p>
</sec>
<sec id="sec14">
<title>Comparison with non-specialist clinicians</title>
<p>To assess clinical utility, we compared the diagnostic performance of the best-performing model with that of four non-specialist physicians. Each physician independently classified the same 400-image test set without access to clinical metadata or contextual information. The performance of each clinician was evaluated using the same metrics as those used for the deep learning models. McNemar&#x2019;s test was used to statistically compare the misclassification rates between the best model and each non-specialist.</p>
</sec>
<sec id="sec15">
<title>Statistical analysis</title>
<p>Model comparison: Pairwise differences in AUC-ROC were tested using DeLong&#x2019;s method for correlated ROC curves. Significance threshold: <italic>&#x03B1;</italic>&#x202F;=&#x202F;0.05 (two-tailed).</p>
<p>Hypothesis: H0: AUCa&#x202F;=&#x202F;AUCBb vs. H1: AUCa =&#x202F;AUCb.</p>
<p>McNemar&#x2019;s test was applied to compare error rates (FP&#x202F;+&#x202F;FN) between models on the same test set:</p>
<disp-formula id="E7">
<mml:math id="M7">
<mml:msup>
<mml:mi>&#x03C7;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi mathvariant="italic">na</mml:mi>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi mathvariant="italic">pb</mml:mi>
</mml:msub>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mo>&#x2223;</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi mathvariant="italic">na</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mi mathvariant="italic">pb</mml:mi>
</mml:msub>
<mml:mo>&#x2223;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mfrac>
</mml:math>
</disp-formula>
<p>where <italic>F</italic><sub>na</sub>&#x202F;=&#x202F;false negatives of Model a, <italic>F</italic><sub>pb</sub>&#x202F;=&#x202F;false positives of Model b. All statistical analyses were performed using Python 3.10.5.</p>
</sec>
</sec>
<sec sec-type="results" id="sec16">
<title>Results</title>
<sec id="sec17">
<title>Training process analysis</title>
<p>During model training, we conducted a comparative evaluation of several widely used deep learning architectures, including Swin-T, ViT_b_16, ResNet152, ResNet34, Inception_v3, and AlexNet, to analyze their performance in automated wound infection classification. As illustrated in <xref ref-type="fig" rid="fig2">Figure 2</xref>, the models were trained over 100 epochs, during which multiple performance metrics were recorded, including accuracy, sensitivity, specificity, precision, F1-score, and AUC. To prevent overfitting and ensure the retention of the best-performing model, the training process incorporated early stopping and model checkpointing strategies.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Training dynamics and validation metrics of five deep learning models.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Grid of performance plots compares five deep learning models: AlexNet, Inception_v3, ResNet34, ResNet152, and Swin-T. Each row displays loss curve, accuracy, validation AUC, validation F1, validation precision, and validation NPV across training epochs.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec18">
<title>Model performance</title>
<p>Five deep learning models were successfully trained and evaluated on the curated wound infection dataset. All models achieve AUC values exceeding 0.9 (0.92&#x2013;0.95), far exceeding random classifiers (AUC&#x202F;=&#x202F;0.5), indicating exceptional classification discrimination capabilities across all five models (<xref ref-type="fig" rid="fig3">Figure 3</xref>). However, the Swin-T model demonstrated the best overall performance. Specifically, Swin-T achieved an accuracy of 0.9025, AUC-ROC of 0.9546, and F1-score of 0.9042, outperforming traditional CNN architectures including AlexNet, ResNet34, ResNet152, and Inception_v3 across most evaluation metrics (<xref ref-type="table" rid="tab1">Table 1</xref>). After clinical decision threshold optimization via the Youden Index, Swin-T achieved a peak accuracy of 90.50% and an AUC-ROC of 0.9546, maintaining an optimized balance between sensitivity (92.00%) and specificity (89.00%). While the ResNet152 model exhibited a higher raw sensitivity of 95.00% under its optimized threshold (0.1554), this came at the cost of a significantly lower specificity (84.00%) (<xref ref-type="table" rid="tab2">Table 2</xref>). In contrast, Swin-T consistently outperformed ResNet152 in terms of F1-score (0.9064) and overall diagnostic stability, providing a more reliable decision boundary for clinical infection screening.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Receiver operating characteristic (ROC) curves for model comparison.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">ROC curve comparison chart shows five models&#x2019; performance with true positive rate versus false positive rate. Swin-t achieves the highest AUC of 0.95, followed by ResNet34 and Inception_v3 at 0.94, ResNet152 at 0.93, and AlexNet at 0.92.</alt-text>
</graphic>
</fig>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>The performance comparison of various deep learning models evaluated on the clinical wound infection dataset.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Metric</th>
<th align="center" valign="top">AlexNet</th>
<th align="center" valign="top">InceptionV3</th>
<th align="center" valign="top">ResNet34</th>
<th align="center" valign="top">ResNet152</th>
<th align="center" valign="top">Swin-T</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Params (M)</td>
<td align="center" valign="top">11.97</td>
<td align="center" valign="top">21.79</td>
<td align="center" valign="top">21.29</td>
<td align="center" valign="top">58.15</td>
<td align="center" valign="top"><bold>86.75</bold></td>
</tr>
<tr>
<td align="left" valign="top">NVIDIA T4(ms)</td>
<td align="center" valign="top">1.8&#x202F;&#x00B1;&#x202F;0.8</td>
<td align="center" valign="top">17.9&#x202F;&#x00B1;&#x202F;3.0</td>
<td align="center" valign="top">7.48&#x202F;&#x00B1;&#x202F;1.7</td>
<td align="center" valign="top">21.8&#x202F;&#x00B1;&#x202F;3.5</td>
<td align="center" valign="top"><bold>27.7&#x202F;&#x00B1;&#x202F;2.0</bold></td>
</tr>
<tr>
<td align="left" valign="top">CPU (ms)</td>
<td align="center" valign="top">109.3&#x202F;&#x00B1;&#x202F;4.3</td>
<td align="center" valign="top">313.7&#x202F;&#x00B1;&#x202F;7.4</td>
<td align="center" valign="top">80.66&#x202F;&#x00B1;&#x202F;7.5</td>
<td align="center" valign="top">320.6&#x202F;&#x00B1;&#x202F;5.7</td>
<td align="center" valign="top"><bold>486.4&#x202F;&#x00B1;&#x202F;5.2</bold></td>
</tr>
<tr>
<td align="left" valign="top">Accuracy</td>
<td align="center" valign="top">0.8325 [0.7928&#x2013;0.8659]</td>
<td align="center" valign="top">0.8625 [0.8253&#x2013;0.8928]</td>
<td align="center" valign="top">0.8650<break/>[0.8280&#x2013;0.8950]</td>
<td align="center" valign="top">0.8850<break/>[0.8500&#x2013;0.9127]</td>
<td align="center" valign="top"><bold>0.9025</bold>
<break/>
<bold>[0.8695&#x2013;0.9279]</bold></td>
</tr>
<tr>
<td align="left" valign="top">Sensitivity</td>
<td align="center" valign="top">0.8550 [0.7995&#x2013;0.8971]</td>
<td align="center" valign="top">0.8450 [0.7884&#x2013;0.8886]</td>
<td align="center" valign="top">0.8400<break/>[0.7829&#x2013;0.8843]</td>
<td align="center" valign="top">0.8750<break/>[0.8220&#x2013;0.9139]</td>
<td align="center" valign="top"><bold>0.8850 [0.8334&#x2013;0.9221]</bold></td>
</tr>
<tr>
<td align="left" valign="top">Specificity</td>
<td align="center" valign="top">0.8100 [0.7500&#x2013;0.8583]</td>
<td align="center" valign="top">0.8800 [0.8277&#x2013;0.9180]</td>
<td align="center" valign="top">0.8900<break/>[0.8391&#x2013;0.9262]</td>
<td align="center" valign="top">0.8950<break/>[0.8448&#x2013;0.9303]</td>
<td align="center" valign="top"><bold>0.9200 [0.8740&#x2013;0.9502]</bold></td>
</tr>
<tr>
<td align="left" valign="top">PPV</td>
<td align="center" valign="top">0.8182 [0.7603&#x2013;0.8646]</td>
<td align="center" valign="top">0.8756 [0.8216&#x2013;0.9150]</td>
<td align="center" valign="top">0.8842<break/>[0.8309&#x2013;0.9223]</td>
<td align="center" valign="top">0.8929<break/>[0.8418&#x2013;0.9288]</td>
<td align="center" valign="top"><bold>0.9171 [0.8696&#x2013;0.9483]</bold></td>
</tr>
<tr>
<td align="left" valign="top">NPV</td>
<td align="center" valign="top">0.8482 [0.7904&#x2013;0.8922]</td>
<td align="center" valign="top">0.8502 [0.7953&#x2013;0.8924]</td>
<td align="center" valign="top">0.8476<break/>[0.7928&#x2013;0.8899]</td>
<td align="center" valign="top">0.8775<break/>[0.8254&#x2013;0.9156]</td>
<td align="center" valign="top"><bold>0.8889 [0.8388&#x2013;0.9248]</bold></td>
</tr>
<tr>
<td align="left" valign="top">F1 Score</td>
<td align="center" valign="top">0.8286 [0.7863&#x2013;0.8670]</td>
<td align="center" valign="top">0.8649 [0.8241&#x2013;0.8964]</td>
<td align="center" valign="top">0.8683<break/>[0.8338&#x2013;0.9016]</td>
<td align="center" valign="top">0.8861<break/>[0.8513&#x2013;0.9169]</td>
<td align="center" valign="top"><bold>0.9042 [0.8727&#x2013;0.9327]</bold></td>
</tr>
<tr>
<td align="left" valign="top">AUC</td>
<td align="center" valign="top">0.9212 [0.8958&#x2013;0.9424]</td>
<td align="center" valign="top">0.9443 [0.9221&#x2013;0.9643]</td>
<td align="center" valign="top">0.9388<break/>[0.9153&#x2013;0.9605]</td>
<td align="center" valign="top">0.9320<break/>[0.9027&#x2013;0.9560]</td>
<td align="center" valign="top"><bold>0.9546 [0.9324&#x2013;0.9755]</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>The values in brackets represent the 95% confidence intervals. Bold values indicate the performance results of the Swin-T model for comparison with other models.</p>
</table-wrap-foot>
</table-wrap>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Comparison of model performance using default (0.5) and optimized probability thresholds.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model</th>
<th align="left" valign="top">Threshold type</th>
<th align="center" valign="top">Optimal threshold</th>
<th align="center" valign="top">Accuracy (%)</th>
<th align="center" valign="top">Sensitivity (%)</th>
<th align="center" valign="top">Specificity (%)</th>
<th align="center" valign="top">F1-Score</th>
<th align="center" valign="top">AUC (95% CI)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="bottom" rowspan="2">Swin-T</td>
<td align="left" valign="bottom"><bold>Default</bold></td>
<td align="char" valign="bottom" char="."><bold>0.5</bold></td>
<td align="center" valign="bottom"><bold>90.25</bold></td>
<td align="center" valign="bottom"><bold>92</bold></td>
<td align="center" valign="bottom"><bold>88.5</bold></td>
<td align="char" valign="bottom" char="."><bold>0.9042</bold></td>
<td align="center" valign="bottom"><bold>0.9546 (0.93&#x2013;0.97)</bold></td>
</tr>
<tr>
<td align="left" valign="bottom"><bold>Optimized</bold></td>
<td align="char" valign="bottom" char="."><bold>0.5848</bold></td>
<td align="center" valign="bottom"><bold>90.5</bold></td>
<td align="center" valign="bottom"><bold>92</bold></td>
<td align="center" valign="bottom"><bold>0.89</bold></td>
<td align="char" valign="bottom" char="."><bold>0.9064</bold></td>
<td align="center" valign="bottom"><bold>0.9546 (0.93&#x2013;0.97)</bold></td>
</tr>
<tr>
<td align="left" valign="bottom" rowspan="2">ResNet152</td>
<td align="left" valign="bottom">Default</td>
<td align="char" valign="bottom" char=".">0.5</td>
<td align="center" valign="bottom">88.5</td>
<td align="center" valign="bottom">89.5</td>
<td align="center" valign="bottom">87.5</td>
<td align="char" valign="bottom" char=".">0.8861</td>
<td align="center" valign="bottom">0.9320 (0.90&#x2013;0.95)</td>
</tr>
<tr>
<td align="left" valign="bottom">Optimized</td>
<td align="char" valign="bottom" char=".">0.1554</td>
<td align="center" valign="bottom">89.5</td>
<td align="center" valign="bottom">95</td>
<td align="center" valign="bottom">0.84</td>
<td align="char" valign="bottom" char=".">0.9005</td>
<td align="center" valign="bottom">0.9320 (0.90&#x2013;0.95)</td>
</tr>
<tr>
<td align="left" valign="bottom" rowspan="2">ResNet34</td>
<td align="left" valign="bottom">Default</td>
<td align="char" valign="bottom" char=".">0.5</td>
<td align="center" valign="bottom">86.5</td>
<td align="center" valign="bottom">89</td>
<td align="center" valign="bottom">84</td>
<td align="char" valign="bottom" char=".">0.8683</td>
<td align="center" valign="bottom">0.9388 (0.91&#x2013;0.95)</td>
</tr>
<tr>
<td align="left" valign="bottom">Optimized</td>
<td align="char" valign="bottom" char=".">0.5541</td>
<td align="center" valign="bottom">87</td>
<td align="center" valign="bottom">89</td>
<td align="center" valign="bottom">0.85</td>
<td align="char" valign="bottom" char=".">0.8725</td>
<td align="center" valign="bottom">0.9388 (0.91&#x2013;0.95)</td>
</tr>
<tr>
<td align="left" valign="bottom" rowspan="2">Inception_v3</td>
<td align="left" valign="bottom">Default</td>
<td align="char" valign="bottom" char=".">0.5</td>
<td align="center" valign="bottom">86.25</td>
<td align="center" valign="bottom">88</td>
<td align="center" valign="bottom">84.5</td>
<td align="char" valign="bottom" char=".">0.8649</td>
<td align="center" valign="bottom">0.9443 (0.92&#x2013;0.96)</td>
</tr>
<tr>
<td align="left" valign="bottom">Optimized</td>
<td align="char" valign="bottom" char=".">0.5817</td>
<td align="center" valign="bottom">87.75</td>
<td align="center" valign="bottom">87</td>
<td align="center" valign="bottom">0.885</td>
<td align="char" valign="bottom" char=".">0.8766</td>
<td align="center" valign="bottom">0.9443 (0.92&#x2013;0.96)</td>
</tr>
<tr>
<td align="left" valign="bottom" rowspan="2">AlexNet</td>
<td align="left" valign="bottom">Default</td>
<td align="char" valign="bottom" char=".">0.5</td>
<td align="center" valign="bottom">83.25</td>
<td align="center" valign="bottom">81</td>
<td align="center" valign="bottom">85.5</td>
<td align="char" valign="bottom" char=".">0.8286</td>
<td align="center" valign="bottom">0.9212 (0.89&#x2013;0.94)</td>
</tr>
<tr>
<td align="left" valign="bottom">Optimized</td>
<td align="char" valign="bottom" char=".">0.4216</td>
<td align="center" valign="bottom">84.25</td>
<td align="center" valign="bottom">87.5</td>
<td align="center" valign="bottom">0.81</td>
<td align="char" valign="bottom" char=".">0.8475</td>
<td align="center" valign="bottom">0.9213 (0.89&#x2013;0.94)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Optimized thresholds were determined using the Youden Index (J&#x202F;=&#x202F;Sensitivity + Specificity &#x2013; 1) to achieve the optimal balance between sensitivity and specificity. Bold values indicate the performance results of the Swin-T model for comparison with other models.</p>
</table-wrap-foot>
</table-wrap>
<p>To further interpret the diagnostic performance of each model, confusion matrices were generated based on the test set predictions (<xref ref-type="fig" rid="fig4">Figure 4</xref>). For the best-performing model, Swin-T, the confusion matrix revealed a high number of true positives (TP&#x202F;=&#x202F;177) and true negatives (TN&#x202F;=&#x202F;184), indicating strong capability in accurately detecting both infected and non-infected wounds. The number of false positives (FP&#x202F;=&#x202F;16) and false negatives (FN&#x202F;=&#x202F;23) remained relatively low, which supports the high specificity (0.920) and sensitivity (0.885) observed in the metric evaluations (<xref ref-type="fig" rid="fig5">Figure 5</xref>).</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Confusion matrices of five deep learning models&#x2014;AlexNet, ResNet34, InceptionV3, ResNet152, and Swin-T&#x2014;used for the classification of infected versus non-infected wounds.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Five confusion matrices compare model performance for infection detection. Swin-T yields 177 true positives, 184 true negatives, 23 false positives, and 16 false negatives. ResNet152 shows 175 true positives, 179 true negatives, 25 false positives, and 21 false negatives. ResNet34 presents 168 true positives, 178 true negatives, 32 false positives, and 22 false negatives. Inception_V3 reports 169 true positives, 176 true negatives, 31 false positives, and 24 false negatives. Alexnet records 171 true positives, 162 true negatives, 29 false positives, and 38 false negatives. All matrices display true and predicted infection and non-infection labels.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Comparative performance of deep learning models in wound infection detection. Radar plot illustrating the performance of five deep learning architectures&#x2014;AlexNet, ResNet34, InceptionV3, ResNet152, and Swin-T&#x2014;across five evaluation metrics: Accuracy, Sensitivity, Specificity, F1-score, and AUC.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Radar chart compares five models&#x2014;Swin-T, ResNet152, ResNet34, Inception_v3, and AlexNet&#x2014;for wound infection detection using accuracy, sensitivity, specificity, F1-score, and AUC. Swin-T outperforms others overall.</alt-text>
</graphic>
</fig>
<p>The computational efficiency of the models was evaluated across different hardware platforms, as summarized in <xref ref-type="table" rid="tab1">Table 1</xref>. Among the tested architectures, the Swin-T model (approx. 86.75&#x202F;M parameters) achieved a high-performance inference speed of 27.7&#x202F;&#x00B1;&#x202F;2.0&#x202F;ms on an industry-standard NVIDIA T4 GPU, significantly meeting the requirements for real-time clinical applications. Even when deployed on a standard CPU, the Swin-T model maintained a responsive inference time of 486.4&#x202F;&#x00B1;&#x202F;5.2&#x202F;ms. While lightweight models like AlexNet showed faster raw speeds, Swin-T provided a more robust balance between architectural capacity and operational latency, ensuring its feasibility for deployment in diverse clinical environments.</p>
</sec>
<sec id="sec19">
<title>Statistical significance testing</title>
<p>To rigorously assess performance differences, DeLong&#x2019;s test and McNemar&#x2019;s test were conducted between Swin-T and all other models. Statistically significant improvements (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05) were observed between Swin-T and AlexNet for both AUC-ROC (<italic>p</italic>&#x202F;=&#x202F;0.0178) and misclassification rate (McNemar&#x2019;s <italic>p</italic>&#x202F;=&#x202F;0.0013). Although differences in AUC between Swin-T and ResNet34, InceptionV3, and ResNet152 did not reach statistical significance, McNemar&#x2019;s test indicated Swin-T produced significantly fewer classification errors compared to ResNet34 (<italic>p</italic>&#x202F;=&#x202F;0.0328) and InceptionV3 (<italic>p</italic>&#x202F;=&#x202F;0.0304), suggesting a consistent advantage in practical diagnostic performance (<xref ref-type="table" rid="tab3">Table 3</xref>).</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Statistical significance tests comparing Swin-T with other models.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Comparison</th>
<th align="center" valign="top">&#x0394;AUC</th>
<th align="center" valign="top">DeLong p</th>
<th align="center" valign="top">DeLong<break/>Sig.</th>
<th align="center" valign="top">McNemar <italic>&#x03C7;</italic><sup>2</sup></th>
<th align="center" valign="top">McNemar p</th>
<th align="center" valign="top">McNemar Sig.</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">AlexNet</td>
<td align="char" valign="top" char=".">&#x2212;0.0334</td>
<td align="char" valign="top" char="."><bold>0.0178</bold></td>
<td align="center" valign="top">&#x2714;</td>
<td align="char" valign="top" char=".">10.41</td>
<td align="char" valign="top" char="."><bold>0.0013</bold></td>
<td align="center" valign="top">&#x2714;</td>
</tr>
<tr>
<td align="left" valign="top">ResNet34</td>
<td align="char" valign="top" char=".">&#x2212;0.0158</td>
<td align="char" valign="top" char=".">0.1767</td>
<td align="center" valign="top">&#x2013;</td>
<td align="char" valign="top" char=".">4.56</td>
<td align="char" valign="top" char="."><bold>0.0328</bold></td>
<td align="center" valign="top">&#x2714;</td>
</tr>
<tr>
<td align="left" valign="top">InceptionV3</td>
<td align="char" valign="top" char=".">&#x2212;0.0103</td>
<td align="char" valign="top" char=".">0.3679</td>
<td align="center" valign="top">&#x2013;</td>
<td align="char" valign="top" char=".">4.69</td>
<td align="char" valign="top" char="."><bold>0.0304</bold></td>
<td align="center" valign="top">&#x2714;</td>
</tr>
<tr>
<td align="left" valign="top">ResNet152</td>
<td align="char" valign="top" char=".">&#x2212;0.0226</td>
<td align="char" valign="top" char=".">0.0993</td>
<td align="center" valign="top">&#x2013;</td>
<td align="char" valign="top" char=".">0.80</td>
<td align="char" valign="top" char=".">0.3711</td>
<td align="center" valign="top">&#x2013;</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values indicate a statistically significant difference (<italic>P</italic> &#x003C; 0.05) when compared with the Swin-T model.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="sec20">
<title>Comparison with non-specialist clinicians</title>
<p>The Swin-T model&#x2019;s diagnostic capability was further benchmarked against four non-specialist clinicians. Swin-T consistently outperformed all four clinicians in accuracy (90.25% vs. 73.00&#x2013;82.25%) and F1-score (0.9042 vs. 0.7003&#x2013;0.8412) (<xref ref-type="table" rid="tab4">Table 4</xref> and <xref ref-type="fig" rid="fig6">Figure 6</xref>).</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Performance comparison between Swin-T and non-specialists.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Metric</th>
<th align="center" valign="top">Swin-T</th>
<th align="center" valign="top">Non-specialist 1</th>
<th align="center" valign="top">Non-specialist 2</th>
<th align="center" valign="top">Non-specialist 3</th>
<th align="center" valign="top">Non-specialist 4</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Accuracy</td>
<td align="center" valign="top">0.9025</td>
<td align="char" valign="top" char=".">0.7475</td>
<td align="char" valign="top" char=".">0.8075</td>
<td align="char" valign="top" char=".">0.8225</td>
<td align="char" valign="top" char=".">0.7300</td>
</tr>
<tr>
<td align="left" valign="top">Sensitivity</td>
<td align="center" valign="top">0.885</td>
<td align="char" valign="top" char=".">0.590</td>
<td align="char" valign="top" char=".">0.770</td>
<td align="char" valign="top" char=".">0.940</td>
<td align="char" valign="top" char=".">0.870</td>
</tr>
<tr>
<td align="left" valign="top">Specificity</td>
<td align="center" valign="top">0.920</td>
<td align="char" valign="top" char=".">0.905</td>
<td align="char" valign="top" char=".">0.845</td>
<td align="char" valign="top" char=".">0.705</td>
<td align="char" valign="top" char=".">0.590</td>
</tr>
<tr>
<td align="left" valign="top">PPV</td>
<td align="center" valign="top">0.9171</td>
<td align="char" valign="top" char=".">0.8613</td>
<td align="char" valign="top" char=".">0.8324</td>
<td align="char" valign="top" char=".">0.7611</td>
<td align="char" valign="top" char=".">0.6797</td>
</tr>
<tr>
<td align="left" valign="top">NPV</td>
<td align="center" valign="top">0.8889</td>
<td align="char" valign="top" char=".">0.6882</td>
<td align="char" valign="top" char=".">0.7860</td>
<td align="char" valign="top" char=".">0.9216</td>
<td align="char" valign="top" char=".">0.8194</td>
</tr>
<tr>
<td align="left" valign="top">F1-score</td>
<td align="center" valign="top">0.9008</td>
<td align="char" valign="top" char=".">0.7003</td>
<td align="char" valign="top" char=".">0.8000</td>
<td align="char" valign="top" char=".">0.8412</td>
<td align="char" valign="top" char=".">0.7632</td>
</tr>
<tr>
<td align="left" valign="top">McNemar<break/><italic>p</italic>-value</td>
<td align="center" valign="top"><bold>&#x2013;&#x2013;</bold></td>
<td align="char" valign="top" char="."><bold>&#x003C;0.0001</bold></td>
<td align="char" valign="top" char="."><bold>&#x003C;0.0001</bold></td>
<td align="char" valign="top" char="."><bold>0.0010</bold></td>
<td align="char" valign="top" char="."><bold>&#x003C;0.0001</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values indicate a statistically significant difference (<italic>P</italic> &#x003C; 0.05) when compared with the Swin-T model.</p>
</table-wrap-foot>
</table-wrap>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Performance comparison between the Swin-T model and non-specialist clinicians in wound infection detection. Radar chart illustrating the performance metrics&#x2014;accuracy, sensitivity, specificity, and F1-score.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g006.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Radar chart comparing performance metrics&#x2014;accuracy, sensitivity, specificity, and F1-score&#x2014;for Swin-T and four non-specialists. Swin-T consistently achieves higher values across all metrics than the non-specialists.</alt-text>
</graphic>
</fig>
<p>To enable a direct comparison with the Swin-T model, confusion matrices were generated for each non-specialist clinician using the same test set predictions (<xref ref-type="fig" rid="fig7">Figure 7</xref>). Compared with Swin-T, confusion matrices of non-specialist clinicians revealed notable performance trade-offs. For example, Non-specialist 3 exhibited the highest sensitivity (94.0%) but also had a markedly higher number of false positives (FP&#x202F;=&#x202F;82), leading to decreased specificity and increased unnecessary alarms. Non-specialist 1 and 4, while more conservative, missed more actual infections (higher FN), resulting in lower sensitivity. These patterns suggest a common challenge among non-specialist clinician in maintaining a balanced classification threshold, often favoring either sensitivity or specificity but not both. McNemar&#x2019;s test revealed statistically significant differences (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.01) between Swin-T and all clinicians, confirming its superior diagnostic precision (<xref ref-type="table" rid="tab4">Table 4</xref>).</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Confusion matrices of non-specialist clinicians&#x2014;used for the classification of infected versus non-infected wounds.</p>
</caption>
<graphic xlink:href="fpubh-14-1772514-g007.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Four color-coded confusion matrices labeled Non-specialist 1 to Non-specialist 4 display predicted versus true labels for Infection and Non-Inf classes. Each matrix presents true and false positive and negative counts, highlighting variation in classification accuracy among the four non-specialists.</alt-text>
</graphic>
</fig>
<p>These results demonstrate the potential of Transformer-based deep learning models as reliable, expert-level diagnostic tools for wound infection detection. The Swin-T model achieved high accuracy across diverse wound types, suggesting clinical viability as an assistive tool for early infection detection, particularly in resource-limited or primary care settings.</p>
</sec>
</sec>
<sec sec-type="discussion" id="sec21">
<title>Discussion</title>
<p>This study proposes a robust and clinically relevant deep learning framework for the automatic detection of wound infections, in which the Swin-T model outperforms traditional CNN architectures. Swin-T demonstrated superior performance in terms of accuracy (90.25%), F1 score (0.9042), and AUC-ROC (0.9546), indicating that Transformer-based architectures may offer significant advantages in handling complex wound image classification tasks.</p>
<p>A key advantage of the Swin-T model lies in its ability to capture both local and global features through its hierarchical structure and sliding window self-attention mechanism (<xref ref-type="bibr" rid="ref17">17</xref>). Unlike traditional CNNs that rely on fixed receptive fields, the Swin-T architecture can dynamically model long-range dependencies, which is particularly useful for interpreting diffuse infection characteristics such as periwound erythema, tissue necrosis, biofilm presence, or periwound inflammation. These complex visual cues are often subtle and may be difficult for non-specialist clinical personnel to accurately identify, highlighting the potential of Swin-T as a decision support tool in primary care settings.</p>
<p>The Swin-T model exhibited significant superiority over traditional CNN architectures (AlexNet, ResNet34, and Inception_v3) and notably outperformed non-specialist clinicians in diagnostic accuracy. Specifically, after clinical decision threshold optimization via the Youden Index, Swin-T achieved a robust balance between sensitivity (92.00%) and specificity (89.00%). This equilibrium is critical in the context of antimicrobial stewardship; while traditional models like ResNet152 or certain non-specialist clinicians may achieve higher raw sensitivity, they often do so by significantly sacrificing specificity. Such an imbalance leads to an influx of false positives, which clinically translates to the over-prescription of antibiotics and the potential exacerbation of antimicrobial resistance. By maintaining high specificity alongside strong sensitivity, the Swin-T model provides a more judicious decision-support tool, ensuring that infections are captured accurately while minimizing unnecessary medical interventions in primary care settings.</p>
<p>Beyond diagnostic accuracy, the operational efficiency of a model is a decisive factor for its integration into fast-paced clinical environments. Our results (<xref ref-type="table" rid="tab1">Table 1</xref>) demonstrate that the Swin-T model, despite having a larger parameter capacity (86.75&#x202F;M), maintains a highly competitive inference latency of 27.7&#x202F;&#x00B1;&#x202F;2.0&#x202F;ms on an NVIDIA T4 GPU. This sub-30&#x202F;ms performance is significant as it supports real-time diagnostic feedback, which is essential for high-throughput screening in outpatient departments. Furthermore, the model&#x2019;s stable performance on a standard CPU (486.4&#x202F;&#x00B1;&#x202F;5.2&#x202F;ms) underscores its hardware agnosticism, suggesting that the system could be reliably deployed in primary care settings or community clinics that lack dedicated GPU servers. By achieving an optimal balance between architectural depth and computational speed, this framework provides a practical solution for bridging the gap between advanced deep learning research and bedside clinical application.</p>
<p>However, this study has some limitations. First, this research was conducted at a single center, which may affect the immediate generalizability of the findings across diverse clinical environments. Nevertheless, this study establishes a foundational framework for broad-spectrum wound infection screening. While the dataset (<italic>n</italic> =&#x202F;4,000) is substantial for clinical research, it is considered moderate from a deep learning perspective. To address this, we implemented data augmentation to improve model robustness. Moving forward, we have established plans for future multi-center collaborations to further optimize and validate the system across diverse geographical regions and clinical settings. Second, infection status in this study was determined based on expert consensus, and although guided by the IWII-WIC framework, some degree of subjective bias may still exist. Lastly, the models used in this study were trained and validated on static images. Future work should consider incorporating time-series images or video data to better capture dynamic wound changes and inflammation patterns.</p>
<p>In conclusion, this study demonstrates the feasibility and effectiveness of deep learning for automated wound infection screening using clinical images. The proposed models, including the Swin-T and conventional CNNs, achieved high diagnostic performance and reduced variability compared with non-specialist clinicians. By providing a rapid, objective, and scalable solution, these AI-driven tools have the potential to support early infection recognition, streamline clinical workflows, and improve antimicrobial stewardship. Such approaches are particularly valuable in resource-limited and decentralized healthcare settings, where they can empower non-specialist clinicians and patients, ultimately enhancing the accessibility, consistency, and quality of wound care.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec22">
<title>Data availability statement</title>
<p>The data analyzed in this study is subject to the following licenses/restrictions: the dataset consists of anonymized clinical wound images collected from hospital records and is subject to patient privacy regulations. Access is restricted and cannot be publicly shared. Use of the dataset requires approval from the corresponding hospital and compliance with local ethical and legal requirements. Requests to access the dataset can be made by contacting the corresponding author, Chao Wang, at <email xlink:href="mailto:wangchaoxy@126.com">wangchaoxy@126.com</email>. Access will require approval from the Ethics Committee of Chengdu Second People&#x2019;s Hospital and compliance with all local ethical and legal requirements.</p>
</sec>
<sec sec-type="ethics-statement" id="sec23">
<title>Ethics statement</title>
<p>The Ethics Committee of Chengdu Second People&#x2019;s Hospital waived informed consent (approval no. [KY]PJ2025216). The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants&#x2019; legal guardians/next of kin because the study used fully anonymized retrospective clinical wound images, which posed no risk to patient privacy or safety, and the protocol was approved by the Ethics Committee.</p>
</sec>
<sec sec-type="author-contributions" id="sec24">
<title>Author contributions</title>
<p>CW: Conceptualization, Data curation, Methodology, Supervision, Writing &#x2013; original draft, Writing &#x2013; review &#x0026; editing. HW: Data curation, Formal analysis, Investigation, Methodology, Supervision, Writing &#x2013; review &#x0026; editing. JH: Data curation, Formal analysis, Investigation, Methodology, Supervision, Writing &#x2013; review &#x0026; editing. ZH: Data curation, Formal analysis, Investigation, Methodology, Supervision, Writing &#x2013; review &#x0026; editing. YY: Data curation, Formal analysis, Methodology, Writing &#x2013; review &#x0026; editing. ZT: Data curation, Formal analysis, Methodology, Writing &#x2013; review &#x0026; editing. DL: Data curation, Formal analysis, Methodology, Supervision, Writing &#x2013; review &#x0026; editing. LW: Data curation, Formal analysis, Methodology, Supervision, Writing &#x2013; review &#x0026; editing.</p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>We are grateful to Chang Peng, Li Ding, Min Zhen, Limei Song, Qi Xu, and Xiaoni Geng from Department of Burn and Plastic Surgery, Chengdu Second People&#x2019;s Hospital, for providing assistance in data collection to this study.</p>
</ack>
<sec sec-type="COI-statement" id="sec25">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec26">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec27">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><label>1.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Uberoi</surname><given-names>A</given-names></name> <name><surname>McCready-Vangi</surname><given-names>A</given-names></name> <name><surname>Grice</surname><given-names>EA</given-names></name></person-group>. <article-title>The wound microbiota: microbial mechanisms of impaired wound healing and infection</article-title>. <source>Nat Rev Microbiol</source>. (<year>2024</year>) <volume>22</volume>:<fpage>507</fpage>&#x2013;<lpage>21</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41579-024-01035-z</pub-id></mixed-citation></ref>
<ref id="ref2"><label>2.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>S</given-names></name> <name><surname>He</surname><given-names>W</given-names></name> <name><surname>Dong</surname><given-names>J</given-names></name> <name><surname>Chan</surname><given-names>YK</given-names></name> <name><surname>Lai</surname><given-names>S</given-names></name> <name><surname>Deng</surname><given-names>Y</given-names></name></person-group>. <article-title>Tailoring versatile nanoheterojunction-incorporated hydrogel dressing for wound bacterial biofilm infection theranostics</article-title>. <source>ACS Nano</source>. (<year>2025</year>) <volume>19</volume>:<fpage>10922</fpage>&#x2013;<lpage>42</lpage>. doi: <pub-id pub-id-type="doi">10.1021/acsnano.4c15743</pub-id>, <pub-id pub-id-type="pmid">40071724</pub-id></mixed-citation></ref>
<ref id="ref3"><label>3.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Polk</surname><given-names>HC</given-names></name> <name><surname>Simpson</surname><given-names>CJ</given-names></name> <name><surname>Simmons</surname><given-names>BP</given-names></name> <name><surname>Alexander</surname><given-names>JW</given-names></name></person-group>. <article-title>Guidelines for prevention of surgical wound infection</article-title>. <source>Arch Surg</source>. (<year>1983</year>) <volume>118</volume>:<fpage>1213</fpage>&#x2013;<lpage>7</lpage>. doi: <pub-id pub-id-type="doi">10.1001/archsurg.1983.01390100075019</pub-id></mixed-citation></ref>
<ref id="ref4"><label>4.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Swanson</surname><given-names>T</given-names></name> <name><surname>Ousey</surname><given-names>K</given-names></name> <name><surname>Haesler</surname><given-names>E</given-names></name> <name><surname>Bjarnsholt</surname><given-names>T</given-names></name> <name><surname>Carville</surname><given-names>K</given-names></name> <name><surname>Idensohn</surname><given-names>P</given-names></name> <etal/></person-group>. <article-title>IWII wound infection in clinical practice consensus document: 2022 update</article-title>. <source>J Wound Care</source>. (<year>2022</year>) <volume>31</volume>:<fpage>S10</fpage>&#x2013;<lpage>21</lpage>. doi: <pub-id pub-id-type="doi">10.12968/jowc.2022.31.Sup12.S10</pub-id>, <pub-id pub-id-type="pmid">36475844</pub-id></mixed-citation></ref>
<ref id="ref5"><label>5.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Snyder</surname><given-names>RJ</given-names></name> <name><surname>Fife</surname><given-names>C</given-names></name> <name><surname>Moore</surname><given-names>Z</given-names></name></person-group>. <article-title>Components and quality measures of DIME (devitalized tissue, infection/inflammation, moisture balance, and edge preparation) in wound care</article-title>. <source>Adv Skin Wound Care</source>. (<year>2016</year>) <volume>29</volume>:<fpage>205</fpage>&#x2013;<lpage>15</lpage>. doi: <pub-id pub-id-type="doi">10.1097/01.ASW.0000482354.01988.b4</pub-id>, <pub-id pub-id-type="pmid">27089149</pub-id></mixed-citation></ref>
<ref id="ref6"><label>6.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Litjens</surname><given-names>G</given-names></name> <name><surname>Kooi</surname><given-names>T</given-names></name> <name><surname>Bejnordi</surname><given-names>BE</given-names></name> <name><surname>Setio</surname><given-names>AAA</given-names></name> <name><surname>Ciompi</surname><given-names>F</given-names></name> <name><surname>Ghafoorian</surname><given-names>M</given-names></name> <etal/></person-group>. <article-title>A survey on deep learning in medical image analysis</article-title>. <source>Med Image Anal</source>. (<year>2017</year>) <volume>42</volume>:<fpage>60</fpage>&#x2013;<lpage>88</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.media.2017.07.005</pub-id>, <pub-id pub-id-type="pmid">28778026</pub-id></mixed-citation></ref>
<ref id="ref7"><label>7.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname><given-names>M</given-names></name> <name><surname>Pi</surname><given-names>Y</given-names></name> <name><surname>Jiang</surname><given-names>Z</given-names></name> <name><surname>Wu</surname><given-names>Y</given-names></name> <name><surname>Bu</surname><given-names>H</given-names></name> <name><surname>Bao</surname><given-names>J</given-names></name> <etal/></person-group>. <article-title>Application of deep learning to identify ductal carcinoma in situ and microinvasion of the breast using ultrasound imaging</article-title>. <source>Quant Imaging Med Surg</source>. (<year>2022</year>) <volume>12</volume>:<fpage>4633</fpage>&#x2013;<lpage>46</lpage>. doi: <pub-id pub-id-type="doi">10.21037/qims-22-46</pub-id>, <pub-id pub-id-type="pmid">36060588</pub-id></mixed-citation></ref>
<ref id="ref8"><label>8.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yi&#x011F;itcan</surname><given-names>&#x00C7;</given-names></name></person-group>. <article-title>Deep learning for automated breast cancer detection in ultrasound: a comparative study of four CNN architectures</article-title>. <source>AIAPP</source>. (<year>2025</year>) <volume>1</volume>:<fpage>13</fpage>&#x2013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.69882/adba.ai.2025073</pub-id></mixed-citation></ref>
<ref id="ref9"><label>9.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alpsalaz</surname><given-names>SD</given-names></name> <name><surname>Aslan</surname><given-names>E</given-names></name> <name><surname>&#x00D6;z&#x00FC;pak</surname><given-names>Y</given-names></name> <name><surname>Alpsalaz</surname><given-names>F</given-names></name> <name><surname>Uzel</surname><given-names>H</given-names></name> <name><surname>Bereznychenko</surname><given-names>V</given-names></name></person-group>. <article-title>Hybrid deep learning with attention fusion for enhanced colon cancer detection</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>:<fpage>45583</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-025-29447-8</pub-id>, <pub-id pub-id-type="pmid">41315603</pub-id></mixed-citation></ref>
<ref id="ref10"><label>10.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aslan</surname><given-names>E</given-names></name> <name><surname>Demirtas Alpsalaz</surname><given-names>S</given-names></name> <name><surname>&#x00D6;z&#x00FC;pak</surname><given-names>Y</given-names></name> <name><surname>Alpsalaz</surname><given-names>F</given-names></name> <name><surname>Uzel</surname><given-names>H</given-names></name></person-group>. <article-title>Alzheimer&#x2019;s classification with a MaxViT-based deep learning model using magnetic resonance imaging</article-title>. <source>JASTT</source>. (<year>2025</year>) <volume>6</volume>, 316&#x2013;27. doi: <pub-id pub-id-type="doi">10.38094/jastt62453</pub-id></mixed-citation></ref>
<ref id="ref11"><label>11.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aslan</surname><given-names>E</given-names></name> <name><surname>&#x00D6;z&#x00FC;pak</surname><given-names>Y</given-names></name></person-group>. <article-title>Comparison of machine learning algorithms for automatic prediction of Alzheimer disease</article-title>. <source>J Chin Med Assoc</source>. (<year>2025</year>) <volume>88</volume>:<fpage>98</fpage>&#x2013;<lpage>107</lpage>. doi: <pub-id pub-id-type="doi">10.1097/JCMA.0000000000001188</pub-id>, <pub-id pub-id-type="pmid">39965789</pub-id></mixed-citation></ref>
<ref id="ref12"><label>12.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>H</given-names></name> <name><surname>Unberath</surname><given-names>M</given-names></name> <name><surname>Dreizin</surname><given-names>D</given-names></name></person-group>. <article-title>Toward automated interpretable AAST grading for blunt splenic injury</article-title>. <source>Emerg Radiol</source>. (<year>2023</year>) <volume>30</volume>:<fpage>41</fpage>&#x2013;<lpage>50</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10140-022-02099-1</pub-id>, <pub-id pub-id-type="pmid">36371579</pub-id></mixed-citation></ref>
<ref id="ref13"><label>13.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Al Sadig</surname><given-names>M</given-names></name> <name><surname>Alsamri</surname><given-names>J</given-names></name> <name><surname>Alharbi</surname><given-names>NH</given-names></name> <name><surname>Asiri</surname><given-names>MM</given-names></name> <name><surname>Ghaleb</surname><given-names>M</given-names></name> <name><surname>Alhashmi</surname><given-names>AA</given-names></name> <etal/></person-group>. <article-title>Investigating blood cell images for enhanced hematologic disorder detection using multi-scale feature learning with a hybrid deep learning model</article-title>. <source>Sci Rep</source>. (<year>2025</year>) <volume>15</volume>:<fpage>41490</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-025-21412-9</pub-id>, <pub-id pub-id-type="pmid">41276546</pub-id></mixed-citation></ref>
<ref id="ref14"><label>14.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yi&#x011F;itcan</surname><given-names>&#x00C7;</given-names></name></person-group>. <article-title>Machine learning approaches for enhanced diagnosis of hematological disorders</article-title>. <source>CSAI</source>. (<year>2025</year>) <volume>1</volume>:<fpage>8</fpage>&#x2013;<lpage>14</lpage>. doi: <pub-id pub-id-type="doi">10.69882/adba.csai.2025072</pub-id></mixed-citation></ref>
<ref id="ref15"><label>15.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Topol</surname><given-names>EJ</given-names></name></person-group>. <article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title>. <source>Nat Med</source>. (<year>2019</year>) <volume>25</volume>:<fpage>44</fpage>&#x2013;<lpage>56</lpage>. doi: <pub-id pub-id-type="doi">10.1038/s41591-018-0300-7</pub-id>, <pub-id pub-id-type="pmid">30617339</pub-id></mixed-citation></ref>
<ref id="ref16"><label>16.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sumithra</surname><given-names>MG</given-names></name> <name><surname>Venkatesan</surname><given-names>C</given-names></name></person-group>. <article-title>SwinDFU-net: deep learning transformer network for infection identification in diabetic foot ulcer</article-title>. <source>Technol Health Care</source>. (<year>2025</year>) <volume>33</volume>:<fpage>39269872</fpage>:<fpage>601</fpage>&#x2013;<lpage>18</lpage>. doi: <pub-id pub-id-type="doi">10.3233/THC-241444</pub-id></mixed-citation></ref>
<ref id="ref17"><label>17.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chlap</surname><given-names>P</given-names></name> <name><surname>Min</surname><given-names>H</given-names></name> <name><surname>Vandenberg</surname><given-names>N</given-names></name> <name><surname>Dowling</surname><given-names>J</given-names></name> <name><surname>Holloway</surname><given-names>L</given-names></name> <name><surname>Haworth</surname><given-names>A</given-names></name></person-group>. <article-title>A review of medical image data augmentation techniques for deep learning applications</article-title>. <source>J Med Imaging Radiat Oncol</source>. (<year>2021</year>) <volume>65</volume>:<fpage>545</fpage>&#x2013;<lpage>63</lpage>. doi: <pub-id pub-id-type="doi">10.1111/1754-9485.13261</pub-id></mixed-citation></ref>
<ref id="ref18"><label>18.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname><given-names>YH</given-names></name> <name><surname>Yi</surname><given-names>JY</given-names></name> <name><surname>Noh</surname><given-names>Y</given-names></name> <name><surname>Jang</surname><given-names>H</given-names></name> <name><surname>Seo</surname><given-names>SW</given-names></name> <name><surname>Na</surname><given-names>DL</given-names></name> <etal/></person-group>. <article-title>On the reliability of deep learning-based classification for Alzheimer's disease: multi-cohorts, multi-vendors, multi-protocols, and head-to-head validation</article-title>. <source>Front Neurosci</source>. (<year>2022</year>) <volume>16</volume>:<fpage>851871</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fnins.2022.851871</pub-id></mixed-citation></ref>
<ref id="ref19"><label>19.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alohali</surname><given-names>MA</given-names></name> <name><surname>El-Rashidy</surname><given-names>N</given-names></name> <name><surname>Alaklabi</surname><given-names>S</given-names></name> <name><surname>Elmannai</surname><given-names>H</given-names></name> <name><surname>Alharbi</surname><given-names>S</given-names></name> <name><surname>Saleh</surname><given-names>H</given-names></name></person-group>. <article-title>Swin-GA-RF: genetic algorithm-based Swin transformer and random forest for enhancing cervical cancer classification</article-title>. <source>Front Oncol</source>. (<year>2024</year>) <volume>14</volume>:<fpage>1392301</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fonc.2024.1392301</pub-id>, <pub-id pub-id-type="pmid">39099689</pub-id></mixed-citation></ref>
<ref id="ref20"><label>20.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>B</given-names></name> <name><surname>Sun</surname><given-names>R</given-names></name> <name><surname>Yang</surname><given-names>X</given-names></name> <name><surname>Niu</surname><given-names>B</given-names></name> <name><surname>Zhang</surname><given-names>T</given-names></name> <name><surname>Zhao</surname><given-names>Y</given-names></name> <etal/></person-group>. <article-title>Recognition of rare microfossils using transfer learning and deep residual networks</article-title>. <source>Biology</source>. (<year>2022</year>) <volume>12</volume>:<fpage>16</fpage>. doi: <pub-id pub-id-type="doi">10.3390/biology12010016</pub-id>, <pub-id pub-id-type="pmid">36671708</pub-id></mixed-citation></ref>
<ref id="ref21"><label>21.</label><mixed-citation publication-type="confproc"><person-group person-group-type="author"><name><surname>Szegedy</surname><given-names>C</given-names></name> <name><surname>Vanhoucke</surname><given-names>V</given-names></name> <name><surname>Ioffe</surname><given-names>S</given-names></name> <name><surname>Shlens</surname><given-names>J</given-names></name> <name><surname>Wojna</surname><given-names>Z</given-names></name></person-group>. <article-title>Rethinking the inception architecture for computer vision</article-title>, in <conf-name>Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR)</conf-name>, <fpage>2818</fpage>&#x2013;<lpage>2826</lpage>, (<year>2016</year>).</mixed-citation></ref>
<ref id="ref22"><label>22.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Saboo</surname><given-names>YS</given-names></name> <name><surname>Kapse</surname><given-names>S</given-names></name> <name><surname>Prasanna</surname><given-names>P</given-names></name></person-group>. <article-title>Convolutional neural networks (CNNs) for pneumonia classification on pediatric chest radiographs</article-title>. <source>Cureus</source>. (<year>2023</year>) <volume>15</volume>:<fpage>e44130</fpage>. doi: <pub-id pub-id-type="doi">10.7759/cureus.44130</pub-id>, <pub-id pub-id-type="pmid">37753018</pub-id></mixed-citation></ref>
<ref id="ref23"><label>23.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Dosovitskiy</surname><given-names>A</given-names></name> <name><surname>Beyer</surname><given-names>L</given-names></name> <name><surname>Kolesnikov</surname><given-names>A</given-names></name> <name><surname>Weissenborn</surname><given-names>D</given-names></name> <name><surname>Zhai</surname><given-names>X</given-names></name> <name><surname>Unterthiner</surname><given-names>T</given-names></name> <etal/>.</person-group> An image is worth 16&#x00D7;16 words: transformers for image recognition at scale, [Epubh ahead of preprint]. doi: <pub-id pub-id-type="doi">10.48550/arXiv.2010.11929</pub-id>, (<year>2020</year>).</mixed-citation></ref>
<ref id="ref24"><label>24.</label><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Howard</surname><given-names>AG</given-names></name> <name><surname>Zhu</surname><given-names>M</given-names></name> <name><surname>Chen</surname><given-names>B</given-names></name> <name><surname>Kalenichenko</surname><given-names>D</given-names></name> <name><surname>Wang</surname><given-names>W</given-names></name> <name><surname>Weyand</surname><given-names>T</given-names></name> <etal/></person-group>. <article-title>MobileNets: efficient CNNs for mobile vision applications</article-title>, [Epubh ahead of preprint]. doi: <pub-id pub-id-type="doi">10.48550/arXiv.1704.04861</pub-id>, (<year>2017</year>).</mixed-citation></ref>
<ref id="ref25"><label>25.</label><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Park</surname><given-names>SH</given-names></name> <name><surname>Goo</surname><given-names>JM</given-names></name> <name><surname>Jo</surname><given-names>CH</given-names></name></person-group>. <article-title>Receiver operating characteristic (ROC) curve: practical review for radiologists</article-title>. <source>Korean J Radiol</source>. (<year>2004</year>) <volume>5</volume>:<fpage>11</fpage>&#x2013;<lpage>8</lpage>. doi: <pub-id pub-id-type="doi">10.3348/kjr.2004.5.1.11</pub-id>, <pub-id pub-id-type="pmid">15064554</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3066053/overview">Ishak Pacal</ext-link>, I&#x011F;d&#x0131;r &#x00DC;niversitesi, T&#x00FC;rkiye</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3081802/overview">Emrah Aslan</ext-link>, Mardin Artuklu University, T&#x00FC;rkiye</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3347171/overview">Davut Ari</ext-link>, Bitlis Eren University, T&#x00FC;rkiye</p>
</fn>
</fn-group>
</back>
</article>