<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Oncol.</journal-id>
<journal-title>Frontiers in Oncology</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Oncol.</abbrev-journal-title>
<issn pub-type="epub">2234-943X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fonc.2025.1598289</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Oncology</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>AI-augmented pathology: the experience of transfer learning and intra-domain data diversity in breast cancer metastasis detection</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Cossio</surname>
<given-names>Manuel</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3012749/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Wiedemann</surname>
<given-names>Nina</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Sanfeliu Torres</surname>
<given-names>Esther</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Barnadas Sole</surname>
<given-names>Esther</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/3014105/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Igual</surname>
<given-names>Laura</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2041125/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Department of Mathematics and Computer Science, Universitat de Barcelona</institution>, <addr-line>Barcelona</addr-line>, <country>Spain</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Institute of Cartography and Geoinformation, ETH</institution>, <addr-line>Z&#xfc;rich, Z&#xfc;rich</addr-line>, <country>Switzerland</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Biomedical Diagnostic Center, Clinic Hospital</institution>, <addr-line>Barcelona</addr-line>, <country>Spain</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Mohamed Shehata, University of Louisville, United States</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Jorge Morales-Montor, National Autonomous University of Mexico, Mexico</p>
<p>Mohamed Elshennawy, Theodor Bilharz Research Institute, Egypt</p>
<p>Hossam Hawash, Zagazig University, Egypt</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Manuel Cossio, <email xlink:href="mailto:manuel.cossio@ub.edu">manuel.cossio@ub.edu</email>
</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>11</day>
<month>06</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>15</volume>
<elocation-id>1598289</elocation-id>
<history>
<date date-type="received">
<day>22</day>
<month>03</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>29</day>
<month>04</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Cossio, Wiedemann, Sanfeliu Torres, Barnadas Sole and Igual</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Cossio, Wiedemann, Sanfeliu Torres, Barnadas Sole and Igual</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<sec>
<title>Background</title>
<p>Metastatic detection in sentinel lymph nodes remains a crucial prognostic factor in breast cancer management, with accurate and timely diagnosis directly impacting treatment decisions. While traditional histopathological assessment relies on microscopic examination of stained tissues, the digitization of slides as whole-slide images (WSI) has enabled the development of computer-aided diagnostic systems. These automated approaches offer potential improvements in detection consistency and efficiency compared to conventional methods.</p>
</sec>
<sec>
<title>Results</title>
<p>This study leverages transfer learning on hematoxylin and eosin (HE) WSIs to achieve computationally efficient metastasis detection without compromising accuracy. We propose an approach for generating segmentation masks by transferring spatial annotations from immunohistochemistry (IHC) WSIs to corresponding H&amp;E slides. Using these masks, four distinct datasets were constructed to fine-tune a pretrained ResNet50 model across eight different configurations, incorporating varied dataset combinations and data augmentation techniques. To enhance interpretability, we developed a visualization tool that employs color-coded probability maps to highlight tumor regions alongside their prediction confidence. Our experiments demonstrated that integrating an external dataset (Camelyon16) during training significantly improved model performance, surpassing the benefits of data augmentation alone. The optimal model, trained on both external and local data, achieved an accuracy and F1-score of 0.98, outperforming existing state-of-the-art methods.</p>
</sec>
<sec>
<title>Conclusions</title>
<p>This study demonstrates that transfer learning architectures, when enhanced with multi-source data integration and interpretability frameworks, can significantly improve metastatic detection in whole slide imaging. Our methodology achieves diagnostic performance comparable to gold-standard techniques while dramatically accelerating analytical workflows. The synergistic combination of external dataset incorporation and probabilistic visualization outputs provides a clinically actionable solution that maintains both computational efficiency and pathological interpretability.</p>
</sec>
</abstract>
<kwd-group>
<kwd>sentinel lymph node</kwd>
<kwd>breast cancer</kwd>
<kwd>metastasis detection</kwd>
<kwd>digital pathology</kwd>
<kwd>transfer learning</kwd>
</kwd-group>
<counts>
<fig-count count="6"/>
<table-count count="6"/>
<equation-count count="5"/>
<ref-count count="61"/>
<page-count count="15"/>
<word-count count="8473"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Breast Cancer</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<label>1</label>
<title>Background</title>
<p>Breast cancer is the most commonly diagnosed cancer among women worldwide, posing a significant global health burden (<xref ref-type="bibr" rid="B1">1</xref>). The prognosis of breast cancer patients is closely tied to the stage of disease progression, with survival rates declining sharply upon the development of distant metastases. As a result, early detection is critical to improving therapeutic outcomes and patient survival (<xref ref-type="bibr" rid="B2">2</xref>, <xref ref-type="bibr" rid="B3">3</xref>).</p>
<p>The evaluation of metastatic status in breast cancer patients involves both preoperative and postoperative assessments. Postoperative analyses provide detailed information about tumor characteristics, including lymphatic and vascular invasion, the extent of necrosis, and the degree of epithelial hyperplasia (<xref ref-type="bibr" rid="B4">4</xref>). While these parameters are valuable indicators of metastatic potential, they are only available after surgical intervention. Preoperative diagnostic tools, such as imaging modalities, offer preliminary insights into tumor behavior. However, definitive assessment of lymph node involvement&#x2014;a key determinant of metastatic spread&#x2014;relies on the analysis of the sentinel lymph node (SLN) obtained during surgery (<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B6">6</xref>) (<xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1A</bold>
</xref>). SLN biopsy is a minimally invasive technique that enables early detection of tumor cells disseminated from the primary site, facilitating precise therapeutic planning and reducing the need for more extensive surgical procedures (<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B5">5</xref>).</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Overview of the metastasis detection process with transfer learning. <bold>(A)</bold> demonstrates the sampling process conducted in the operating room, wherein the surgeon performs a procedure to locate and extract the sentinel lymph node (SLN) for subsequent pathology processing. <bold>(B)</bold> depicts the biological sample processing and image recording stages, which comprises two primary components. The first component involves slide preparation, hematoxylin and eosin (HE) staining, and immunohistochemistry (IHC) with cytokeratin staining. The second component involves slide identification via a barcode, image scanning as a whole slide image (WSI) file, and storage within the hospital server. <bold>(C)</bold> the computational section, comprises three distinct components. C1 pertains to the procedures associated with data generation and data processing, C2 is dedicated to the fine-tuning of the ResNet50 model, and the patch-level testing thereof, while C3 is focused on whole-slide testing. C1, in particular, encompasses the generation of 4 distinct patch datasets and encompasses the data preprocessing stage, which includes the creation of patches and various data augmentation techniques. C2 is characterized by the detailed procedures for fine-tuning the ResNet50 model, employing various configurations, including the different dataset types and augmentation settings. In this phase, patch-level testing is performed, and metrics such as accuracy (ACC), sensitivity (SEN), specificity (SPE), and F1-score (F1) are reported. C3 provides an elaborate account of the patch-level testing, using patient WSI of HE stained tissue sections. This phase also involves the reporting of metrics such as ACC, SEN, SPE and F1. Furthermore, it elucidates the implementation of a visual interface designed for medical professionals to facilitate the interpretation of model predictions. This interface supports the analysis and assessment of the model&#x2019;s performance, thereby enhancing its practical utility in clinical settings. FIB, tumor associated fibrosis; AUG, augmentations.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g001.tif"/>
</fig>
<p>Following SLN biopsy, the tissue is processed and evaluated by pathologists through microscopic examination. Tissue structures are differentiated using staining techniques, such as hematoxylin and eosin (HE) for general morphology or immunohistochemistry (IHC) for specific biomarkers (<xref ref-type="bibr" rid="B7">7</xref>) (<xref ref-type="fig" rid="f2">
<bold>Figures&#xa0;2A, B</bold>
</xref>, respectively). IHC is particularly valuable for identifying metastatic cells within normal parenchyma, especially in cases where morphological differentiation between normal and abnormal tissue is challenging using HE staining alone (<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B9">9</xref>) (<xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1B</bold>
</xref>). The IHC process involves the use of primary antibodies that bind to specific antigens in the tissue, followed by enzyme-conjugated secondary antibodies that produce a visible signal, typically a brown precipitate, at the target site (<xref ref-type="bibr" rid="B10">10</xref>).</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>Details of the mask generation process. <bold>(A)</bold> Patch of a whole slide image (WSI) stained with hematoxylin and eosin (HE) with tumoral cells. <bold>(B)</bold> The consecutive slice of the same surgical piece, stained with Cytokeratin immunohistochemistry (IHC, brown deposits indicating tumoral cells). <bold>(C)</bold> Mask generated only over tumoral cells. <bold>(D)</bold> Mask generated over tumoral cells and tumor-associated fibrosis (FIB). <bold>(A, B)</bold> were generated during the pathology section. <bold>(C, D)</bold> Were generated during the computational section.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g002.tif"/>
</fig>
<p>Advances in medical technology have enabled the digitization of histopathological slides, resulting in whole-slide images (WSIs). WSIs capture multiple magnifications of the same tissue region within a single file, allowing pathologists to navigate seamlessly from a macroscopic overview to high-resolution details. This digital transformation has facilitated the application of computer vision algorithms, particularly convolutional neural networks (CNNs), to automate the detection of tumor regions in biopsy specimens. Researchers have developed CNN-based models capable of identifying various cancer types, including breast cancer, within WSIs (<xref ref-type="bibr" rid="B11">11</xref>&#x2013;<xref ref-type="bibr" rid="B17">17</xref>). These models are trained on annotated datasets containing both tumor and healthy tissue samples, enabling them to learn discriminative patterns for accurate classification (<xref ref-type="bibr" rid="B18">18</xref>). To manage the computational demands of processing large WSIs, images are often divided into smaller patches at the highest resolution. The CNN generates patch-wise predictions, which are subsequently aggregated to produce comprehensive diagnostic visualizations (<xref ref-type="bibr" rid="B19">19</xref>) (<xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1C</bold>
</xref>).</p>
<p>Despite its potential, computational pathology faces two key challenges. First, the limited availability of patient data often requires multi-institutional collaborations, delaying research progress (<xref ref-type="bibr" rid="B20">20</xref>). Second, the computational resources needed for training models are immense, with studies requiring terabytes of storage and hundreds of thousands of GPU hours (<xref ref-type="bibr" rid="B21">21</xref>). These demands strain infrastructure and contribute to environmental concerns, as medical imaging accounts for approximately 1% of global greenhouse gas emissions (<xref ref-type="bibr" rid="B22">22</xref>).</p>
<p>This study proposes a transfer learning approach to address critical challenges in computational pathology for metastatic tumor detection. Building on the established success of deep learning in medical imaging, we investigate whether pretrained models can be effectively adapted to overcome data scarcity and computational constraints in clinical settings. Our work is guided by three principal hypotheses: First, that transfer learning with strategic fine-tuning can extract diagnostically relevant patterns from limited histopathology datasets while maintaining clinical-grade accuracy. Second, combining institutional data with curated public repositories will yield more robust feature representations than either source alone. Third, integrating intuitive visual interpretability tools can render model decisions transparent to pathologists without requiring specialized technical expertise.</p>
<p>The experimental framework employs a ResNet50 architecture initialized with ImageNet weights, subsequently fine-tuned on annotated whole-slide images of sentinel lymph node biopsies from Hospital Clinic Barcelona<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref>. Through carefully designed comparisons across eight experimental conditions, we systematically evaluate the impact of key factors including: (i) data composition strategies balancing local and public datasets, (ii) augmentation techniques for improved generalization, and (iii) evaluation protocols that assess both computational and clinical viability.</p>
<p>A novel aspect of this research is its dual-focus evaluation methodology, which combines traditional performance metrics with clinician-centric visualization tools. This approach seeks to establish whether computational predictions can simultaneously satisfy quantitative benchmarks and operational requirements for pathological diagnosis. The study design intentionally avoids presupposing optimal configurations, instead examining how varying degrees of data diversity and explainability influence model behavior in diagnostically relevant scenarios.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Related work</title>
<p>The detection of metastasis from primary tumors is a critical step in tumor staging, with SLN biopsies serving as the gold standard for assessing lymph node involvement (<xref ref-type="bibr" rid="B5">5</xref>, <xref ref-type="bibr" rid="B6">6</xref>). Historically, pathologists relied on optical microscopes to examine processed and stained tissue samples (<xref ref-type="bibr" rid="B23">23</xref>). However, the introduction of digital slide scanners has transformed this process, enabling the creation and storage of WSIs that can be analyzed computationally (<xref ref-type="bibr" rid="B24">24</xref>).</p>
<p>A central challenge in automated WSI analysis is the accurate classification of tumor regions. This requires the generation of masks&#x2014;annotations that distinguish cancerous cells from healthy tissue. Masks can be binary, indicating the presence or absence of cancer, or multi-label, providing additional information such as malignancy grades or the presence of immune cells within the tumor microenvironment (<xref ref-type="bibr" rid="B25">25</xref>, <xref ref-type="bibr" rid="B26">26</xref>). These masks are typically created by pathologists or derived from ICH techniques. In the latter approach, IHC is used to identify cancer regions in a tissue slice adjacent to the HE-stained slide. The spatial information from the IHC slice is then transferred to the HE slide to generate the corresponding mask (<xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B27">27</xref>).</p>
<p>In machine learning-based analysis of WSIs, the images are typically divided into smaller patches, where patch size serves as a critical parameter affecting both computational efficiency and the model&#x2019;s capacity to capture relevant contextual information (<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B29">29</xref>). To mitigate challenges associated with limited dataset size and variability, data augmentation techniques&#x2014;including rotation, contrast adjustment, noise injection, and grayscale conversion&#x2014;are widely adopted (<xref ref-type="bibr" rid="B15">15</xref>, <xref ref-type="bibr" rid="B16">16</xref>, <xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B31">31</xref>). While such augmentations can enhance model robustness and generalization, excessive application may introduce artificial patterns or distort underlying data distributions (<xref ref-type="bibr" rid="B32">32</xref>). Notably, some studies mention some benefits of augmentation without providing full comparative analyses against non-augmented baselines, raising questions about their necessity in certain contexts (<xref ref-type="bibr" rid="B33">33</xref>).</p>
<p>Two primary methodologies are commonly employed for training models in WSI patch classification. The first is transfer learning, wherein a model pretrained on a large-scale dataset (e.g., ImageNet) is fine-tuned for the target histopathological task. A key variant of this approach is end-to-end learning, which unifies feature extraction and classification within a single framework, enabling direct learning from image patches. This method is particularly beneficial when working with limited annotated data, as it leverages pre-existing feature representations. Widely used architectures for transfer learning include ResNet, AlexNet, DenseNet, EfficientNet, Inception V3, and VGG (<xref ref-type="bibr" rid="B34">34</xref>&#x2013;<xref ref-type="bibr" rid="B37">37</xref>). For example, ResNet-50 has demonstrated strong performance in breast cancer histopathology detection (<xref ref-type="bibr" rid="B38">38</xref>). However, a notable limitation of these deep learning models is their inherent &#x201c;black-box&#x201d; nature, often lacking interpretability in decision-making processes.</p>
<p>The second methodology separates feature extraction from classification. Handcrafted feature extraction involves using domain specific knowledge and algorithmic methods to identify and quantify relevant features within an image. These features are then fed into a separate classifier, such as a Support Vector Machine (SVM) or Random Forest (<xref ref-type="bibr" rid="B34">34</xref>, <xref ref-type="bibr" rid="B39">39</xref>, <xref ref-type="bibr" rid="B40">40</xref>). Examples of handcrafted features include those derived from texture analysis (e.g., Local Binary Patterns, Gabor filters), edge detection (e.g., Canny edge detector), or feature descriptors (e.g., SIFT, HOG) (<xref ref-type="bibr" rid="B41">41</xref>&#x2013;<xref ref-type="bibr" rid="B43">43</xref>). While handcrafted features are interpretable, they require extensive domain expertise and may not capture high-level abstractions as effectively as deep learning models.</p>
<p>The performance of deep learning models in histopathology is typically evaluated using a combination of quantitative metrics and qualitative visualizations. Standard quantitative measures include accuracy, sensitivity, specificity, precision, recall, F1-score, and the Area Under the Receiver Operating Characteristic Curve (AUC-ROC) (<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B40">40</xref>). On the qualitative side, visual&#xa0;assessments often involve the use of heatmaps, which highlight regions of interest within WSIs. These heatmaps can be displayed independently or overlaid on the original tissue images, occasionally requiring grayscale conversion to improve visual clarity (<xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B44">44</xref>). Such visualizations play a critical role in aiding medical professionals by enhancing the interpretability of&#xa0;model predictions and by facilitating the identification of cancerous regions, which in turn can support the development of new training datasets (<xref ref-type="bibr" rid="B28">28</xref>).</p>
<p>Among the various visualization techniques, Class Activation Map (CAM)-based methods are widely used for explaining CNN outputs. These techniques work by projecting back the weights from the final convolutional layers to generate saliency maps that highlight spatial regions most influential to a model&#x2019;s prediction (<xref ref-type="bibr" rid="B45">45</xref>). While effective for small-scale images, applying CAM-based methods to WSIs&#x2014;which can exceed 10 billion pixels at high resolution&#x2014;poses significant challenges. First, generating maps for each of the thousands of patches in a WSI can be computationally expensive. Second, the resulting pixel-level heatmaps across such a massive canvas can become visually overwhelming, introducing excessive detail that obscures broader diagnostic patterns. Third, in contrast to natural images where objects of interest are often visually prominent, regions of interest in pathology images tend to be subtle and lack strong saliency or clear boundaries. This makes it difficult for CAM-based methods to consistently localize relevant features, often resulting in under- or over-activation of regions and reducing the interpretability of the visualization at the clinical scale (<xref ref-type="bibr" rid="B45">45</xref>&#x2013;<xref ref-type="bibr" rid="B47">47</xref>).</p>
<p>In summary, despite notable advancements in automated WSIs analysis, key challenges persist in addressing data variability, optimizing data augmentation, and improving model interpretability. Our work advances current methodologies by (1) incorporating external datasets to better capture domain variability, (2) systematically evaluating the influence of augmentations on model performance, and (3) developing visualization tools to increase the transparency of deep learning models in metastasis detection.</p>
</sec>
<sec id="s3">
<label>3</label>
<title>Data and methods</title>
<sec id="s3_1">
<label>3.1</label>
<title>Data</title>
<p>In this study, WSIs were obtained from the Pathological Anatomy Service of the Biomedical Diagnostic Center at the Hospital Clinic Barcelona (BDC-HCB), Barcelona. The images were digitized using the Roche Diagnostics Ventana DP200 slide scanner at a final magnification of 40X. To ensure data privacy, the images were stored on a secure virtual disk with encryption and password-protected access. The Ethics Committee of HCB authorized the use of the images for this study.<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref> BDC-HCB provided two distinct batches of images. The first batch comprised 10 hematoxylin and eosin (HE) images and 5 immunohistochemical (IHC) images, which were utilized for model fine-tuning and patch-based testing. These images had an average resolution of 0.5 &#xb5;m/px and an average size of 75,000 px &#xd7; 50,000 px. Among the HE images, five were confirmed as positive for metastasis in the sentinel lymph node (SLN) (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, columns HE&#x2019;+&#x2019;), as validated by IHC using the mouse monoclonal antibody Cytokeratin 19 (A53-B/A2.26) against simple and complex epithelia, developed by Cell Marque Tissue Diagnostics (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, columns IHC-GT).</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Outline of the patch dataset generation procedure concerning image acquisition origins.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" colspan="6" align="center">WSI</th>
<th valign="top" colspan="4" align="center">Patches Production</th>
<th valign="middle" rowspan="2" colspan="2" align="center">Patch Datasets</th>
</tr>
<tr>
<th valign="top" colspan="2" align="center">HE (+)</th>
<th valign="top" colspan="2" align="center">HE (-)</th>
<th valign="top" colspan="2" align="center">IHC-GT</th>
<th valign="top" colspan="2" align="center">Creation</th>
<th valign="top" align="center" colspan="2">Label</th>
</tr>
<tr>
<th valign="top" align="center">Origin</th>
<th valign="top" align="center">WSI</th>
<th valign="top" align="center">Origin</th>
<th valign="top" align="center">WSI</th>
<th valign="top" align="center">Origin</th>
<th valign="top" align="center">WSI</th>
<th valign="top" align="center">MC</th>
<th valign="top" align="center">Size</th>
<th valign="top" align="center">(+)</th>
<th valign="top" align="center">(-)</th>
<th valign="top" align="center">ID</th>
<th valign="top" align="center">PN</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">1. HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">YES</td>
<td valign="top" align="center">90x90</td>
<td valign="top" align="center">TC &gt;20%</td>
<td valign="top" align="center">TC = 0</td>
<td valign="top" align="left">Tumor</td>
<td valign="top" align="left">23,900</td>
</tr>
<tr>
<td valign="top" align="left">2. HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">YES</td>
<td valign="top" align="center">90x90</td>
<td valign="top" align="center">TC &gt;20%</td>
<td valign="top" align="center">TC = 0</td>
<td valign="top" align="left">Tumor + FIB</td>
<td valign="top" align="left">23,900</td>
</tr>
<tr>
<td valign="top" align="left">3. HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">YES</td>
<td valign="top" align="center">90x90</td>
<td valign="top" align="center">TC &gt;20%</td>
<td valign="top" align="center">TC = 0</td>
<td valign="top" align="left">Tumor Extended</td>
<td valign="top" align="left">227,923</td>
</tr>
<tr>
<td valign="top" align="left">4. HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">HCB</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">YES</td>
<td valign="top" align="center">90x90</td>
<td valign="top" align="center">TC &gt;20%</td>
<td valign="top" align="center">TC = 0</td>
<td valign="middle" rowspan="2" align="left">Interhospital</td>
<td valign="middle" rowspan="2" align="left">104,107</td>
</tr>
<tr>
<td valign="top" align="center">5. CAM16</td>
<td valign="top" align="center">n = 5</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">0</td>
<td valign="top" align="center">NO</td>
<td valign="top" align="center">90x90</td>
<td valign="top" align="center">TC &gt;20%</td>
<td valign="top" align="center">TC = 0</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The presented table provides a comprehensive enumeration of whole slide image (WSIs) processing steps for the creation of specific patch datasets. Each row in the table corresponds to a distinct patch dataset and outlines various essential parameters. The &#x201c;WSI&#x201d; columns denote the source of the WSI to create patches, distinguishing between Hospital Clinic Barcelona (HCB) and Camelyon 16 (CAM). Within the table, &#x201c;HE +&#x201d; signifies the count of positive hematoxylin and eosin (HE) WSI, &#x201c;HE -&#x201d; represents the count of negative HE WSI, and &#x201c;IHC-GT&#x201d; indicates the number of immunohistochemistry WSI featuring brown deposits used as ground truth (GT) references. Utilizing the information from IHC-GT, masks for HE + patches are constructed, denoted in the mask creation &#x201c;MC&#x201d; column with a value of YES. CAM WSIs come pre-equipped with accompanying masks that have been published alongside the images, therefore the &#x201c;MC&#x201d; column is designated as NO. The &#x201c;Size&#x201d; column provides the dimensions of the resultant patches, while the &#x201c;Label&#x201d; column indicates the tumor content (TC) necessary for classifying a patch as positive or negative. The &#x201c;ID&#x201d; column specifies the corresponding patch dataset identification, and the &#x201c;PN&#x201d; column enumerates patch number. Notably, the &#x2018;Interhospital&#x2019; patch dataset, as delineated in rows 4 and 5, amalgamates data from both HCB and CAM sources, consolidating patches into a unified dataset. FIB, tumor associated fibrosis.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Consequently, each positive case included two corresponding images: one in bright field HE and another in bright field IHC. A key diagnostic feature of cancer cells in IHC images is the presence of brown deposits. The remaining five HE images, which were negative for SLN metastasis, were provided exclusively in bright field HE format (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, columns HE&#x2019;-&#x2019;). The second batch, consisting of 5 positive HE images and 5 corresponding IHC images, was reserved for whole slide testing.</p>
<p>Following the pathology section, the computational section was initiated, beginning with the annotation process. Tumor masks were generated by manually delineating tumor regions in HE images from positive cases (<xref ref-type="fig" rid="f2">
<bold>Figures&#xa0;2C, D</bold>
</xref>; <xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, column MC). Tumor areas were identified based on the presence of the marker in the corresponding IHC WSIs, as confirmed by pathologists (<xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2B</bold>
</xref>). The delineation of these masks was performed using the <italic>labelme</italic>
<xref ref-type="fn" rid="fn3">
<sup>3</sup>
</xref> software, applied to each previously identified tumor region. Each mask assigned a value&#xa0;of 1 to pixels corresponding to tumor cells and 0 to non-tumor cells. The alignment of the masks was rigorously validated through a three-stage visual inspection process: first by the pathologist, then by the imaging technician, and finally by the computational engineer.</p>
<p>Patches measuring 90 &#xd7; 90 pixels were extracted from the WSIs, each containing regions labeled as positive (tumor), negative (non-tumor), or a mix of both. A patch was classified as positive if more than 20% of its pixels were tumor-associated (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, column Label &#x2018;+&#x2019;). Critically, no upper threshold was imposed&#x2014;patches with varying tumor burdens (e.g., 30%, 50%, 80%, etc.) were randomly included in the positive class to improve model generalizability and reduce the risk of misclassifying partially tumorous patches. Conversely, patches with no tumor pixels were labeled as negative (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, column Label &#x2018;-&#x2019;). To address class imbalance, the more abundant negative class was subsampled to match the number of positive patches. Finally, a subset of patches was manually reviewed to verify label accuracy.</p>
<p>Four patch datasets were constructed using the first batch of WSIs, which included 10 HE images (5 positive and 5 negative cases), for model training (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, columns Patch datasets). The first dataset, labeled &#x2018;Tumor&#x2019;, exclusively contained patches with tumor cells, as defined by the masks of positive patches (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, first row). The second dataset, &#x2018;Tumor + FIB&#x2019;, incorporated tumor-related fibrosis alongside tumor cells within the masks of its positive patches (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, second row). The third dataset, &#x2018;Tumor extended&#x2019;, was created by expanding the &#x2018;Tumor&#x2019; dataset with additional patches (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, third row). The fourth dataset, &#x2018;Interhospital&#x2019;, comprised a 50%/50% mixture of patches from the &#x2018;Tumor&#x2019; dataset and samples from the Camelyon 16 dataset of equivalent size (<xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>, fourth row). The Camelyon 16 dataset was developed by retrospectively sampling sentinel lymph nodes (SLNs) from 399 patients who underwent breast cancer surgery at two Dutch hospitals: Radboud University Medical Center (RUMC) and University Medical Center Utrecht (UMCU). The slides were digitized using a Pannoramic 250 Flash II scanner (3DHISTECH) with a 20x objective lens. Metastases on the slides were annotated under the supervision of expert pathologists. While obvious metastases were annotated without IHC staining, IHC (anti-Cytokeratin [CAM 5.2], BD Biosciences) was employed for ambiguous cases (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B48">48</xref>). No image normalization techniques&#x2014;such as histogram equalization or style transfer&#x2014;were applied to the Camelyon16 dataset, as its inherent variability was intentionally preserved to assess its impact on model performance. As shown in <xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>, noticeable differences in contrast and saturation are evident across samples from different sources.</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Illustration showcasing patches extracted from our two image sources. The first row showcases 90x90 image patches with a positive label obtained from the Camelyon 16 dataset (<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B48">48</xref>). The second row presents 90x90 positive image patches sourced from the Biomedical Diagnostic Center of the Hospital Clinic (BDC-HC) in Barcelona.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g003.tif"/>
</fig>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Data augmentation and preprocessing</title>
<p>The image preprocessing pipeline incorporated multiple data augmentation techniques to enhance model robustness. First, color space transformations were applied by converting images to grayscale followed by transformation from RGB to HSV color space, with saturation values randomly sampled from a uniform distribution between 1 and 180. Noise augmentation included two components: Gaussian noise with mean values randomly selected from 0 to 10, and salt-and-pepper noise applied with probabilities ranging uniformly from 0 to 1. Additional image manipulations consisted of RGB channel separation and contrast adjustment through random scaling factors <italic>&#x3b1;</italic> uniformly distributed between 0 and 10, combined with brightness modification using random offsets <italic>&#x3b2;</italic> sampled from 0 to 10. Finally, all images underwent normalization to the (0,1) range, with subsequent calculation of per-channel mean and standard deviation values across the entire dataset.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Deep neural network model</title>
<p>We employed a pre-trained ResNet-50 architecture, implementing a transfer learning strategy that has demonstrated effectiveness in both general image classification tasks (<xref ref-type="bibr" rid="B49">49</xref>) and specialized medical imaging applications including breast cancer pathology detection (<xref ref-type="bibr" rid="B39">39</xref>, <xref ref-type="bibr" rid="B50">50</xref>). The ResNet-50 model consists of five hierarchical stages, each containing one convolution block and one identity block, with every block comprising three convolutional layers. This architecture, containing approximately 23 million trainable parameters (<xref ref-type="bibr" rid="B51">51</xref>), has proven particularly successful in cancer pathology detection tasks as evidenced by previous studies (<xref ref-type="bibr" rid="B52">52</xref>, <xref ref-type="bibr" rid="B53">53</xref>).</p>
<p>The model initialization utilized weights pre-trained on the ImageNet dataset (1000 classes), followed by task-specific fine-tuning for our histopathology application. All training experiments were conducted in a Google Colaboratory environment<xref ref-type="fn" rid="fn4">
<sup>4</sup>
</xref> with NVIDIA Tesla K80 GPU acceleration. During fine-tuning, we adopted a strategic approach where the first three convolutional layers remained frozen to preserve their pre-trained weights, while all subsequent layers including the classification head were made trainable. This design allowed the model to maintain robust low-level feature extractors while adapting higher-level features to our specific histopathology domain.</p>
<p>The training protocol employed cross-entropy loss function with a batch size of 128 across 18 epochs. We implemented weight decay (L2 regularization) with a coefficient of 1 &#xd7; 10<sup>&#x2212;4</sup> and used an adaptive learning rate schedule with an initial maximum value of 2 &#xd7; 10<sup>&#x2212;2</sup> that followed a cosine annealing pattern. This combination of hyperparameters was carefully selected to balance efficient convergence with effective regularization, enabling the model to adapt its higher-level representations while maintaining the fundamental visual feature extractors learned from the large-scale ImageNet dataset.</p>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Experimental setup</title>
<sec id="s3_4_1">
<label>3.4.1</label>
<title>Patch-based testing</title>
<p>For patch-based testing, eight fine-tuning experiments were conducted, each designed to address specific training conditions as described in <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>. These experiments involved four distinct patch datasets, with two experiments performed per dataset: one using original, unaltered patches and the other utilizing augmented patches (see <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>). The data were split into training and testing sets using a 70/30 ratio. After training and testing, the model weights were saved and exported, resulting in a total of eight preserved models. Additionally, validation measures were computed, as outlined in section 3.4.3.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Comprehensive summary of the experiments conducted to fine-tune and evaluate a ResNet50 model for sentinel lymph node biopsy classification.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" colspan="3" align="center">Fine-tuning</th>
<th valign="middle" align="center">Patch-based Testing</th>
<th valign="middle" colspan="2" align="center">Whole Slide Testing</th>
</tr>
<tr>
<th valign="middle" align="left">Experiment</th>
<th valign="middle" align="left">Dataset</th>
<th valign="middle" align="center">Augmentation</th>
<th valign="middle" align="center">Metrics across models (T4)</th>
<th valign="middle" align="center">Metrics across models (T5)</th>
<th valign="middle" align="center">Metrics across WSI (T6) 1</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">1</td>
<td valign="middle" align="left">Tumor</td>
<td valign="middle" align="left">0</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">2</td>
<td valign="middle" align="left">Tumor</td>
<td valign="middle" align="left">1</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">3</td>
<td valign="middle" align="left">Tumor + FIB</td>
<td valign="middle" align="left">0</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">4</td>
<td valign="middle" align="left">Tumor + FIB</td>
<td valign="middle" align="left">1</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">5</td>
<td valign="middle" align="left">Tumor Extended</td>
<td valign="middle" align="left">0</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">6</td>
<td valign="middle" align="left">Tumor Extended</td>
<td valign="middle" align="left">1</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">7</td>
<td valign="middle" align="left">Interhospital</td>
<td valign="middle" align="left">0</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
<tr>
<td valign="middle" align="left">8</td>
<td valign="middle" align="left">Interhospital</td>
<td valign="middle" align="left">1</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
<td valign="middle" align="center">ACC, SEN, SPE, F1 (Mean &#xb1; SD)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The table presents a two-pronged analytical approach for assessing the performance of eight ResNet50 models, initially fully trained on the ImageNet dataset, in classifying sentinel lymph node (SLN) status. The models underwent a series of fine-tuning experiments (column 1) using different datasets (column 2), with the 10 application of data augmentation assessed in a binary fashion (column 3). Performance metrics, Accuracy (ACC), Sensitivity (SEN), Specificity (SPE), and weighted F1 score, were evaluated under two testing conditions. The &#x2019;Patch-based Testing&#x2019; (column 4) determined the efficacy of models using patches from the same whole slide images (WSIs) used for fine-tuning, while the &#x2019;Whole Slide Testing&#x2019; (columns 5 and 6) involved a comprehensive assessment with unseen WSI, first by collating the aforementioned metrics across identical models (column 5) for each WSI (16, 18, 23, 25, and 27) and computing the mean and standard deviation (SD), then by aggregating these metrics for each WSI across all models (column 6), also calculating mean and SD for a thorough inter-model and intra-WSI comparison. This dual analysis elucidates both the collective and individual model performances relative to each WSI, offering a granular view of model robustness and diagnostic precision. T4, T5 and T6 refer to <xref ref-type="table" rid="T4">
<bold>Tables&#xa0;4</bold>
</xref>, <xref ref-type="table" rid="T5">
<bold>5</bold>
</xref> and <xref ref-type="table" rid="T6">
<bold>6</bold>
</xref> containing reported results.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3_4_2">
<label>3.4.2</label>
<title>Whole slide testing</title>
<p>To assess the robustness of the models generated in each experiment, we performed whole slide testing. This involved making predictions on WSIs that were not used for generating training patches. Specifically, the models were tested on a set of 5 WSIs from new cancer patients, each containing varying percentages of cancer pixels (see <xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>, column 1). Notably, these WSIs were entirely unseen by any of the models during training. To summarize the results, we generated tables reporting the mean and standard deviation (SD) of model performance. These tables were organized either by experiment or by patient WSI (refer to <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>, columns 5 and 6). When results were grouped by model, the mean and SD were calculated across all five patients. Conversely, when grouped by patient, the results were averaged across all four trained models. This final analysis enabled us to identify the patients for whom the models encountered the most significant performance challenges.</p>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Details of pixel content in whole slide testing samples.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="center">Case</th>
<th valign="top" align="center">HE</th>
<th valign="top" align="center">Case</th>
<th valign="top" align="center">IHC</th>
<th valign="top" align="center">Tissue (%)</th>
<th valign="top" align="center">Tumor (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">16</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">17</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">99.97</td>
<td valign="top" align="center">0.03</td>
</tr>
<tr>
<td valign="top" align="center">18</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">19</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">99.65</td>
<td valign="top" align="center">0.35</td>
</tr>
<tr>
<td valign="top" align="center">23</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">24</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">70.59</td>
<td valign="top" align="center">29.41</td>
</tr>
<tr>
<td valign="top" align="center">25</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">26</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">81.36</td>
<td valign="top" align="center">18.64</td>
</tr>
<tr>
<td valign="top" align="center">27</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">28</td>
<td valign="top" align="center">1</td>
<td valign="top" align="center">94.89</td>
<td valign="top" align="center">5.11</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Percentage of tissue and tumor pixels per whole slide image (WSI) are detailed in the last columns.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s3_4_3">
<label>3.4.3</label>
<title>Validation measures</title>
<p>To thoroughly evaluate the performance of the trained models in each experiment, we first generated confusion matrices for the patch-based testing stage. These matrices were created using Fastai&#x2019;s built-in functionality, with text-format labels assigned to each patch. For the whole slide testing stage, we implemented a custom sliding window method (illustrated in <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4</bold>
</xref>). This method utilized entire WSIs as the source of patches, with a binary mask serving as the ground truth reference. The output of this process was a probability matrix, where the prediction for each patch was spatially linked to its neighboring patches (see <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4C</bold>
</xref>).</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Protocol for prediction interpretation and visual aid. The methodology includes 5 steps, as follows: <bold>(A)</bold> the sliding window technique is applied to extract consecutive non-overlapping patches of 90x90 pixels from a whole slide image (WSI). <bold>(B)</bold> these image patches are then input to a pre-trained model, which generates predictions for each patch. <bold>(C)</bold> the output of the model generates a prediction matrix, where each prediction is assigned to the corresponding location of the patch. <bold>(D)</bold> a binary color map is utilized to visualize the predicted values above the threshold of 0.5, with red indicating the patches that meet this criterion. <bold>(E)</bold> a heatmap is constructed, utilizing red, green, and blue colors, to display the predicted probabilities for each patch. The color codes represent high (above 0.9), medium (0.7 to 0.9), and low probability (0.5 to 0.7) predictions, respectively.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g004.tif"/>
</fig>
<p>To classify patches as positive or negative, a threshold of 0.5 was applied. Patches with probabilities above this threshold were classified as positive, while those below were classified as negative. By comparing the patches with their corresponding labels in the binary mask, we identified true positives (TP) and true negatives (TN) where the labels matched. Patches with mismatched labels were classified as false positives (FP) or false negatives (FN).</p>
<p>We then calculated several validation metrics to assess model performance. These included accuracy (ACC) for analytical validation, sensitivity (SEN), specificity (SPE), and weighted F1-score (F1) for clinical validation, in accordance with the International Medical Device Regulator Forum&#x2019;s (IMDRF) guidelines for the clinical evaluation of Software as a Medical Device (SaMD)<xref ref-type="fn" rid="fn5">
<sup>5</sup>
</xref>. Additionally, AUC-ROC curves were constructed to further evaluate model performance.</p>
<p>Accuracy was calculated as the ratio of correctly classified instances (TP + TN) to the total number of instances (TP + TN + FP + FN). Sensitivity was determined as the ratio of TP to the sum of TP and FN. Specificity was computed as the ratio of TN to the sum of TN and FP. The F1 score, which combines precision (TP/(TP + FP)) and recall (sensitivity), was derived as the harmonic mean of these two metrics, providing a balanced measure of model performance.</p>
<disp-formula>
<mml:math display="block" id="M1">
<mml:mrow>
<mml:mtext>Accuracy</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<mml:math display="block" id="M2">
<mml:mrow>
<mml:mtext>Sensitivity</mml:mtext>
<mml:mi>&#x2009;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>Recall</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<mml:math display="block" id="M3">
<mml:mrow>
<mml:mtext>Specificity</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>N</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<mml:math display="block" id="M4">
<mml:mrow>
<mml:mtext>Precision</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula>
<mml:math display="block" id="M5">
<mml:mrow>
<mml:mtext>F</mml:mtext>
<mml:mn>1</mml:mn>
<mml:mtext>&#x2009;</mml:mtext>
<mml:mi>score</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext>Precision</mml:mtext>
<mml:mo>&#xd7;</mml:mo>
<mml:mtext>Recall</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext>Precision</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext>Recall</mml:mtext>
</mml:mrow>
</mml:mfrac>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
</sec>
<sec id="s3_4_4">
<label>3.4.4</label>
<title>Visual aid for classification interpretation</title>
<p>To enhance prediction interpretability and foster collaboration between physicians and data scientists, we developed a visual aid system based on tumor classification prediction matrices, featuring two visual tools. This system directly visualizes model predictions on image patches, avoiding excessive detail when composing the complete WSI, and bypasses further computations on model weights.</p>
<p>The first tool is a high-contrast monochromatic map that highlights all regions of the image with a tumor prediction probability above 0.5 (as illustrated in <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4D</bold>
</xref>). The second tool is a heatmap that uses three distinct colors to differentiate patches based on their prediction probabilities (as shown in <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4E</bold>
</xref>). The color-coding for the heatmap is defined as follows: red for high-probability predictions (above 0.9), green for medium-probability predictions (between 0.7 and 0.9), and blue for low-probability predictions (between 0.5 and 0.7). These visual tools are designed to enhance the clarity of the model&#x2019;s predictions and support effective communication between healthcare professionals and data scientists.</p>
</sec>
</sec>
</sec>
<sec id="s4" sec-type="results">
<label>4</label>
<title>Results</title>
<p>In this section, we present the results of our experiments, as summarized in <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>. Initially, we computed histograms for the 4 patch datasets generated. <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure&#xa0;1</bold>
</xref> illustrates that the histograms for positive samples in all 4 patch datasets exhibited a similar distribution to those of the negative samples, confirming an expected increase in the relative frequency of pixels with value 255 due to white patches in negative samples.</p>
<p>In the subsequent fine-tuning and patch-based testing phase, both processes utilized patches derived from the same WSIs. As detailed in <xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>, optimal performance across all models was consistently achieved when fine-tuning was performed without the application of data augmentations. Specifically, the &#x2018;Interhospital&#x2019; model (Experiment 7) demonstrated the highest accuracy (0.881), while the &#x2018;Tumor + FIB&#x2019; model (Experiment 3) exhibited the best sensitivity (0.756) and weighted F1-score. Notably, the &#x2018;Tumor&#x2019; model (Experiment 1) achieved the highest specificity (0.987). Regarding computational efficiency, model training times, conducted using NVIDIA Tesla K80 hardware for a fixed 18 epochs per dataset, varied considerably, and this variation appears correlated with dataset size. The &#x2018;Tumor&#x2019; and &#x2018;Tumor + FIB&#x2019; datasets, comprising 23,900 samples each, showed the most rapid training, both completing in approximately 5.87 minutes. In contrast, the &#x2018;Interhospital&#x2019; dataset, with 104,107 samples, required a significantly longer duration of 25.55 minutes, and the &#x2018;Tumor extended&#x2019; dataset, the largest with 227,923 samples, presented the longest training time at 55.93 minutes.</p>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Mean of patch-based testing metrics for each model with and without augmentations.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Model</th>
<th valign="middle" rowspan="2" align="center">AUG</th>
<th valign="top" colspan="4" align="center">Testing</th>
</tr>
<tr>
<th valign="top" align="center">ACC</th>
<th valign="top" align="center">SEN</th>
<th valign="top" align="center">SPE</th>
<th valign="top" align="center">F1</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">Tumor</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.673</bold>
<break/>0.671</td>
<td valign="top" align="center">0.232<break/>0.232</td>
<td valign="top" align="center">
<bold>0.987</bold>
<break/>0.983</td>
<td valign="top" align="center">
<bold>0.856</bold>
<break/>0.843</td>
</tr>
<tr>
<td valign="top" align="left">Tumor + FIB</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.878</bold>
<break/>0.729</td>
<td valign="top" align="center">
<bold>0.756</bold>
<break/>0.425</td>
<td valign="top" align="center">
<bold>0.970</bold>
<break/>0.956</td>
<td valign="top" align="center">
<bold>0.913</bold>
<break/>0.865</td>
</tr>
<tr>
<td valign="top" align="left">Tumor extended</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.477</bold>
<break/>0.467</td>
<td valign="top" align="center">0.001<break/>0.001</td>
<td valign="top" align="center">
<bold>0.973</bold>
<break/>0.968</td>
<td valign="top" align="center">
<bold>0.723</bold>
<break/>0.695</td>
</tr>
<tr>
<td valign="middle" align="left">Interhospital</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.881</bold>
<break/>0.787</td>
<td valign="top" align="center">0.000<break/>0.000</td>
<td valign="top" align="center">
<bold>0.948</bold>
<break/>0.895</td>
<td valign="top" align="center">
<bold>0.890</bold>
<break/>0.810</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>FIB, tumor related fibrosis; AUG, augmentations; ACC, accuracy; SEN, sensitivity; SPE, specificity; F1, weighted F1-score. Bold values indicate the highest value of the 2 metrics per model (with and without AUG).</p>
</table-wrap-foot>
</table-wrap>
<p>Following the fine-tuning phase, our research extended to the whole slide testing phase (<xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>, Whole Slide Testing). The metrics obtained during this phase differed significantly from the preliminary patch-based testing metrics (<xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>). To capture these differences, two tables were generated. <xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref> outlined the results for each fine-tuned model with and without augmentations, while <xref ref-type="table" rid="T6">
<bold>Table&#xa0;6</bold>
</xref> detailed the results for each image (individual patient) across all models, both with and without augmentations. These analyses aimed to identify robust models across all images and assess the difficulty levels in detecting metastasis for each image using all models.</p>
<table-wrap id="T5" position="float">
<label>Table&#xa0;5</label>
<caption>
<p>Mean and standard deviation of whole slide testing metrics for each model with and without augmentations.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Model</th>
<th valign="middle" rowspan="2" align="center">AUG</th>
<th valign="top" colspan="2" align="center">ACC</th>
<th valign="top" colspan="2" align="center">SEN</th>
<th valign="top" colspan="2" align="center">SPE</th>
<th valign="top" colspan="2" align="center">F1</th>
</tr>
<tr>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">Tumor</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.978<break/>0.978</td>
<td valign="top" align="center">0.019<break/>0.015</td>
<td valign="top" align="center">0.001<break/>
<bold>0.106</bold>
</td>
<td valign="top" align="center">0.002<break/>0.055</td>
<td valign="top" align="center">
<bold>0.995</bold>
<break/>0.994</td>
<td valign="top" align="center">0.008<break/>0.008</td>
<td valign="top" align="center">0.971<break/>
<bold>0.975</bold>
</td>
<td valign="top" align="center">0.029<break/>0.023</td>
</tr>
<tr>
<td valign="top" align="center">Tumor + FIB</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.979</bold>
<break/>0.978</td>
<td valign="top" align="center">0.020<break/>0.014</td>
<td valign="top" align="center">0.001<break/>
<bold>0.241</bold>
</td>
<td valign="top" align="center">0.002<break/>0.132</td>
<td valign="top" align="center">0.991<break/>
<bold>0.992</bold>
</td>
<td valign="top" align="center">0.004<break/>0.002</td>
<td valign="top" align="center">0.972<break/>
<bold>0.977</bold>
</td>
<td valign="top" align="center">0.029<break/>0.021</td>
</tr>
<tr>
<td valign="top" align="center">Tumor extended</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.977<break/>
<bold>0.981</bold>
</td>
<td valign="top" align="center">0.020<break/>0.019</td>
<td valign="top" align="center">
<bold>0.005</bold>
<break/>0.000</td>
<td valign="top" align="center">0.009<break/>0.000</td>
<td valign="top" align="center">0.995<break/>
<bold>0.998</bold>
</td>
<td valign="top" align="center">0.010<break/>0.002</td>
<td valign="top" align="center">0.971<break/>
<bold>0.973</bold>
</td>
<td valign="top" align="center">0.028<break/>0.029</td>
</tr>
<tr>
<td valign="top" align="center">Interhospital</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.958<break/>
<bold>0.982</bold>
</td>
<td valign="top" align="center">0.050<break/>0.015</td>
<td valign="top" align="center">0.219<break/>
<bold>0.256</bold>
</td>
<td valign="top" align="center">0.367<break/>0.143</td>
<td valign="top" align="center">0.971<break/>
<bold>0.995</bold>
</td>
<td valign="top" align="center">0.055<break/>0.003</td>
<td valign="top" align="center">0.963<break/>
<bold>0.978</bold>
</td>
<td valign="top" align="center">0.033<break/>0.022</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>FIB, tumor related fibrosis; AUG, augmentations; ACC, accuracy; SEN, sensitivity; SPE, specificity; F1, weighted F1-score; SD, standard deviation. Bold values indicate the highest value of the 2 metrics per model (with and without AUG).</p>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T6" position="float">
<label>Table&#xa0;6</label>
<caption>
<p>Mean and standard deviation of whole slide testing metrics for each sample, with and without augmentations.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">WSI</th>
<th valign="middle" rowspan="2" align="center">AUG</th>
<th valign="top" colspan="2" align="center">ACC</th>
<th valign="top" colspan="2" align="center">SEN</th>
<th valign="top" colspan="2" align="center">SPE</th>
<th valign="top" colspan="2" align="center">F1</th>
</tr>
<tr>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
<th valign="top" align="center">Mean</th>
<th valign="top" align="center">SD</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">16</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.999</bold>
<break/>0.996</td>
<td valign="top" align="center">0.000<break/>0.003</td>
<td valign="top" align="center">0.003<break/>
<bold>0.167</bold>
</td>
<td valign="top" align="center">0.005<break/>0.177</td>
<td valign="top" align="center">
<bold>0.999</bold>
<break/>0.996</td>
<td valign="top" align="center">0.000<break/>0.003</td>
<td valign="top" align="center">
<bold>0.999</bold>
<break/>0.998</td>
<td valign="top" align="center">0.000<break/>0.001</td>
</tr>
<tr>
<td valign="top" align="center">18</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">
<bold>0.996</bold>
<break/>0.989</td>
<td valign="top" align="center">0.003<break/>0.007</td>
<td valign="top" align="center">0.002<break/>
<bold>0.094</bold>
</td>
<td valign="top" align="center">0.005<break/>0.077</td>
<td valign="top" align="center">
<bold>0.997</bold>
<break/>0.989</td>
<td valign="top" align="center">0.003<break/>0.007</td>
<td valign="top" align="center">
<bold>0.997</bold>
<break/>0.993</td>
<td valign="top" align="center">0.001<break/>0.004</td>
</tr>
<tr>
<td valign="top" align="center">23</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.958<break/>
<bold>0.961</bold>
</td>
<td valign="top" align="center">0.003<break/>0.002</td>
<td valign="top" align="center">0.060<break/>
<bold>0.127</bold>
</td>
<td valign="top" align="center">0.114<break/>0.129</td>
<td valign="top" align="center">0.996<break/>
<bold>0.997</bold>
</td>
<td valign="top" align="center">0.003<break/>0.004</td>
<td valign="top" align="center">0.941<break/>
<bold>0.947</bold>
</td>
<td valign="top" align="center">0.008<break/>0.008</td>
</tr>
<tr>
<td valign="top" align="center">25</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.964<break/>
<bold>0.966</bold>
</td>
<td valign="top" align="center">0.000<break/>0.002</td>
<td valign="top" align="center">0.000<break/>
<bold>0.117</bold>
</td>
<td valign="top" align="center">0.001<break/>0.085</td>
<td valign="top" align="center">
<bold>0.999</bold>
<break/>0.997</td>
<td valign="top" align="center">0.000<break/>0.002</td>
<td valign="top" align="center">0.946<break/>
<bold>0.954</bold>
</td>
<td valign="top" align="center">0.000<break/>0.006</td>
</tr>
<tr>
<td valign="top" align="center">27</td>
<td valign="top" align="center">0<break/>1</td>
<td valign="top" align="center">0.948<break/>
<bold>0.988</bold>
</td>
<td valign="top" align="center">0.050<break/>0.002</td>
<td valign="top" align="center">0.219<break/>
<bold>0.249</bold>
</td>
<td valign="top" align="center">0.422<break/>0.217</td>
<td valign="top" align="center">0.956<break/>
<bold>0.995</bold>
</td>
<td valign="top" align="center">0.055<break/>0.003</td>
<td valign="top" align="center">0.963<break/>
<bold>0.986</bold>
</td>
<td valign="top" align="center">0.026<break/>0.003</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>WSI, whole slide image; AUG, augmentations; ACC, accuracy; SEN, sensitivity; SPE, specificity; F1, weighted F1-score; SD, standard deviation. Bold values indicate the highest value of the 2 metrics per model (with and without AUG).</p>
</table-wrap-foot>
</table-wrap>
<p>Upon analysis of this testing phase, models fine-tuned with augmentations consistently outperformed their non-augmented counterparts, as evidenced by the results in <xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref>. Particularly, the &#x2018;Interhospital&#x2019; model (Experiment 8) achieved the highest accuracy (0.982), sensitivity (0.256), and F1-score (0.978). Additionally, &#x2018;Tumor Extended&#x2019; (Experiment 6) demonstrated the highest specificity (0.998). The Area Under the ROC Curve (AUC-ROC) further supported &#x2018;Interhospital&#x2019; as the leading model (<xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure&#xa0;2</bold>
</xref>).</p>
<p>Examining patients individually, superior metrics were also evident in cases where augmentations were utilized (<xref ref-type="table" rid="T6">
<bold>Table&#xa0;6</bold>
</xref>), aligning with the trends previously observed in the whole slide testing phase (<xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref>). This observation emphasized the potential impact of augmentations on model performance, especially in scenarios involving domain changes and unseen images. Notably, patients 16 and 18 achieved the best accuracy and F1-score metrics, despite having low tumor patch content in their WSIs (0.03% and 0.35%, <xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>). This outcome indicated the models&#x2019; proficiency in identifying healthy tissue, offering potential benefits in the screening of patients with normal SLN biopsies.</p>
<p>Quantitative metrics analysis revealed that superior metrics were achieved without augmentations during model fine-tuning and patch-based testing (<xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>). This result was likely due to the low variability between the training and testing sets, sourced from the same WSIs. However, during whole slide testing, where domain changes occurred, models trained with augmentations demonstrated better performance (<xref ref-type="table" rid="T5">
<bold>Tables&#xa0;5</bold>
</xref>, <xref ref-type="table" rid="T6">
<bold>6</bold>
</xref>). This underscored the importance of using models trained on diverse data when transitioning domains, enhancing their generalization capabilities.</p>
<p>The qualitative metrics section served as a vital complement to the quantitative findings, offering valuable clinical insights. Cases with minimal tumor pixels, such as cases 16 and 18 (<xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref>), displayed distinct patterns from those with higher tumor counts. These cases exhibited tumor pixel rates of 0.03% and 0.35% across their WSIs, respectively (<xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>). The top-performing model, Interhospital, showed intersections with tumor regions in its predictions, but false positives were scattered throughout the images. Notably, false positives were also observed above areas of heterogeneous tissue with some degree of fibrosis in case 18 (<xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref>).</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Visual interface for interpretation of classification results for WSI in whole slide testing. The figure consists of four columns, where the first column (WSI HE) presents the original whole slide image (WSI) in hematoxylin and eosin (HE), while the second column shows the ground truth marking with immunohistochemistry (IHC) for Cytokeratin (WSI IHC). The brown deposits inthe IHC marking represent cancer cells, and a red arrow is utilized as a visual aid to locate brown deposits in cases that may be too small to identify. The third column depicts a binary map (BIinary MAP), with red masking indicating patches with predictions above 0.5. Finally, the fourth column exhibits a heatmap that portrays the prediction probability (Prediction MAP), where red, green, and blue indicate high (above 0.9), medium (between 0.7 and 0.9), and low probability (between 0.5 and 0.7) predictions, respectively. The rows <bold>(A&#x2013;E)</bold> represent different tissue sections of specific patients (16, 18, 25 and 27). Rows C and D represent different portions of WSI 25 (upper and lower, respectively). TC stands for tumor content and represents the percentage of tumor pixels in the WSI.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g005.tif"/>
</fig>
<p>Furthermore, we evaluated the models&#x2019; performance in cases featuring a substantial proportion of tumor pixels per image. Specifically, we analyzed WSI case 25 (depicted in the fourth and fifth rows) and case 27 (shown in the final row of <xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref>).</p>
<p>These cases displayed tumor pixel ratios of 18.64% and 5.11%, respectively (refer to <xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>). A significant increase of at least 1600% in pixel percentage was observed when comparing cases with low and high tumor pixel percentages. This considerable rise in tumor pixels correlated with the tumor cell count, indicating an extended period of tumor growth from the initial metastatic cell invasion of the SLN to the biopsy. Consequently, this prolonged duration likely contributed to the distinct differences between tumor cells and normal cells visual phenotype, improving the model&#x2019;s performance.</p>
<p>In the context of WSI 25, both tissue slices exhibited striking similarities between the ground truth area for IHC and the model&#x2019;s predictions, as demonstrated in <xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref> cases 25.1 and 25.2. Conversely, in WSI 27, while the IHC and binary predictions showed a significant overlap (highlighted in red), the color-coded prediction map provided valuable insights (<xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref> case 27). Specifically, regions with the highest predicted probabilities (0.9 and above, indicated in red) almost entirely coincided with tumor areas. The remaining predictions, characterized by lower probabilities (below 0.9, depicted in green and yellow), played a vital role in distinguishing false positives from true positives (highlighted in red). This detailed analysis underscored the model&#x2019;s accurate identification of tumor regions in cases with varying tumor pixel percentages, offering crucial information for clinical evaluation and diagnosis.</p>
<p>Moreover, upon a meticulous inspection of the prediction maps generated by all models overlaid on the same image (<xref ref-type="fig" rid="f6">
<bold>Figure&#xa0;6</bold>
</xref>), the distinctive outcomes align with the quantitative metrics, as illustrated in the F1 score results in <xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref> and the AUC ROC values presented in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Figure&#xa0;2</bold>
</xref>. Notably, the &#x2018;Interhospital&#x2019; model consistently outperforms other models across all three scenarios, followed by &#x2018;Tumor + FIB&#x2019;. The preceding analysis provides two critical insights. First, it highlights the importance of the visual interface for comprehensively evaluating model performance. The interface enables direct visualization of model outputs in the context of the original histopathology images, conferring clinical meaning to quantitative metrics. Second, it underscores the significance of heterogeneity in the whole slide testing. As observed, the top performing models were fine-tuned on the most heterogeneous datasets. Specifically, the &#x2018;Interhospital&#x2019; model was trained on samples from two distinct sources - locally generated data and the public Camelyon dataset. Meanwhile, the &#x2018;Tumor + FIB&#x2019; model was trained on patches containing not only tumor cells, but also associated fibrous stromal tissue.</p>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Differences in prediction map output for each model. The initial image represents the ground truth in immunohistochemistry (IHC), highlighting tumor areas. Images 2 to 5 depict prediction maps generated by individual models. The selected models for evaluation were those trained with augmentations, as they demonstrated superior performance. Heatmaps are constructed, utilizing red, green, and blue colors, to display the predicted probabilities for each patch. The color codes represent high (above 0.9), medium (0.7 to 0.9), and low probability (0.5 to 0.7) predictions, respectively.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fonc-15-1598289-g006.tif"/>
</fig>
</sec>
<sec id="s5" sec-type="discussion">
<label>5</label>
<title>Discussion</title>
<p>This study demonstrates the effectiveness of transfer learning in fine-tuning ResNet50 models for accurately detecting metastatic regions in SLN biopsy images. By leveraging Cytokeratin markers in IHC images to generate masks, we created four distinct patch datasets for fine-tuning, incorporating both augmented and non-augmented samples. Patch-level testing revealed that the highest accuracy was achieved without augmentations, likely due to overlaps between test and training samples from the same WSIs. However, whole-slide testing showed improved performance with augmentations, underscoring the role of data variability in enhancing model generalization. Notably, incorporating additional variability&#x2014;such as introducing samples from Camelyon into the Interhospital dataset&#x2014;provided the optimal diversity level, leading to superior performance on unseen images. These findings suggest that supplementing data augmentation with diverse datasets that maintain domain specificity can further enhance model robustness.</p>
<p>Our results provide support for our three initial hypotheses. First, transfer learning with appropriate fine-tuning indeed achieved clinically relevant performance (accuracy up to 0.982) despite using only 25 WSIs for training, supporting that pretrained models can overcome data scarcity limitations. Second, the hybrid Interhospital dataset consistently outperformed models trained on single-institution data (accuracy improvements of 3-5%), demonstrating that combining institutional with public datasets enhances generalization. Third, our visual explainability tools were able to bridge the technical-clinical gap - pathologists could interpret the heatmaps and binary maps with minimal training, validating their utility in real diagnostic workflows. These findings collectively help demonstrate that our proposed approach addresses the key challenges of data efficiency, generalizability, and clinical adoption in computational pathology.</p>
<p>The observed variations in model training times reveal important insights about computational efficiency. Models trained on smaller datasets (&#x2018;Tumor&#x2019; and &#x2018;Tumor + FIB&#x2019; with 23,900 samples) completed training significantly faster than those using larger collections. The &#x2018;Interhospital&#x2019; dataset (104,107 samples) showed moderate training duration, while the extensive &#x2018;Tumor Extended&#x2019; dataset (227,923 samples) required the longest processing time, confirming the expected scaling of computational most accurate model (patch-based testing: 0.881; whole-slide: 0.982 with augmentation), emphasizing that data diversity and quality outweigh sheer volume for metastatic detection tasks. This finding has important practical implications for clinical implementation, where both accuracy and computational efficiency are paramount.</p>
<p>We also noted a difference in performance between evaluating individual patches and entire slides. While the model consistently showed high specificity, meaning it was good at correctly identifying non-tumor samples, its sensitivity decreased when analyzing whole slides. This indicates a potential difficulty in detecting new tumor regions not seen during training. Data augmentation proved useful, as models trained with augmented data showed better tumor detection without sacrificing specificity. The best accuracy we achieved (0.98) was with an augmented model, which is comparable to results from prior studies classifying similar lymph node samples using GoogleNet (0.98), AlexNet (0.92), VGG16 (0.97), and FaceNet (0.96) (<xref ref-type="bibr" rid="B54">54</xref>&#x2013;<xref ref-type="bibr" rid="B57">57</xref>). Other research on different types of oncology images, also using transfer learning and augmentations (including techniques like edge detection we didn&#x2019;t explore), reported similar (VGG 19, ResNet 50, ResNet 152, MobileNet) and even higher accuracies (VGG 16, ResNet 101V2, DenseNet 169) (<xref ref-type="bibr" rid="B58">58</xref>).</p>
<p>Our evaluation metrics systematically assessed model performance across different testing conditions. When tested on previously unseen WSIs, models exhibited overall accuracy improvements, likely due to the higher proportion of non-tumor regions in WSIs, which enabled models to leverage their high specificity. However, the observed decrease in sensitivity highlights ongoing challenges in generalizing to new tumor data. The increase in weighted F1-score with additional training data suggests that models benefit from the larger proportion of healthy tissue in whole-slide images compared to tumor-enriched patch datasets. The ability to reliably identify non-tumor regions could facilitate computational pathology workflows by enabling automatic masking of healthy areas, allowing pathologists to focus on uncertain regions.</p>
<p>A key aspect of this work was creating a visual interface to improve clinical understanding. Overlaying identified metastatic areas onto the complete patient slides allowed pathologists to visually evaluate the model&#x2019;s predictions within the original context. Our results demonstrated accurate tumor patch detection when tumor content was above 16%, consistent with our quantitative findings. Indeed, our patch-based heatmap overlay offered the ideal level of detail for pathologists to differentiate zones of tumor development in the predictions. As noted, larger metastatic deposits tend to exhibit greater cellular diversity due to longer growth periods and potential variations in oxygen and nutrient access, leading to distinct morphologies within the same cluster (<xref ref-type="bibr" rid="B59">59</xref>).</p>
<p>Additionally, our approach offers a time-efficient alternative to traditional IHC procedures. Generating IHC-stained slides from biopsy tissue can take up to 16 hours (<xref ref-type="bibr" rid="B60">60</xref>, <xref ref-type="bibr" rid="B61">61</xref>), whereas our models process an HE-stained WSI in approximately 18 minutes&#x2014;a 98% reduction in processing time. By enabling rapid, automated identification of tumor and healthy regions, this approach has the potential to streamline computational pathology workflows and reduce clinician workload. However, challenges persist in cases with low tumor cell content or poorly differentiated morphology&#x2014;such as Cases 16 and 18&#x2014;where tumor cells closely resemble normal tissue. In addition to the lack of clear differentiation, these tumor regions covered a relatively small surface area, resulting in fewer representative patches with this morphology in the training data. This raises an important question: is the issue primarily due to poor differentiation, or is it the limited representation of such cases in the training dataset? To explore this, future studies should focus on datasets containing a sufficient number of poorly differentiated tumor samples alongside normal tissue. Despite these limitations, the model&#x2019;s ability to accurately exclude healthy tissue and flag uncertain regions for expert review highlights its time-saving potential compared to manual IHC slide analysis.</p>
<p>The clinical implications of this work extend beyond technical performance metrics. Our framework demonstrates how computational pathology can be strategically adapted to institutional settings through three key principles: leveraging transfer learning to overcome data limitations, curating hybrid datasets that balance domain specificity with diversity, and developing interpretation tools that align with pathologists&#x2019; diagnostic workflows. This approach not only achieves diagnostic accuracy comparable to manual assessment but does so while respecting the practical constraints of clinical environments, where computational resources and annotation bandwidth are often limited.</p>
<p>Looking forward, two parallel pathways emerge for advancing this research. First, technical refinements should focus on improving detection of diagnostically challenging cases, particularly micrometastases and tumors with ambiguous morphology. Second, clinical validation studies must establish how best to integrate such systems into real-world workflows, including optimal division of labor between AI and pathologists. The interpretability tools developed here provide a foundation for this transition, enabling collaborative decision-making where computational efficiency complements human expertise. Together, these directions promise to accelerate the translation of AI-assisted pathology from research laboratories to routine clinical practice.</p>
</sec>
<sec id="s6" sec-type="conclusions">
<label>6</label>
<title>Conclusions</title>
<p>This study validates three fundamental advancements in computational pathology for metastatic detection. First, we demonstrate that transfer learning with targeted fine-tuning can achieve clinical-grade accuracy using limited training data, overcoming establish that intuitive visualization tools effectively bridge the gap between computational outputs and clinical interpretation, facilitating pathologist adoption.</p>
<p>The success of our framework stems from its synergistic approach: (1) optimizing data efficiency through strategic transfer learning, (2) improving robustness via multi-source dataset integration, and (3) ensuring clinical relevance through interpretable visual analytics. These innovations collectively address the tripartite challenge of data scarcity, domain generalization, and workflow integration that has hindered widespread adoption of AI in pathology.</p>
<p>While demonstrating significant improvements over conventional methods, two key opportunities emerge for future research: enhancing detection of minimal residual disease and improving performance on morphologically ambiguous cases. Clinical translation of this validated pipeline could transform metastatic screening protocols by combining the efficiency of AI with pathological expertise, potentially establishing new standards for rapid, accurate lymph node assessment. The principles developed here may extend to other histopathological applications where data limitations and clinical interpretability remain persistent challenges.</p>
</sec>
</body>
<back>
<sec id="s7" sec-type="data-availability">
<title>Data availability statement</title>
<p>The WSI HE images and ground truth IHC from the Pathology Laboratory of Hospital Clinic Barcelona are inaccessible to the public to uphold individual privacy in compliance with the European General Data Protection Regulation. However, WSI HE images and ground truth masks from the Camelyon 16 dataset are accessible to the public via this site: <uri xlink:href="https://camelyon16.grand-challenge.org/Data/">https://camelyon16.grand-challenge.org/Data/</uri>. Requests to access these datasets should be directed to Geert Litjens, <email xlink:href="mailto:Geert.Litjens@radboudumc.nl">Geert.Litjens@radboudumc.nl</email>.</p>
</sec>
<sec id="s8" sec-type="ethics-statement">
<title>Ethics statement</title>
<p>The studies involving humans were approved by Comite de Etica del Hospital Clinic de Barcelona (CEIm). The studies were conducted in accordance with the local legislation and institutional requirements. The ethics committee/institutional review board waived the requirement of written informed consent for participation from the participants or the participants&#x2019; legal guardians/next of kin because of the retrospective nature of the study.</p>
</sec>
<sec id="s9" sec-type="author-contributions">
<title>Author contributions</title>
<p>MC: Formal analysis, Conceptualization, Methodology, Writing &#x2013; original draft, Project administration, Visualization, Resources, Data curation, Writing &#x2013; review &amp; editing. NW: Writing &#x2013; original draft, Software, Visualization, Formal analysis, Validation, Methodology, Data curation, Writing &#x2013; review &amp; editing. ET: Writing &#x2013; original draft, Methodology, Formal analysis, Data curation, Conceptualization. ES: Data curation, Writing &#x2013; review &amp; editing, Methodology, Formal analysis, Project administration. LI: Formal analysis, Investigation, Resources, Writing &#x2013; original draft, Conceptualization, Writing &#x2013; review &amp; editing, Methodology, Supervision.</p>
</sec>
<sec id="s10" sec-type="funding-information">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. This work was partially supported by the MICINN Grant PID2022-136436NB-I00, AGAUR Grant 2021-SGR-01104 and by the CRAI partnership at Universitat de Barcelona. </p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>We express our gratitude to Francesco Schettini for facilitating the connection between the Computer Vision Group at Universitat de Barcelona and the Pathology Service at Hospital Clinic and for his encouragement.</p>
</ack>
<sec id="s11" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s12" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec id="s13" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s14" sec-type="supplementary-material">
<title>Supplementary Material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fonc.2025.1598289/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fonc.2025.1598289/full#supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet1.pdf" id="SM1" mimetype="application/pdf"/>
</sec>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://www.clinicbarcelona.org/">https://www.clinicbarcelona.org/</ext-link>.</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>See Section 6.1: Ethics Approval and Consent to Participate.</p>
</fn>
<fn id="fn3">
<label>3</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://pypi.org/project/labelme/">https://pypi.org/project/labelme/</ext-link>.</p>
</fn>
<fn id="fn4">
<label>4</label>
<p>Using Fastai (1.0.61), Pytorch (1.0.0), and Openslide (1.2.0) packages.</p>
</fn>
<fn id="fn5">
<label>5</label>
<p>IMDRF/SaMD WG/N41; Software as a Medical Device (SaMD): Clinical Evaluation.</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Siegel</surname> <given-names>RL</given-names>
</name>
<name>
<surname>Miller</surname> <given-names>KD</given-names>
</name>
<name>
<surname>Jemal</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>Cancer statistics, 2019</article-title>. <source>CA: Cancer J Clin</source>. (<year>2019</year>) <volume>69 </volume>:<fpage>7</fpage>&#x2013;<lpage>34</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3322/caac.21551</pub-id>
</citation>
</ref>
<ref id="B2">
<label>2</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Edge</surname> <given-names>SB</given-names>
</name>
</person-group>. <source>Ajcc cancer staging manual</source> Vol. <volume>7</volume>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2010</year>) p. <fpage>97</fpage>&#x2013;<lpage>100</lpage>.</citation>
</ref>
<ref id="B3">
<label>3</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Litjens</surname> <given-names>G</given-names>
</name>
<name>
<surname>Bandi</surname> <given-names>P</given-names>
</name>
<name>
<surname>Bejnordi</surname> <given-names>BE</given-names>
</name>
<name>
<surname>Geessink</surname> <given-names>O</given-names>
</name>
<name>
<surname>Balkenhol</surname> <given-names>M</given-names>
</name>
<name>
<surname>Bult</surname> <given-names>P</given-names>
</name>
<etal/>
</person-group>. <article-title>1399 h&amp;e-stained sentinel lymph node sections of breast cancer patients: the camelyon dataset</article-title>. <source>GigaScience</source>. (<year>2018</year>) <volume>7</volume>:<fpage>giy065</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/gigascience/giy065</pub-id>
</citation>
</ref>
<ref id="B4">
<label>4</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yenidunya</surname> <given-names>S</given-names>
</name>
<name>
<surname>Bayrak</surname> <given-names>R</given-names>
</name>
<name>
<surname>Haltas</surname> <given-names>H</given-names>
</name>
</person-group>. <article-title>Predictive value of pathological and immunohistochemical parameters for axillary lymph node metastasis in breast carcinoma</article-title>. <source>Diagn Pathol</source>. (<year>2011</year>) <volume>6</volume>:<fpage>1</fpage>&#x2013;<lpage>9</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/1746-1596-6-18</pub-id>
</citation>
</ref>
<ref id="B5">
<label>5</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhou</surname> <given-names>L-Q</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>X-L</given-names>
</name>
<name>
<surname>Huang</surname> <given-names>S-Y</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>G-G</given-names>
</name>
<name>
<surname>Ye</surname> <given-names>H-R</given-names>
</name>
<name>
<surname>Wei</surname> <given-names>Qi</given-names>
</name>
<etal/>
</person-group>. <article-title>Lymph node metastasis prediction from primary breast cancer us images using deep learning</article-title>. <source>Radiology</source>. (<year>2020</year>) <volume>294</volume>:<fpage>19</fpage>&#x2013;<lpage>28</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1148/radiol.2019190372</pub-id>
</citation>
</ref>
<ref id="B6">
<label>6</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fleming</surname> <given-names>FJ</given-names>
</name>
<name>
<surname>Kavanagh</surname> <given-names>D</given-names>
</name>
<name>
<surname>Crotty</surname> <given-names>TB</given-names>
</name>
<name>
<surname>Quinn</surname> <given-names>CM</given-names>
</name>
<name>
<surname>McDermott</surname> <given-names>EW</given-names>
</name>
<name>
<surname>O&#x2019;higgins</surname> <given-names>N</given-names>
</name>
<etal/>
</person-group>. <article-title>Factors affecting metastases to non-sentinel lymph nodes in breast cancer</article-title>. <source>J Clin Pathol</source>. (<year>2004</year>) <volume>57</volume>:<page-range>73&#x2013;6</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1136/jcp.57.1.73</pub-id>
</citation>
</ref>
<ref id="B7">
<label>7</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Klevesath</surname> <given-names>MB</given-names>
</name>
<name>
<surname>Bobrow</surname> <given-names>LG</given-names>
</name>
<name>
<surname>Pinder</surname> <given-names>SE</given-names>
</name>
<name>
<surname>Purushotham</surname> <given-names>AD</given-names>
</name>
</person-group>. <article-title>The value of immunohistochemistry in sentinel lymph node histopathology in breast cancer</article-title>. <source>Br J Cancer</source>. (<year>2005</year>) <volume>92</volume>:<page-range>2201&#x2013;5</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/sj.bjc.6602641</pub-id>
</citation>
</ref>
<ref id="B8">
<label>8</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>King</surname> <given-names>TA</given-names>
</name>
<name>
<surname>Ganaraj</surname> <given-names>A</given-names>
</name>
<name>
<surname>Fey</surname> <given-names>JV</given-names>
</name>
<name>
<surname>Tan</surname> <given-names>LK</given-names>
</name>
<name>
<surname>Hudis</surname> <given-names>C</given-names>
</name>
<name>
<surname>Norton</surname> <given-names>L</given-names>
</name>
<etal/>
</person-group>. <article-title>Cytokeratin-positive cells in sentinel lymph nodes in breast cancer are not random events: experience in patients undergoing prophylactic mastectomy</article-title>. <source>Cancer</source>. (<year>2004</year>) <volume>101</volume>:<page-range>926&#x2013;33</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/cncr.v101:5</pub-id>
</citation>
</ref>
<ref id="B9">
<label>9</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Giuliano</surname> <given-names>AE</given-names>
</name>
<name>
<surname>Dale</surname> <given-names>PS</given-names>
</name>
<name>
<surname>Turner</surname> <given-names>RR</given-names>
</name>
<name>
<surname>Morton</surname> <given-names>DL</given-names>
</name>
<name>
<surname>Evans</surname> <given-names>SW</given-names>
</name>
<name>
<surname>Krasne</surname> <given-names>DL</given-names>
</name>
</person-group>. <article-title>Improved axillary staging of breast cancer with sentinel lymphadenectomy</article-title>. <source>Ann Surg</source>. (<year>1995</year>) <volume>222</volume>:<fpage>394</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1097/00000658-199509000-00016</pub-id>
</citation>
</ref>
<ref id="B10">
<label>10</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kabiraj</surname> <given-names>A</given-names>
</name>
<name>
<surname>Gupta</surname> <given-names>J</given-names>
</name>
<name>
<surname>Khaitan</surname> <given-names>T</given-names>
</name>
<name>
<surname>Bhattacharya</surname> <given-names>PT</given-names>
</name>
</person-group>. <article-title>Principle and techniques of immunohistochemistry&#x2014;a review</article-title>. <source>Int J Biol Med Res</source>. (<year>2015</year>) <volume>6</volume>:<page-range>5204&#x2013;10</page-range>.</citation>
</ref>
<ref id="B11">
<label>11</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ho</surname> <given-names>C</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>XF</given-names>
</name>
<name>
<surname>Sauer</surname> <given-names>J</given-names>
</name>
<name>
<surname>Saraf</surname> <given-names>SA</given-names>
</name>
<name>
<surname>Jialdasani</surname> <given-names>R</given-names>
</name>
<etal/>
</person-group>. <article-title>A promising deep learning-assistive algorithm for histopathological screening of colorectal cancer</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume>:<fpage>1</fpage>&#x2013;<lpage>9</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-022-06264-x</pub-id>
</citation>
</ref>
<ref id="B12">
<label>12</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Coudray</surname> <given-names>N</given-names>
</name>
<name>
<surname>Ocampo</surname> <given-names>PS</given-names>
</name>
<name>
<surname>Sakellaropoulos</surname> <given-names>T</given-names>
</name>
<name>
<surname>Narula</surname> <given-names>N</given-names>
</name>
<name>
<surname>Snuderl</surname> <given-names>M</given-names>
</name>
<name>
<surname>Feny&#xf6;</surname> <given-names>D</given-names>
</name>
<etal/>
</person-group>. <article-title>Classification and mutation prediction from non&#x2013;small cell lung cancer histopathology images using deep learning</article-title>. <source>Nat Med</source>. (<year>2018</year>) <volume>24</volume>:<page-range>1559&#x2013;67</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41591-018-0177-5</pub-id>
</citation>
</ref>
<ref id="B13">
<label>13</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Santos</surname> <given-names>GCD</given-names>
</name>
<name>
<surname>Ara&#xfa;jo</surname> <given-names>ALu&#xed;zaD</given-names>
</name>
<name>
<surname>Amorim</surname> <given-names>HAde</given-names>
</name>
<name>
<surname>Giraldo-Rold&#xe1;n</surname> <given-names>D</given-names>
</name>
<name>
<surname>Sousa-Neto</surname> <given-names>Sebasti&#xe3;oSilv&#xe9;riode</given-names>
</name>
<name>
<surname>Vargas</surname> <given-names>PA</given-names>
</name>
<etal/>
</person-group>. <article-title>Feasibility study of resnet-50 in the distinction of intraoral neural tumors using histopathological images</article-title>. <source>J Oral Pathol Med</source>. (<year>2024</year>). doi:&#xa0;<pub-id pub-id-type="doi">10.1111/jop.13560</pub-id>
</citation>
</ref>
<ref id="B14">
<label>14</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cruz-Roa</surname> <given-names>A</given-names>
</name>
<name>
<surname>Gilmore</surname> <given-names>H</given-names>
</name>
<name>
<surname>Basavanhally</surname> <given-names>A</given-names>
</name>
<name>
<surname>Feldman</surname> <given-names>M</given-names>
</name>
<name>
<surname>Ganesan</surname> <given-names>S</given-names>
</name>
<name>
<surname>Shih</surname> <given-names>NNC</given-names>
</name>
<etal/>
</person-group>. <article-title>Accurate and reproducible invasive breast cancer detection in whole-slide images: A deep learning approach for quantifying tumor extent</article-title>. <source>Sci Rep</source>. (<year>2017</year>) <volume>7</volume>:<fpage>1</fpage>&#x2013;<lpage>14</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/srep46450</pub-id>
</citation>
</ref>
<ref id="B15">
<label>15</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>M</given-names>
</name>
<name>
<surname>Hu</surname> <given-names>L</given-names>
</name>
<name>
<surname>Tang</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>C</given-names>
</name>
<name>
<surname>He</surname> <given-names>Yu</given-names>
</name>
<name>
<surname>Zeng</surname> <given-names>C</given-names>
</name>
<etal/>
</person-group>. <article-title>A deep learning method for breast cancer classification in the pathology images</article-title>. <source>IEEE J Biomed Health Inf</source>. (<year>2022</year>) <volume>26 </volume>:<page-range>5025&#x2013;32</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/JBHI.2022.3187765</pub-id>
</citation>
</ref>
<ref id="B16">
<label>16</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yan</surname> <given-names>R</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>F</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>L</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>T</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Y</given-names>
</name>
<etal/>
</person-group>. <article-title>Breast cancer histopathological image classification using a hybrid deep neural network</article-title>. <source>Methods</source>. (<year>2020</year>) <volume>173</volume>:<fpage>52</fpage>&#x2013;<lpage>60</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ymeth.2019.06.014</pub-id>
</citation>
</ref>
<ref id="B17">
<label>17</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Varavadekar</surname> <given-names>V</given-names>
</name>
<name>
<surname>Sajwan</surname> <given-names>V</given-names>
</name>
<name>
<surname>Mazo</surname> <given-names>C</given-names>
</name>
</person-group>. <article-title>Transfer learning approach for detecting breast cancer tumour subtypes in whole-slide images</article-title>. In: <source>2024 IEEE international symposium on biomedical imaging (ISBI)</source>. <publisher-loc>Athens, Greece</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2024</year>). p. <fpage>1</fpage>&#x2013;<lpage>4</lpage>.</citation>
</ref>
<ref id="B18">
<label>18</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>J</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Q</given-names>
</name>
<name>
<surname>Xie</surname> <given-names>H</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>H</given-names>
</name>
</person-group>. <article-title>Boosted efficientnet: Detection of lymph node metastases in breast cancer using convolutional neural networks</article-title>. <source>Cancers</source>. (<year>2021</year>) <volume>13</volume>:<fpage>661</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/cancers13040661</pub-id>
</citation>
</ref>
<ref id="B19">
<label>19</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kim</surname> <given-names>Y-G</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>S</given-names>
</name>
<name>
<surname>Cho</surname> <given-names>CE</given-names>
</name>
<name>
<surname>Song</surname> <given-names>InH</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>HJ</given-names>
</name>
<name>
<surname>Ahn</surname> <given-names>S</given-names>
</name>
<etal/>
</person-group>. <article-title>Effectiveness of transfer learning for enhancing tumor classification with a convolutional neural network on frozen sections</article-title>. <source>Sci Rep</source>. (<year>2020</year>) <volume>10</volume>:<fpage>1</fpage>&#x2013;<lpage>9</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-020-78129-0</pub-id>
</citation>
</ref>
<ref id="B20">
<label>20</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rajpurkar</surname> <given-names>P</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>E</given-names>
</name>
<name>
<surname>Banerjee</surname> <given-names>O</given-names>
</name>
<name>
<surname>Topol</surname> <given-names>EJ</given-names>
</name>
</person-group>. <article-title>Ai in health and medicine</article-title>. <source>Nat Med</source>. (<year>2022</year>) <volume>28</volume>:<page-range>31&#x2013;8</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41591-021-01614-0</pub-id>
</citation>
</ref>
<ref id="B21">
<label>21</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Kartasalo</surname> <given-names>K</given-names>
</name>
<name>
<surname>Valkonen</surname> <given-names>M</given-names>
</name>
<name>
<surname>Larsson</surname> <given-names>C</given-names>
</name>
<name>
<surname>Ruusuvuori</surname> <given-names>P</given-names>
</name>
<name>
<surname>Hartman</surname> <given-names>J</given-names>
</name>
<etal/>
</person-group>. <article-title>Predicting molecular phenotypes from histopathology images: a transcriptome-wide expression-morphology analysis in breast cancer</article-title>. <source>arXiv preprint arXiv:2009.08917</source>. (<year>2020</year>). doi:&#xa0;<pub-id pub-id-type="doi">10.1158/0008-5472.CAN-21-0482</pub-id>
</citation>
</ref>
<ref id="B22">
<label>22</label>
<citation citation-type="journal">
<article-title>The lancet digital health</article-title>. <source>Curbing carbon footprint Health Care</source>. (<year>2023</year>) <volume>5</volume>(<issue>12</issue>):<elocation-id>e848</elocation-id>.</citation>
</ref>
<ref id="B23">
<label>23</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ribatti</surname> <given-names>D</given-names>
</name>
<name>
<surname>Finato</surname> <given-names>N</given-names>
</name>
<name>
<surname>Crivellato</surname> <given-names>E</given-names>
</name>
<name>
<surname>Guidolin</surname> <given-names>D</given-names>
</name>
<name>
<surname>Longo</surname> <given-names>V</given-names>
</name>
<name>
<surname>Mangieri</surname> <given-names>D</given-names>
</name>
<etal/>
</person-group>. <article-title>Angiogenesis and mast cells in human breast cancer sentinel lymph nodes with and without micrometastases</article-title>. <source>Histopathology</source>. (<year>2007</year>) <volume>51</volume>:<page-range>837&#x2013;42</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1111/j.1365-2559.2007.02869.x</pub-id>
</citation>
</ref>
<ref id="B24">
<label>24</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Al-Janabi</surname> <given-names>S</given-names>
</name>
<name>
<surname>Huisman</surname> <given-names>Andr&#xe9;</given-names>
</name>
<name>
<surname>Diest</surname> <given-names>PJV</given-names>
</name>
</person-group>. <article-title>Digital pathology: current status and future perspectives</article-title>. <source>Histopathology</source>. (<year>2012</year>). doi:&#xa0;<pub-id pub-id-type="doi">10.1111/j.1365-2559.2011.03814.x</pub-id>
</citation>
</ref>
<ref id="B25">
<label>25</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>S</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>DM</given-names>
</name>
<name>
<surname>Rong</surname> <given-names>R</given-names>
</name>
<name>
<surname>Zhan</surname> <given-names>X</given-names>
</name>
<name>
<surname>Xiao</surname> <given-names>G</given-names>
</name>
</person-group>. <article-title>Pathology image analysis using segmentation deep learning algorithms</article-title>. <source>Am J Pathol</source>. (<year>2019</year>) <volume>189</volume>:<page-range>1686&#x2013;98</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ajpath.2019.05.007</pub-id>
</citation>
</ref>
<ref id="B26">
<label>26</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Xu</surname> <given-names>J</given-names>
</name>
<name>
<surname>Cai</surname> <given-names>C</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Yao</surname> <given-names>Bo</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>G</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>X</given-names>
</name>
<etal/>
</person-group>. <article-title>Multi-tissue partitioning for whole slide images of colorectal cancer histopathology images with deeptissue net</article-title>. In: <source>Digital pathology: 15th european congress, ECDP 2019, warwick, UK, april 10&#x2013;13, 2019, proceedings</source>, vol. <volume>15</volume>. <publisher-loc>Warwick, UK</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2019</year>). p. <page-range>100&#x2013;8</page-range>.</citation>
</ref>
<ref id="B27">
<label>27</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shamai</surname> <given-names>G</given-names>
</name>
<name>
<surname>Livne</surname> <given-names>A</given-names>
</name>
<name>
<surname>Pol&#xf3;nia</surname> <given-names>Ant&#xf3;nio</given-names>
</name>
<name>
<surname>Sabo</surname> <given-names>E</given-names>
</name>
<name>
<surname>Cretu</surname> <given-names>A</given-names>
</name>
<name>
<surname>Bar-Sela</surname> <given-names>G</given-names>
</name>
<etal/>
</person-group>. <article-title>Deep learning-based image analysis predicts pd-l1 status from h&amp;e-stained histopathology images in breast cancer</article-title>. <source>Nat Commun</source>. (<year>2022</year>) <volume>13</volume>:<fpage>6753</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41467-022-34275-9</pub-id>
</citation>
</ref>
<ref id="B28">
<label>28</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Fan</surname> <given-names>K</given-names>
</name>
<name>
<surname>Wen</surname> <given-names>S</given-names>
</name>
<name>
<surname>Deng</surname> <given-names>Z</given-names>
</name>
</person-group>. <article-title>Deep learning for detecting breast cancer metastases on wsi</article-title>. In: <source>In <italic>innovation in medicine and healthcare systems, and multimedia: proceedings of KES-inMed-19 and KES-IIMSS-19 conferences</italic>
</source>. <publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2019</year>). p. <page-range>137&#x2013;45</page-range>.</citation>
</ref>
<ref id="B29">
<label>29</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khened</surname> <given-names>M</given-names>
</name>
<name>
<surname>Kori</surname> <given-names>A</given-names>
</name>
<name>
<surname>Rajkumar</surname> <given-names>H</given-names>
</name>
<name>
<surname>Krishnamurthi</surname> <given-names>G</given-names>
</name>
<name>
<surname>Srinivasan</surname> <given-names>B</given-names>
</name>
</person-group>. <article-title>A generalized deep learning framework for whole-slide image segmentation and analysis</article-title>. <source>Sci Rep</source>. (<year>2021</year>) <volume>11</volume>:<fpage>11579</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-021-90444-8</pub-id>
</citation>
</ref>
<ref id="B30">
<label>30</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yari</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Nguyen</surname> <given-names>TV</given-names>
</name>
<name>
<surname>Nguyen</surname> <given-names>HT</given-names>
</name>
</person-group>. <article-title>Deep learning applied for histological diagnosis of breast cancer</article-title>. <source>IEEE Access</source>. (<year>2020</year>) <volume>8</volume>:<page-range>162432&#x2013;48</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/Access.6287639</pub-id>
</citation>
</ref>
<ref id="B31">
<label>31</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nguyen</surname> <given-names>CP</given-names>
</name>
<name>
<surname>Vo</surname> <given-names>AH</given-names>
</name>
<name>
<surname>Nguyen</surname> <given-names>BT</given-names>
</name>
</person-group>. <article-title>Breast cancer histology image classification using deep learning</article-title>. In: <source>2019 19th international symposium on communications and information technologies (ISCIT)</source>. <publisher-loc>US</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2019</year>). p. <page-range>366&#x2013;70</page-range>.</citation>
</ref>
<ref id="B32">
<label>32</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Romero Aquino</surname> <given-names>N</given-names>
</name>
<name>
<surname>Gutoski</surname> <given-names>M</given-names>
</name>
</person-group>. <article-title>Leandro T Hattori, and Heitor S Lopes. The effect of data augmentation on the performance of convolutional neural networks</article-title>. <source>Braz Soc Comput Intell</source>. (<year>2017</year>) <volume>10</volume>.</citation>
</ref>
<ref id="B33">
<label>33</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rodriguez</surname> <given-names>Jo&#xe3;oPM</given-names>
</name>
<name>
<surname>Rodriguez</surname> <given-names>R</given-names>
</name>
<name>
<surname>Silva</surname> <given-names>VWK</given-names>
</name>
<name>
<surname>Kitamura</surname> <given-names>FC</given-names>
</name>
<name>
<surname>Corradi</surname> <given-names>GCAnt&#xf4;nio</given-names>
</name>
<name>
<surname>Marchi</surname> <given-names>ACBde</given-names>
</name>
<etal/>
</person-group>. <article-title>Artificial intelligence as a tool for diagnosis in digital pathology whole slide images: a systematic review</article-title>. <source>J Pathol Inf</source>. (<year>2022</year>) <volume>13</volume>:<fpage>100138</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jpi.2022.100138</pub-id>
</citation>
</ref>
<ref id="B34">
<label>34</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sharmay</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Ehsany</surname> <given-names>L</given-names>
</name>
<name>
<surname>Syed</surname> <given-names>S</given-names>
</name>
<name>
<surname>Brown</surname> <given-names>DE</given-names>
</name>
</person-group>. <article-title>Histotransfer: understanding transfer learning for histopathology</article-title>. In: <source>In <italic>2021 IEEE EMBS international conference on biomedical and health informatics (BHI)</italic>
</source>. <publisher-loc>Athens, Greece</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2021</year>). p. <fpage>1</fpage>&#x2013;<lpage>4</lpage>.</citation>
</ref>
<ref id="B35">
<label>35</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nahid</surname> <given-names>A-A</given-names>
</name>
<name>
<surname>Kong</surname> <given-names>Y</given-names>
</name>
</person-group>. <article-title>Histopathological breast-image classification using local and frequency domains by convolutional neural network</article-title>. <source>Information</source>. (<year>2018</year>) <volume>9</volume>:<fpage>19</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/info9010019</pub-id>
</citation>
</ref>
<ref id="B36">
<label>36</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Benhammou</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Achchab</surname> <given-names>B</given-names>
</name>
<name>
<surname>Herrera</surname> <given-names>F</given-names>
</name>
<name>
<surname>Tabik</surname> <given-names>S</given-names>
</name>
</person-group>. <article-title>Breakhis based breast cancer automatic diagnosis using deep learning: Taxonomy, survey and insights</article-title>. <source>Neurocomputing</source>. (<year>2020</year>) <volume>375</volume>:<fpage>9</fpage>&#x2013;<lpage>24</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.neucom.2019.09.044</pub-id>
</citation>
</ref>
<ref id="B37">
<label>37</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Cascianelli</surname> <given-names>S</given-names>
</name>
<name>
<surname>Bello-Cerezo</surname> <given-names>R</given-names>
</name>
<name>
<surname>Bianconi</surname> <given-names>F</given-names>
</name>
<name>
<surname>Fravolini</surname> <given-names>ML</given-names>
</name>
<name>
<surname>Belal</surname> <given-names>M</given-names>
</name>
<name>
<surname>Palumbo</surname> <given-names>B</given-names>
</name>
<etal/>
</person-group>. <source>In <italic>intelligent interactive multimedia systems and services 2017</italic>
</source>, vol. <volume>10</volume>. <publisher-loc>Algarve, Portugal</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <fpage>21</fpage>&#x2013;<lpage>30</lpage>.</citation>
</ref>
<ref id="B38">
<label>38</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Gao</surname> <given-names>J</given-names>
</name>
<name>
<surname>Kan</surname> <given-names>H</given-names>
</name>
<name>
<surname>Huang</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Tang</surname> <given-names>F</given-names>
</name>
<name>
<surname>Li</surname> <given-names>W</given-names>
</name>
<etal/>
</person-group>. <article-title>Resnet for histopathologic cancer detection, the deeper, the better</article-title>? <source>J Data Sci Intelligent Syst</source>. (<year>2024</year>) <volume>2</volume>:<page-range>212&#x2013;20</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.47852/bonviewJDSIS3202744</pub-id>
</citation>
</ref>
<ref id="B39">
<label>39</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zewdie</surname> <given-names>ET</given-names>
</name>
<name>
<surname>Tessema</surname> <given-names>AW</given-names>
</name>
<name>
<surname>Simegn.</surname> <given-names>GL</given-names>
</name>
</person-group>. <article-title>Classification of breast cancer types, subtypes and grade from histopathological images using deep learning technique</article-title>. <source>Health Technol</source>. (<year>2021</year>) <volume>11</volume>:<page-range>1277&#x2013;90</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s12553-021-00592-0</pub-id>
</citation>
</ref>
<ref id="B40">
<label>40</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Vo-Le</surname> <given-names>C</given-names>
</name>
<name>
<surname>Son</surname> <given-names>NH</given-names>
</name>
<name>
<surname>Muoi</surname> <given-names>PV</given-names>
</name>
<name>
<surname>Phuong</surname> <given-names>NH</given-names>
</name>
</person-group>. <article-title>Breast cancer detection from histopathological biopsy images using transfer learning</article-title>. In: <source>In <italic>2020 IEEE eighth international conference on communications and electronics (ICCE)</italic>
</source>. <publisher-loc>Phu Quo, Vietnam</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2021</year>). p. <page-range>408&#x2013;12</page-range>.</citation>
</ref>
<ref id="B41">
<label>41</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Erfankhah</surname> <given-names>H</given-names>
</name>
<name>
<surname>Yazdi</surname> <given-names>M</given-names>
</name>
<name>
<surname>Babaie</surname> <given-names>M</given-names>
</name>
<name>
<surname>Tizhoosh</surname> <given-names>HR</given-names>
</name>
</person-group>. <article-title>Heterogeneity-aware local binary patterns for retrieval of histopathology images</article-title>. In: <source>IEEE access</source> <publisher-loc>Piscataway, US</publisher-loc>, vol. <volume>7</volume>. (<year>2019</year>). p. <page-range>18354&#x2013;67</page-range>.</citation>
</ref>
<ref id="B42">
<label>42</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Peikari</surname> <given-names>M</given-names>
</name>
<name>
<surname>Gangeh</surname> <given-names>MJ</given-names>
</name>
<name>
<surname>Zubovits</surname> <given-names>J</given-names>
</name>
<name>
<surname>Clarke</surname> <given-names>G</given-names>
</name>
<name>
<surname>Martel</surname> <given-names>AL</given-names>
</name>
</person-group>. <article-title>Triaging diagnostically relevant regions from pathology whole slides of breast cancer: A texture based approach</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2015</year>) <volume>35</volume>:<page-range>307&#x2013;15</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TMI.2015.2470529</pub-id>
</citation>
</ref>
<ref id="B43">
<label>43</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Xing</surname> <given-names>F</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>L</given-names>
</name>
</person-group>. <article-title>Robust nucleus/cell detection and segmentation in digital pathology and microscopy images: a comprehensive review</article-title>. <source>IEEE Rev Biomed Eng</source>. (<year>2016</year>) <volume>9</volume>:<page-range>234&#x2013;63</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/RBME.2016.2515127</pub-id>
</citation>
</ref>
<ref id="B44">
<label>44</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Singhal</surname> <given-names>N</given-names>
</name>
<name>
<surname>Soni</surname> <given-names>S</given-names>
</name>
<name>
<surname>Bonthu</surname> <given-names>S</given-names>
</name>
<name>
<surname>Chattopadhyay</surname> <given-names>N</given-names>
</name>
<name>
<surname>Samanta</surname> <given-names>P</given-names>
</name>
<name>
<surname>Joshi</surname> <given-names>U</given-names>
</name>
<etal/>
</person-group>. <article-title>A deep learning system for prostate cancer diagnosis and grading in whole slide images of core needle biopsies</article-title>. <source>Sci Rep</source>. (<year>2022</year>) <volume>12</volume> <fpage>3383</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-022-07217-0</pub-id>
</citation>
</ref>
<ref id="B45">
<label>45</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Abdullah</surname> <given-names>W</given-names>
</name>
<name>
<surname>Tolba</surname> <given-names>A</given-names>
</name>
<name>
<surname>Elmasry</surname> <given-names>A</given-names>
</name>
<name>
<surname>Mostafa</surname> <given-names>NN</given-names>
</name>
</person-group>. <article-title>Visioncam: a comprehensive xai toolkit for interpreting image-based deep learning models</article-title>. <source>Sustain Mach Intell J</source>. (<year>2024</year>) <volume>8</volume>:<fpage>4</fpage>&#x2013;<lpage>46</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.61356/SMIJ.2024.8290</pub-id>
</citation>
</ref>
<ref id="B46">
<label>46</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rony</surname> <given-names>J&#xe9;r&#xf4;me</given-names>
</name>
<name>
<surname>Belharbi</surname> <given-names>S</given-names>
</name>
<name>
<surname>Dolz</surname> <given-names>J</given-names>
</name>
<name>
<surname>Ayed</surname> <given-names>IB</given-names>
</name>
<name>
<surname>McCaffrey</surname> <given-names>L</given-names>
</name>
<name>
<surname>Granger.</surname> <given-names>E</given-names>
</name>
</person-group>. <article-title>Deep weakly-supervised learning methods for classification and localization in histology images: A survey</article-title>. <source>J Mach Learn Biomed Imaging</source>. (<year>2023</year>) <volume>2</volume>:<fpage>96</fpage>&#x2013;<lpage>150</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.59275/j.melba.2023-5g54</pub-id>
</citation>
</ref>
<ref id="B47">
<label>47</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname> <given-names>Y-x</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>F</given-names>
</name>
<name>
<surname>Shi</surname> <given-names>J-j</given-names>
</name>
<name>
<surname>Huang</surname> <given-names>Y-l</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>M</given-names>
</name>
</person-group>. <article-title>Convolutional neural networks for classifying cervical cancer types using histological images</article-title>. <source>J Digital Imaging</source>. (<year>2023</year>) <volume>36</volume>:<page-range>441&#x2013;9</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s10278-022-00722-8</pub-id>
</citation>
</ref>
<ref id="B48">
<label>48</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bandi</surname> <given-names>P</given-names>
</name>
<name>
<surname>Geessink</surname> <given-names>O</given-names>
</name>
<name>
<surname>Manson</surname> <given-names>Q</given-names>
</name>
<name>
<surname>Dijk</surname> <given-names>MV</given-names>
</name>
<name>
<surname>Balkenhol</surname> <given-names>M</given-names>
</name>
<name>
<surname>Hermsen</surname> <given-names>M</given-names>
</name>
<etal/>
</person-group>. <article-title>From detection of individual metastases to classification of lymph node status at the patient level: the camelyon17 challenge</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2018</year>) <volume>38</volume>:<page-range>550&#x2013;60</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TMI.2018.2867350</pub-id>
</citation>
</ref>
<ref id="B49">
<label>49</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sarwinda</surname> <given-names>D</given-names>
</name>
<name>
<surname>Paradisa</surname> <given-names>RH</given-names>
</name>
<name>
<surname>Bustamam</surname> <given-names>A</given-names>
</name>
<name>
<surname>Anggia</surname> <given-names>P</given-names>
</name>
</person-group>. <article-title>Deep learning in image classification using residual network (resnet) variants for detection of colorectal cancer</article-title>. <source>Proc Comput Sci</source>. <publisher-loc>Las vegas, US</publisher-loc> (<year>2021</year>) <volume>179</volume>:<page-range>423&#x2013;31</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.procs.2021.01.025</pub-id>
</citation>
</ref>
<ref id="B50">
<label>50</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Eldin</surname> <given-names>SN</given-names>
</name>
<name>
<surname>Hamdy</surname> <given-names>JK</given-names>
</name>
<name>
<surname>Adnan</surname> <given-names>GT</given-names>
</name>
<name>
<surname>Hossam</surname> <given-names>M</given-names>
</name>
<name>
<surname>Elmasry</surname> <given-names>N</given-names>
</name>
<name>
<surname>Mohammed</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>Deep learning approach for breast cancer diagnosis from microscopy biopsy images</article-title>. In: <source>In <italic>2021 international mobile, intelligent, and ubiquitous computing conference (MIUCC)</italic>
</source>. <publisher-loc>Cairo, Egypt</publisher-loc>: <publisher-name>IEEE</publisher-name> (<year>2021</year>). p. <page-range>216&#x2013;22</page-range>.</citation>
</ref>
<ref id="B51">
<label>51</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>S</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>J</given-names>
</name>
</person-group>. <article-title>Deep residual learning for image recognition</article-title>. In: <source>In <italic>Proceedings of the IEEE conference on computer vision and pattern recognition</italic>
</source> <publisher-loc>Las Vegas, US</publisher-loc> (<year>2016</year>). p. <page-range>770&#x2013;8</page-range>.</citation>
</ref>
<ref id="B52">
<label>52</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Le</surname> <given-names>A</given-names>
</name>
<name>
<surname>Li</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Tang</surname> <given-names>H</given-names>
</name>
<name>
<surname>Yang</surname> <given-names>H</given-names>
</name>
</person-group>. <article-title>A new breast cancer diagnosis application based on resnet50</article-title>. In: <source>Second IYSF academic symposium on artificial intelligence and computer engineering</source>, vol. <volume>12079</volume>. <publisher-loc>Xi'an, China</publisher-loc>: <publisher-name>SPIE</publisher-name> (<year>2021</year>). p. <page-range>615&#x2013;21</page-range>.</citation>
</ref>
<ref id="B53">
<label>53</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Vesal</surname> <given-names>S</given-names>
</name>
<name>
<surname>Ravikumar</surname> <given-names>N</given-names>
</name>
<name>
<surname>Davari</surname> <given-names>A</given-names>
</name>
<name>
<surname>Ellmann</surname> <given-names>S</given-names>
</name>
<name>
<surname>Maier</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>Classification of breast cancer histology images using transfer learning</article-title>. In: <source>
<italic>Image analysis and recognition: 15th international conference, ICIAR 2018, p&#xf3;voa de varzim, Portugal, june 27&#x2013;29, 2018, proceedings</italic> 15</source>. <publisher-loc>Povoa de Varzim, Portugal</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2018</year>). p. <page-range>812&#x2013;9</page-range>.</citation>
</ref>
<ref id="B54">
<label>54</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname> <given-names>C</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>W</given-names>
</name>
<name>
<surname>Jia</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Sermanet</surname> <given-names>P</given-names>
</name>
<name>
<surname>Reed</surname> <given-names>S</given-names>
</name>
<name>
<surname>Anguelov</surname> <given-names>D</given-names>
</name>
<etal/>
</person-group>. <article-title>Going deeper with convolutions</article-title>. In: <source>In <italic>Proceedings of the IEEE conference on computer vision and pattern recognition</italic>
</source> (<year>2015</year>). p. <fpage>1</fpage>&#x2013;<lpage>9</lpage>.</citation>
</ref>
<ref id="B55">
<label>55</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krizhevsky</surname> <given-names>A</given-names>
</name>
<name>
<surname>Sutskever</surname> <given-names>I</given-names>
</name>
<name>
<surname>Hinton</surname> <given-names>GE</given-names>
</name>
</person-group>. <article-title>Imagenet classification with deep convolutional neural networks</article-title>. <source>Adv Neural Inf Process Syst</source>. (<year>2012</year>) <volume>25</volume>.</citation>
</ref>
<ref id="B56">
<label>56</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname> <given-names>K</given-names>
</name>
<name>
<surname>Zisserman</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <source>arXiv preprint arXiv:1409.1556</source>. (<year>2014</year>) <publisher-loc>NY, US</publisher-loc>.</citation>
</ref>
<ref id="B57">
<label>57</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>D</given-names>
</name>
<name>
<surname>Otto</surname> <given-names>C</given-names>
</name>
<name>
<surname>Jain</surname> <given-names>AK</given-names>
</name>
</person-group>. <source>arXiv preprint arXiv:1507.07242</source> (<year>2015</year>).</citation>
</ref>
<ref id="B58">
<label>58</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kaur</surname> <given-names>H</given-names>
</name>
<name>
<surname>Sharma</surname> <given-names>R</given-names>
</name>
<name>
<surname>Kaur</surname> <given-names>J</given-names>
</name>
</person-group>. <article-title>Comparison of deep transfer learning models for classification of cervical cancer from pap smear images</article-title>. <source>Sci Rep</source>. (<fpage>3945</fpage>
<year>2025</year>) <volume>15</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-024-74531-0</pub-id>
</citation>
</ref>
<ref id="B59">
<label>59</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Naidoo</surname> <given-names>K</given-names>
</name>
<name>
<surname>Pinder</surname> <given-names>SE</given-names>
</name>
</person-group>. <article-title>Micro-and macro-metastasis in the axillary lymph node: A review</article-title>. <source>surgeon</source>. <publisher-loc>Edinburgh, Ireland</publisher-loc> (<year>2017</year>) <volume>15 </volume>:<fpage>76</fpage>&#x2013;<lpage>82</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.surge.2016.07.002</pub-id>
</citation>
</ref>
<ref id="B60">
<label>60</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Renshaw</surname> <given-names>S</given-names>
</name>
</person-group>. <source>Immunohistochemistry: methods express</source>. <publisher-name>Scion Publishing Ltd</publisher-name> (<year>2006</year>).</citation>
</ref>
<ref id="B61">
<label>61</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pina</surname> <given-names>Helo&#xef;se</given-names>
</name>
<name>
<surname>Salleron</surname> <given-names>J</given-names>
</name>
<name>
<surname>Gilson</surname> <given-names>P</given-names>
</name>
<name>
<surname>Husson</surname> <given-names>M</given-names>
</name>
<name>
<surname>Rouyer</surname> <given-names>M</given-names>
</name>
<name>
<surname>Leroux</surname> <given-names>A</given-names>
</name>
<etal/>
</person-group>. <article-title>ntraoperative prediction of non-sentinel lymph node metastases in breast cancer using cytokeratin 19 mrna copy number: A retrospective analysis</article-title>. <source>Mol Clin Oncol</source>. (<year>2022</year>) <volume>16</volume>:<fpage>1</fpage>&#x2013;<lpage>8</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3892/mco.2022.2491</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>