<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Comput. Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Computer Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Comput. Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-9898</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fcomp.2025.1646311</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>An explainable digital twin framework for skin cancer analysis using early activation meta-learner</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Sampath</surname> <given-names>Pradeepa</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<uri xlink:href="https://loop.frontiersin.org/people/3125207"/>
</contrib>
<contrib contrib-type="author">
<name><surname>S.</surname> <given-names>Gopika</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Vimal</surname> <given-names>S.</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<uri xlink:href="https://loop.frontiersin.org/people/1778405"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Kang</surname> <given-names>Yoonje</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<uri xlink:href="https://loop.frontiersin.org/people/3166953"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Seo</surname> <given-names>Sanghyun</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<uri xlink:href="https://loop.frontiersin.org/people/1814205"/>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Information Technology, School of Computing, SASTRA Deemed University, Thanjavur</institution>, <city>Tamilnadu</city>, <country country="in">India</country></aff>
<aff id="aff2"><label>2</label><institution>School of Computing and Emerging Technologies, Department of CSE, Kalaignarkarunanidhi Institute of Technology, Coimbatore</institution>, <city>Tamilnadu</city>, <country country="in">India</country></aff>
<aff id="aff3"><label>3</label><institution>Department of Applied Art and Technology, Chung-Ang University, Anseong</institution>, <city>Kyunggi-do</city>, <country country="kp">Republic of Korea</country></aff>
<aff id="aff4"><label>4</label><institution>School of Art and Technology, Chung-Ang University, Anseong</institution>, <city>Kyunggi-do</city>, <country country="kp">Republic of Korea</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Sanghyun Seo, <email xlink:href="mailto:sanghyun@cau.ac.kr">sanghyun@cau.ac.kr</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-09-01">
<day>01</day>
<month>09</month>
<year>2025</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>7</volume>
<elocation-id>1646311</elocation-id>
<history>
<date date-type="received">
<day>13</day>
<month>06</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>29</day>
<month>07</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2025 Sampath, S., Vimal, Kang and Seo.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Sampath, S., Vimal, Kang and Seo</copyright-holder>
<license>
<ali:license_ref start_date="2025-09-01">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Skin cancer is among the most common cancers globally, which calls for timely and precise diagnosis for successful therapy. Conventional methods of diagnosis, including dermoscopy and histopathology, are significantly dependent on expert judgment and therefore are time-consuming and susceptible to inconsistencies. Deep learning algorithms have shown potential in skin cancer classification but tend to consume a substantial amount of computational resources and large training sets. To overcome these issues, we introduce a new hybrid computer-aided diagnosis (CAD) system that integrates Stem Block for feature extraction and machine learning for classification. The International Skin Imaging Collaboration (ISIC) skin cancer dermoscopic images were collected from Kaggle, and essential features were collected from the Stem Block of a deep learning (DL) algorithm. The selected features, which were standardized using StandardScaler to achieve zero mean and unit variance, were then classified using a meta-learning classifier to enhance precision and efficiency. In addition, a digital twin framework was introduced to simulate and analyze the diagnostic process virtually, enabling real-time feedback and performance monitoring. This virtual replication aids in continuous improvement and supports the deployment of the CAD system in clinical environments. To improve transparency and clinical reliability, explainable artificial intelligence (XAI) methods were incorporated to visualize and interpret model predictions. Compared to state-of-the-art approaches, our system reduced training complexities without compromising high classification precision. Our proposed model attained an accuracy level of 96.25%, demonstrating its consistency and computationally efficient status as a screening tool for detecting skin cancer.</p></abstract>
<kwd-group>
<kwd>skin cancer</kwd>
<kwd>dermoscopic images</kwd>
<kwd>digital twin</kwd>
<kwd>deep learning</kwd>
<kwd>meta-learning</kwd>
<kwd>Grad-CAM</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. RS-2023-00218176) and the MSIT (Ministry of Science and ICT), Korea, under the Convergence security core talent training business support program (IITP-2025-RS-2023-00266605) supervised by the IITP (Institute for Information &#x00026; Communications Technology Planning &#x00026; Evaluation).</funding-statement>
</funding-group>
<counts>
<fig-count count="18"/>
<table-count count="2"/>
<equation-count count="31"/>
<ref-count count="40"/>
<page-count count="0"/>
<word-count count="7532"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Human-Media Interaction</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Skin cancer is one of the most widespread and serious health conditions globally, affecting millions of people annually. The skin is the largest human body organ, and it serves very important functions such as protecting against external elements, controlling temperature, and preventing excessive water loss. Due to its prolonged exposure to ultraviolet (UV) light, harmful environmental factors, and hereditary tendencies, the skin is susceptible to cancer. Skin cancer occurs when the skin cells grow abnormally and uncontrollably, usually because of extensive exposure to UV light (<xref ref-type="bibr" rid="B17">Melarkode et al., 2023</xref>).</p>
<p>The World Health Organization (WHO) states that skin cancer is responsible for one-third of all cancers worldwide, and melanoma is the deadliest and most malignant type (<xref ref-type="bibr" rid="B5">Brancaccio et al., 2024</xref>). Non-melanoma skin cancer (NMSC) accounts for an estimated 2 to 3 million cases annually, while melanoma recorded approximately 330,000 new cases and 60,000 deaths globally in 2022. The incidence of skin cancer is notably higher in regions such as Oceania, North America, and Europe (<xref ref-type="bibr" rid="B34">Wang et al., 2025</xref>). Australia reports rates 10 to 20 times higher than those in European populations. Melanoma has a five-year survival rate of 99.6% when diagnosed at an early stage, and delayed diagnosis markedly decreases survival rates, indicating the importance of early and correct detection techniques (<xref ref-type="bibr" rid="B23">Salinas et al., 2024</xref>).</p>
<p>Conventional skin cancer diagnosis is mainly based on dermatological inspection, dermoscopy, and histopathological examination (<xref ref-type="bibr" rid="B18">Naqvi et al., 2023</xref>). Dermoscopy is a painless imaging method that facilitates visualization of structures beneath the surface of lesions that are invisible to the naked eye, contributing to a high diagnostic precision (<xref ref-type="bibr" rid="B35">Wei et al., 2024</xref>). The reliability of dermoscopy is highly dependent on the dermatologist, making it vulnerable to inter-observer variation (<xref ref-type="bibr" rid="B2">Akter et al., 2025</xref>). Histopathological examination, the gold standard for skin cancer diagnosis, entails tissue biopsy followed by microscopic examination, but it is time-consuming, expensive, and invasive (<xref ref-type="bibr" rid="B28">Tai et al., 2024</xref>). Consequently, the need for faster, objective, and automated diagnostic tools has grown in recent times (<xref ref-type="bibr" rid="B6">Claret et al., 2024</xref>). With the recent advancements in artificial intelligence (AI) and deep learning (DL), computer-aided diagnosis (CAD) systems have proven to be efficient tools for detecting skin cancer (<xref ref-type="bibr" rid="B16">Mahmud et al., 2023</xref>).</p>
<p>Convolutional neural networks (CNNs) have achieved outstanding performance in automatically detecting hierarchical features of dermoscopic images to perform accurate skin lesion classification (<xref ref-type="bibr" rid="B11">Himel et al., 2024</xref>). Traditional end-to-end deep models are computationally demanding, requiring large labeled datasets, time-consuming training procedures, and vast computational power, which, in practical terms, seems to be not feasible in real-world settings (<xref ref-type="bibr" rid="B29">Thompson et al., 2020</xref>). Features with high dimensionality in a single CNN layer can complicate the classification task and result in overfitting, thus weakening the generalizability of models (<xref ref-type="bibr" rid="B9">Georgiou et al., 2020</xref>).</p>
<p>Many deep learning models have been developed to classify skin cancer; however, most of the current approaches have inherent limitations, which include concerns regarding excessive computational cost, overfitting due to limited training data, and lack of interpretability. In addition, traditional end-to-end classification models usually fail to take advantage of the complementary benefits of multiple learning methods. Our research addresses these limitations through a hybrid framework that employs deep feature extraction using a lightweight pre-trained model combined with a meta-learning-based stacking classifier. Unlike previous studies, our proposed framework takes into account the need for efficient feature selection, better generalization, and model interpretability in a skin cancer classification system designed for real-world clinical applications, where computing power and labeled data may be limited.</p>
<p>In view of these challenges, this study innovatively developed a new hybrid CAD system for the diagnosis of skin cancer that combines feature extraction from deep learning and machine learning-based classification.</p>
<list list-type="bullet">
<list-item><p>The proposed approach utilizes a pre-trained deep learning model to extract high-level features from dermoscopic images for skin cancer classification.</p></list-item>
<list-item><p>Instead of using all deep features indiscriminately, t-distributed stochastic neighbor embedding (t-SNE) visualization is employed to identify the most relevant and informative layers, enabling effective and selective feature extraction.</p></list-item>
<list-item><p>This feature selection strategy reduces redundancy and retains only the most discriminative representations of skin lesions, improving model efficiency.</p></list-item>
<list-item><p>The selected deep features are input into a meta-learning classifier, which effectively handles high-dimensional data and enhances classification performance.</p></list-item>
<list-item><p>The primary contribution is the development of a novel hybrid diagnostic framework that combines early activation deep feature extraction with meta-learning classification.</p></list-item>
<list-item><p>The methodology emphasizes the use of t-SNE-based layer selection to identify the most learnable layers in the early activation block of the model, ensuring optimal feature input.</p></list-item>
<list-item><p>By avoiding the use of all model layers, the approach significantly reduces computational complexity and training time without compromising accuracy.</p></list-item>
<list-item><p>The integration of a meta-learning classifier improves diagnostic accuracy when compared to traditional end-to-end deep learning models.</p></list-item>
<list-item><p>A digital Twin of the diagnostic system is conceptualized to simulate the end-to-end diagnostic workflow, enabling iterative testing, real-time monitoring, and safe validation before clinical deployment.</p></list-item>
<list-item><p>To support clinical trust and transparency, the framework incorporates explainable AI (XAI) techniques (e.g., Grad-CAM, saliency maps) to visually interpret model decisions and highlight key image regions influencing classification outcomes.</p></list-item>
</list>
</sec>
<sec sec-type="materials|methods" id="s2">
<label>2</label>
<title>Materials and methods</title>
<sec>
<label>2.1</label>
<title>Related works</title>
<p><xref ref-type="table" rid="T1">Table 1</xref> is a comparative overview of different deep learning and hybrid approaches utilized in recent research for skin lesion classification. The table summarizes each method&#x00027;s accuracy and its limitations. It is an exhaustive overview of available approaches based on model performance and complexity. This overview assists in determining the gaps and challenges that still exist in the field, including low performance, high complexity, and dependence on a single model type.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Literature review.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Author and year</bold></th>
<th valign="top" align="left"><bold>Methodology</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="left"><bold>Limitations</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B10">Gouda et al. (2022)</xref></td>
<td valign="top" align="left">CNN ResNet-50 Inception-ResNet Inception</td>
<td valign="top" align="center">85.7%</td>
<td valign="top" align="left">Low performance Perform preprocessing step</td>
</tr>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B12">Inthiyaz et al. (2023)</xref></td>
<td valign="top" align="left">CNN, Softmax classifier</td>
<td valign="top" align="center">87.42%</td>
<td valign="top" align="left">Low performance</td>
</tr>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B4">Amin et al. (2020)</xref></td>
<td valign="top" align="left">AlexNet and VGG-16, Bagged tree classifier</td>
<td valign="top" align="center">98.71%</td>
<td valign="top" align="left">Perform more preprocessing steps High complexity</td>
</tr>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B39">Zhang et al. (2020)</xref></td>
<td valign="top" align="left">Whale optimization CNN</td>
<td valign="top" align="center">93%</td>
<td valign="top" align="left">Only depends on DL</td>
</tr>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B31">Toga&#x000E7;ar et al. (2021)</xref></td>
<td valign="top" align="left">Two MobileNets ANN</td>
<td valign="top" align="center">95.27%</td>
<td valign="top" align="left">Used the same model twice High complexity</td>
</tr>
<tr>
<td valign="top" align="left"><xref ref-type="bibr" rid="B36">Wu et al. (2020)</xref></td>
<td valign="top" align="left">ARDT-DenseNet</td>
<td valign="top" align="center">87.8%</td>
<td valign="top" align="left">Low performance</td>
</tr></tbody>
</table>
</table-wrap>
<p>Deep learning algorithms are commonly employed for the extraction of intricate features from dermoscopic images, with research achieving accuracies of 98.75% using sophisticated feature extraction methods such as local optimal-oriented pattern (LOOP) and gray-level co-occurrence matrix (GLCM) (<xref ref-type="bibr" rid="B3">Alharbi et al., 2024</xref>). Hybrid systems that combine deep learning and machine learning classifiers have also shown promising results such as DenseNet201 with support vector machines (SVMs), which registered 91.09% accuracy on the International Skin Imaging Collaboration (ISIC) dataset. Transfer learning methodologies with pre-trained models such as EfficientNet-B7 and ResNet50 have also improved performance, with EfficientNet-B7 recording 84.4% accuracy (<xref ref-type="bibr" rid="B14">Kanchana et al., 2024</xref>). To overcome shortcomings such as computational inefficiency and large dataset requirement, a novel hybrid CAD system has been proposed, comprising a Stem Block for feature extraction and a meta-learning classifier for classification. The Stem Block identifies key features from dermoscopic images through deep learning algorithms, and the meta-learning classifier promotes accuracy by basing its approach on knowledge generalization across different datasets. A comparative evaluation was conducted between direct classification using the MobileNetV2 model and the proposed hybrid approach. In the direct classification setup, MobileNetV2 achieved an accuracy of 90.00%, whereas the proposed approach demonstrated a significant improvement with an accuracy of 96.25%. This enhancement can be attributed to the use of refined deep features and the ability of the meta-learning classifier to effectively combine multiple learners for better generalization. This model exhibited a high accuracy of 96.25%, with decreased training complexities, rendering it appropriate for real-time clinical use. In comparison to current state-of-the-art models, the proposed model performs better while being computationally efficient.</p>
</sec>
<sec>
<label>2.2</label>
<title>Methodology</title>
<p>During the implementation of the proposed hybrid CAD framework, several technical challenges were addressed to ensure optimal performance. One major challenge was the selection of meaningful deep features from the pre-trained MobileNetV2 model. Using all feature maps without evaluating their relevance led to increased computational cost and a higher risk of overfitting. To resolve this, t-SNE-based analysis was applied to identify and retain only the most discriminative layers, allowing the model to focus on high-impact features. Another key consideration was the design and tuning of the stacking classifier, where multiple base learners were integrated. It was essential to carefully validate these learners to ensure that they provided complementary decision boundaries without redundancy. The hyperparameters of both the base learners and the meta-classifier were fine-tuned through empirical testing to achieve optimal balance and performance. These methodological choices contributed to the overall robustness, efficiency, and high accuracy of the proposed hybrid model in skin cancer classification. <xref ref-type="fig" rid="F1">Figure 1</xref> shows the overall architecture diagram for the proposed model.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Work flow of the proposed model.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0001.tif">
<alt-text>Flowchart illustrating a machine learning pipeline for skin lesion classification. It starts with data collection images, followed by feature extraction using convolutional layers and bottleneck blocks. The features undergo visualization with t-SNE and are processed by various classifiers, including Cat Boost, XG Boost, SVC, Random Forest, Gradient Boosting, and KNN. Logistic regression integrates the meta-learning stage, leading to final classification.</alt-text>
</graphic>
</fig>
<sec>
<label>2.2.1</label>
<title>Dataset description</title>
<p>This section describes the benchmark dataset used to evaluate the performance of the proposed model. The skin lesion images are part of the dataset, and they are downloadable from the International Skin Imaging Collaboration (ISIC) repository. The dataset contains two classes: malignant and benign tumors, along with respective image annotations provided by medical doctors who performed biopsies to diagnose them. Dermoscopy imaging was performed to better visualize tumor structures. The dataset contains 3,297 images, of which 1,800 are benign and 1,497 are malignant. The dataset is publicly available on Kaggle. <xref ref-type="fig" rid="F2">Figure 2</xref> displays some sample images from the malignant class, and <xref ref-type="fig" rid="F3">Figure 3</xref> displays some sample images from the benign class. The images are scaled down to 224 &#x000D7; 224 pixels.</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>Skin cancer: malignant.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0002.tif">
<alt-text>Eight skin lesion images arranged in two rows and four columns. Each image shows different characteristics, such as variations in color, texture, and shape. The lesions range from reddish and pinkish hues to darker spots, with some displaying irregular borders.</alt-text>
</graphic>
</fig>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>Skin cancer: benign.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0003.tif">
<alt-text>Close-up images of skin lesions arranged in two rows, showing various colors, shapes, and textures. Each image highlights different characteristics of skin abnormalities, possibly for medical analysis or diagnosis.</alt-text>
</graphic>
</fig>
</sec>
<sec>
<label>2.2.2</label>
<title>Early activation block</title>
<p>The Stem Block serves as the foundational feature extractor, as shown in <xref ref-type="fig" rid="F4">Figure 4</xref>. It processes input images into a lower-resolution, high-channel representation suitable for subsequent layers (<xref ref-type="bibr" rid="B24">Sandler et al., 2018</xref>). The Stem Block consists of an initial 3 &#x000D7; 3 convolution with a stride of 2, followed by batch normalization and a ReLU6 activation, forming the conv1_relu layer (<xref ref-type="bibr" rid="B40">Zhu et al., 2024</xref>).</p>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>Architecture of the deep learning model used.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0004.tif">
<alt-text>Flowchart of a neural network architecture showing layers starting with Conv1 (3x3, 32), followed by sequential Bottleneck layers, Conv1 (1x1, 1280), Conv7, and more Bottleneck layers. The flow leads to Global Averaging, Dense layer, and ends with Extracted Features.</alt-text>
</graphic>
</fig>
<p>Depthwise separable convolution is a computationally efficient alternative to regular convolution and is utilized extensively in lightweight neural network frameworks. As opposed to conventional convolution where the same convolution filter is applied for all input channels, it performs the operation in two distinct steps: depthwise convolution and pointwise convolution. This considerably alleviates the cost of computation and improves model efficiency without significantly degrading performance. In the depthwise convolution process, one filter is applied to each input channel independently, instead of summing up all channels together using several filters. <xref ref-type="disp-formula" rid="EQ1">Equation 1</xref> represents the formula for the computational cost of depthwise convolution, where h is the height of the input feature map, w is the width of the input feature map, c is the number of input channels, and k is the size of the convolutional kernel.</p>
<disp-formula id="EQ1"><mml:math id="M1"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>h</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>c</mml:mi><mml:mo>&#x000D7;</mml:mo></mml:mstyle><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(1)</label></disp-formula>
<p><xref ref-type="disp-formula" rid="EQ2">Equation 2</xref> represents the computational cost of pointwise convolution. In pointwise convolution, 1 &#x000D7; 1 convolutions are employed to fuse the features of all channels, effectively combining channel information. Here, d is the number of output channels.</p>
<disp-formula id="EQ2"><mml:math id="M2"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>P</mml:mi><mml:mi>o</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>t</mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>h</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>c</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>d</mml:mi></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(2)</label></disp-formula>
<p><xref ref-type="disp-formula" rid="EQ3">Equation 3</xref> represents the overall computational cost of depthwise separable convolution, which is the sum of the depthwise and pointwise convolution costs.</p>
<disp-formula id="EQ3"><mml:math id="M3"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi><mml:mi>o</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mi>t</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>h</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>&#x000D7;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>&#x000D7;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>&#x000D7;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>d</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(3)</label></disp-formula>
<p>One of the key reasons for choosing the Stem Block is its complexity vs. performance trade-off. The architecture is tailored to improve feature representation through linear bottlenecks and expansion layers such that it can learn rich spatial and channel-wise information efficiently. While it is lightweight, it has been demonstrated to produce excellent classification accuracy on large-scale databases such as ImageNet, where it outperforms the majority of conventional architectures based on an efficiency-to-performance ratio.</p>
<p><xref ref-type="disp-formula" rid="EQ4">Equation 4</xref> represents the structure of an inverted residual block, where the input x is initially subjected to a 1 &#x000D7; 1 convolution to increase the number of channels, followed by a 3 &#x000D7; 3 depthwise convolution, then compressed using another 1 &#x000D7; 1 convolution. Linear activation occurs before residual connection addition. This design assists in maintaining the input while facilitating economical feature transformation at lower computational expense. In addition, it has also proven to be highly suitable for medical image classification tasks, such as the ISIC skin cancer dataset employed in this study (<xref ref-type="bibr" rid="B27">Surya et al., 2023</xref>).</p>
<disp-formula id="EQ4"><mml:math id="M4"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>B</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mtext>&#x02009;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>v</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>p</mml:mi><mml:mi>t</mml:mi><mml:mi>h</mml:mi><mml:mi>w</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>v</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>3</mml:mn></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>3</mml:mn></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>v</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(4)</label></disp-formula>
<p>The expansion layer is a key component in the architecture of deep learning models. Its purpose is to expand the number of channels before applying depthwise convolution. This operation allows the network to learn richer and more complex feature representations. The expansion layer increases the number of channels by a factor t, which is set to 6. The input channels <italic>C</italic><sub><italic>in</italic></sub> are multiplied by this factor to produce a higher-dimensional intermediate representation known as the hidden channels <italic>C</italic><sub><italic>hidden</italic></sub>. <xref ref-type="disp-formula" rid="EQ5">Equation 5</xref> represents the number of channels after expansion. It captures finer details and richer spatial features.</p>
<disp-formula id="EQ5"><mml:math id="M6"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>h</mml:mi><mml:mi>i</mml:mi><mml:mi>d</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;</mml:mtext><mml:mo>=</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>&#x000D7;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>6</mml:mn></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>C</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>n</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(5)</label></disp-formula>
<p>Medical imaging demands models that effectively extract relevant patterns at a computationally reasonable cost, particularly with big datasets. Its low-cost feature extraction mechanism (<xref ref-type="bibr" rid="B25">Sawal and Kathait, 2023</xref>) makes it a very good fit for transfer learning. This guarantees that the model retains richer feature representations along with fitting itself to the sensitivities of medical image classification (<xref ref-type="bibr" rid="B30">Tika Adilah and Kristiyanti, 2023</xref>). In comparison with heavyweight architectures, it provides equivalent classification accuracy but with much reduced computational requirements, which makes it a perfect fit for large-scale processing. In addition, its strong ability to generalize well across different tasks and datasets aligns with the scope of this study, where accurate and efficient classification of images of skin cancer is of high importance (<xref ref-type="bibr" rid="B21">Ragab et al., 2022</xref>). Due to the advantage of its lightweight design, powerful feature extraction ability, high efficiency, and excellent performance in medical image processing, it is a perfect model for this research (<xref ref-type="bibr" rid="B8">Ekmekyapar and Ta&#x0015F;ci, 2023</xref>). Its computation cost at the expense of meaningful feature extraction makes it perfect for skin lesion classification, finally improving diagnostic accuracy and clinical decision-making (<xref ref-type="bibr" rid="B37">Xu and Mohammadi, 2024</xref>).</p>
</sec>
<sec>
<label>2.2.3</label>
<title>Feature extraction</title>
<p>In this study, we used t-SNE to identify the most learnable feature representations from the model&#x00027;s intermediate layers. t-SNE is a non-linear dimensionality reduction method that maintains local structures in high-dimensional data by probabilistic similarity matching (<xref ref-type="bibr" rid="B32">Van der Maaten and Hinton, 2008</xref>). t-SNE first computes pairwise similarities between high-dimensional feature vectors. For two features <italic>f</italic><sub><italic>i</italic></sub> <italic>and f</italic><sub><italic>j</italic></sub> from a layer l, it calculates how similar they are using a Gaussian kernel. <xref ref-type="disp-formula" rid="EQ6">Equation 6</xref> expresses the similarity of point j to point i as a conditional probability, where <italic>f</italic><sub><italic>i</italic></sub> <italic>and f</italic><sub><italic>j</italic></sub> are feature vectors in high-dimensional space, &#x003C3; is the bandwidth of the Gaussian kernel, and <inline-formula><mml:math id="M7"><mml:mo>&#x02225;</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>-</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mo>&#x02225;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> is the squared Euclidean distance between the features.</p>
<disp-formula id="EQ6"><mml:math id="M8"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>p</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle><mml:mo>&#x02223;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mfrac><mml:mrow><mml:mi>exp</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mo>&#x02225;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:msub><mml:msup><mml:mo>&#x02225;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msup><mml:mo>/</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle><mml:msubsup><mml:mi>&#x003C3;</mml:mi><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msubsup><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext></mml:mrow><mml:mrow><mml:mstyle displaystyle='true'><mml:msub><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle><mml:mo>&#x02260;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mi>exp</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mo>&#x02212;</mml:mo><mml:mo>&#x02225;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:msup><mml:mo>&#x02225;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msup><mml:mo>/</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle><mml:msubsup><mml:mi>&#x003C3;</mml:mi><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mfrac></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(6)</label></disp-formula>
<p>To make the similarity symmetric and ensure consistency between i and j, <inline-formula><mml:math id="M9"><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi><mml:mo>|</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mi>p</mml:mi></mml:mrow><mml:mrow><mml:mi>j</mml:mi><mml:mo>|</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mi>N</mml:mi></mml:mrow></mml:mfrac></mml:math></inline-formula> is defined, where N is the number of samples. t-SNE maps high-dimensional data into a low-dimensional space using a student&#x00027;s t-distribution, which is a heavy-tailed distribution that helps prevent crowding, with one degree of freedom. <xref ref-type="disp-formula" rid="EQ7">Equation 7</xref> represents the similarity between the points <italic>y</italic><sub><italic>i</italic></sub> and <italic>y</italic><sub><italic>j</italic></sub>, which are low-dimensional points, and the denominator ensures that all <italic>q</italic><sub><italic>ij</italic></sub> sum to 1 to make it a proper probability distribution.</p>
<disp-formula id="EQ7"><mml:math id="M10"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>q</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mfrac><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mo>+</mml:mo><mml:mo>&#x02225;</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:msub><mml:msup><mml:mo>&#x02225;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mstyle displaystyle='true'><mml:msub><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle><mml:mo>&#x02260;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo>&#x02225;</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>l</mml:mi></mml:mstyle></mml:msub><mml:msup><mml:mo>&#x02225;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mo>&#x02212;</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mrow></mml:mstyle></mml:mrow></mml:mfrac></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(7)</label></disp-formula>
<p>As a next step, t-SNE minimizes the difference between the high-dimensional and low-dimensional similarity distributions using the Kullback&#x02013;Leibler (KL) divergence. <xref ref-type="disp-formula" rid="EQ8">Equation 8</xref> is the objective function to minimize the difference, where <italic>p</italic><sub><italic>ij</italic></sub> is the similarity between the points i and j in the high-dimensional space, and <italic>q</italic><sub><italic>ij</italic></sub> is the similarity between the points <italic>i</italic> and <italic>j</italic> in the low-dimensional embedding. This cost function encourages similar points in the high-dimensional space to stay close together in the low-dimensional map. It allows us to visually interpret clusters, patterns, and outliers in complex datasets.</p>
<disp-formula id="EQ8"><mml:math id="M11"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mstyle displaystyle="true"><mml:mo>&#x02211;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mo>&#x02260;</mml:mo><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>p</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mo class="qopname">log</mml:mo><mml:mstyle mathvariant="bold"></mml:mstyle><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>p</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>q</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(8)</label></disp-formula>
<p>Given the complexity of skin lesion classification, effective feature selection plays an important role in improving classification performance (<xref ref-type="bibr" rid="B38">Yuan et al., 2024</xref>; <xref ref-type="bibr" rid="B33">Villa-Pulgarin et al., 2022</xref>). Instead of arbitrarily choosing an intermediate layer for feature extraction, we utilized t-SNE to analyze the clustering behavior of feature representations from multiple layers within the pre-trained model&#x00027;s early activation block. Feature maps were extracted from several convolutional and dense layers. Each layer&#x00027;s output was reshaped into a feature vector representation. The extracted high-dimensional features were projected into a two-dimensional space using t-SNE. The objective was to observe how well the feature representations of benign and malignant skin lesions were clustered. Layers exhibiting well-separated clusters of benign and malignant lesions were considered the most learnable, as they provided high discriminative power. We selected the Conv1_relu layer from the Stem Block as it produced a well-clustered plot, as shown in <xref ref-type="fig" rid="F12">Figure 12</xref>. The features from the layers were selected for further processing in the meta-learning classifier.</p>
<p>The Conv1_ReLU layer in the Stem Block consists of a convolutional operation followed by batch normalization and a ReLU6 activation function. The first layer of the network is responsible for extracting initial low-level features from the input image. It starts with a standard 3 &#x000D7; 3 convolution that processes the input RGB image X &#x02208; <italic>R</italic><sup>224 <italic>X</italic> 224 <italic>X</italic> 3</sup>. The convolution uses a filter tensor W &#x02208; <italic>R</italic><sup>3 <italic>X</italic> 3 <italic>X</italic> 3 <italic>X</italic> 32</sup>, which consists of 32 filters, each of size 3 &#x000D7; 3 &#x000D7; 3, along with a bias term B &#x02208; <italic>R</italic><sup>32</sup>. <xref ref-type="disp-formula" rid="EQ9">Equation 9</xref> defines the operation as follows:</p>
<disp-formula id="EQ9"><mml:math id="M12"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>Z</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>X</mml:mi><mml:mtext>&#x000A0;</mml:mtext><mml:mo>*</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:mi>W</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>B</mml:mi></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(9)</label></disp-formula>
<p>where <sup>&#x0002A;</sup> denotes the convolution operation with a stride of 2. This operation reduces the spatial resolution and expands the depth channels from 3 to 32. Following convolution, the output is passed through batch normalization to stabilize training and improve convergence. For each output channel K &#x02208; {1, ... ,32}, the feature map is normalized using the channel-wise mean &#x003BC;<sup>(<italic>k</italic>)</sup> and standard deviation &#x003C3;<sup>(<italic>k</italic>)</sup>, as shown in <xref ref-type="disp-formula" rid="EQ10">Equation 10</xref>. <xref ref-type="disp-formula" rid="EQ11">Equation 11</xref> denotes that the normalized value is scaled and shifted using learnable parameters &#x003B3;<sup>(<italic>k</italic>)</sup>, &#x003B2;<sup>(<italic>k</italic>)</sup>.</p>
<disp-formula id="EQ10"><mml:math id="M13"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mfrac><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mstyle mathvariant="bold"><mml:mo>-</mml:mo></mml:mstyle><mml:msup><mml:mrow><mml:mi>&#x003BC;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:mi>&#x003C3;</mml:mi><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mi>&#x003F5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(10)</label></disp-formula>
<disp-formula id="EQ11"><mml:math id="M14"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msup><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:msup><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msup><mml:mrow><mml:mi>&#x003B2;</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(11)</label></disp-formula>
<p>After normalization, the layer applies the ReLU6 activation function, which clips the values between 0 and 6. It is defined as ReLU6(y) = min(max(y,0), 6) and ReLU6(y) = min(max(y,0), 6). This helps maintain stability in low-precision computations. <xref ref-type="disp-formula" rid="EQ12">Equation 12</xref> shows the final output of this layer.</p>
<disp-formula id="EQ12"><mml:math id="M15"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>F</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>U</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>6</mml:mn></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>Y</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x02208;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>R</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>112</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mn>112</mml:mn><mml:mo>&#x000D7;</mml:mo><mml:mn>32</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(12)</label></disp-formula>
</sec>
<sec>
<label>2.2.4</label>
<title>Meta-learning classifier</title>
<p>Meta-learning is realized as a stacking classifier, which is a form of ensemble-based meta-learning. We do not perform task-level learning (similar to few-shot learning), but instead model-level learning, where the meta-learner is trained to optimally combine the predictions of many base classifiers. Each base learner captures a different facet of the input feature space extracted from the pre-trained deep model; their outputs are given to a higher-level learner (the meta-classifier), which learns to produce the final prediction. This also turns the model into some sort of a stacked architecture, enabling the model to utilize diverse classifiers with complementary strengths to enhance model robustness and generalization capabilities. This approach is common in standard meta-learning methods, as it uses knowledge across a range of models to improve classification performance, especially when noise or unbalanced data are present, such as in our classification of skin lesions.</p>
<p>Meta-learning is a method in which predictions of different base models are combined with a meta-classifier (<xref ref-type="bibr" rid="B13">Itoo and Garg, 2022</xref>), as shown in <xref ref-type="fig" rid="F5">Figure 5</xref>. The meta-classifier is trained to weigh the base model outputs optimally and combine them for better predictive accuracy.</p>
<disp-formula id="EQ13"><mml:math id="M16"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mstyle></mml:mrow></mml:msup><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mi>t</mml:mi><mml:mi>a</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle></mml:mrow></mml:msup></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>B</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(13)</label></disp-formula>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>Meta-learner.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0005.tif">
<alt-text>The Figure 5 shows a stacking ensemble model where six classifiers generate predictions fed into a logistic regression meta-learner. This meta-learning approach combines outputs from diverse models to improve classification accuracy and robustness in decision-making.</alt-text>
</graphic>
</fig>
<p><xref ref-type="disp-formula" rid="EQ13">Equation 13</xref> shows how the meta-classifier computes the prediction from all the base models. Here, B is the number of base models, <italic>y</italic><sup><italic>b</italic></sup> is the prediction from the base models, and f-meta is the final estimator, which is a Lasso-regularized logistic regression in this case.</p>
<p>In a meta-learning classifier, prediction is performed through a two-tier process involving multiple base models and a meta-model. The base layer consists of multiple individual models <italic>f</italic><sub><italic>b</italic></sub>, each trained independently. For every input X, each base model generates its own prediction <italic>y</italic><sup><italic>b</italic></sup>, as shown in <xref ref-type="disp-formula" rid="EQ14">Equation 14</xref>. Here, <italic>b</italic> indexes each base learner, and <italic>f</italic><sub><italic>b</italic></sub>(X) denotes the prediction made by the b-th model on input X.</p>
<disp-formula id="EQ14"><mml:math id="M17"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>X</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>&#x02200;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi><mml:mo>&#x02208;</mml:mo></mml:mstyle><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>,</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>&#x02026;</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>,</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>6</mml:mn></mml:mstyle></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(14)</label></disp-formula>
<p>The predictions from all base learners are then passed to a meta-layer. It learns to combine them optimally. This is performed using logistic regression as the meta-model, with L1 regularization. <xref ref-type="disp-formula" rid="EQ15">Equation 15</xref> shows the final prediction being class 1, given all base predictions. Here, <italic>y</italic><sup><italic>b</italic></sup> is the prediction from the b-th base model, <italic>Bw</italic><sub><italic>b</italic></sub> is the weight assigned to each base model&#x00027;s output by the meta-model, <italic>w</italic><sub>0</sub> is the bias term in the logistic regression, and &#x003C3;() is the sigmoid activation function used to produce the final probability. To prevent overfitting and increase sparsity, an L1 regularization constraint is placed on the weight vector w, such that ||<italic>w</italic>||<sub>1</sub> &#x02264; &#x003BB;. Here, ||<italic>w</italic>||<sub>1</sub> represents the L1 norm of the weights, and &#x003BB; is a hyperparameter controlling the strength of regularization.</p>
<disp-formula id="EQ15"><mml:math id="M18"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>P</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mtext>&#x000A0;</mml:mtext><mml:mo>|</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mi>&#x003C3;</mml:mi><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle displaystyle="true"><mml:mo>&#x02211;</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>B</mml:mi></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>b</mml:mi></mml:mstyle></mml:mrow></mml:msup><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>0</mml:mn></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(15)</label></disp-formula>
<p>In the meta-classifier, six base models are employed&#x02014;namely, CatBoost, XGBoost, SVM, random forest, gradient boosting, and k-nearest neighbors (KNN). CatBoost is a gradient boosting decision tree that is specifically used for categorical data. It is used for effective and precise classification and regression. It encodes the categorical features with target statistics from previous rows (<xref ref-type="bibr" rid="B20">Prokhorenkova et al., 2018</xref>). <xref ref-type="disp-formula" rid="EQ16">Equation 16</xref> is the encoded value for the categorical feature of the <italic>i</italic>-th sample. Here, <italic>y</italic><sub><italic>j</italic></sub> represents the true target values of previous rows, and i-1 ensures that only preceding samples are used, preserving causality and avoiding target leakage. <xref ref-type="disp-formula" rid="EQ17">Equation 17</xref> represents the objective function that includes both the loss term and regularization term. Here, n is the total number of samples, <italic>L</italic> (<italic>y</italic><sub><italic>i</italic></sub>, <italic>F</italic>(<italic>x</italic><sub><italic>i</italic></sub>)) is the loss function, F(x) = <inline-formula><mml:math id="M19"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow></mml:munderover><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:math></inline-formula> (x) is the model&#x00027;s cumulative prediction over T decision trees, and &#x003A9;(<italic>f</italic><sub><italic>t</italic></sub>) is a regularization function applied to each individual tree <italic>f</italic><sub><italic>t</italic></sub>.</p>
<disp-formula id="EQ16"><mml:math id="M20"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mfrac><mml:mrow><mml:mstyle displaystyle="true"><mml:msubsup><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mo>-</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msubsup></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>-</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(16)</label></disp-formula>
<disp-formula id="EQ17"><mml:math id="M21"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle><mml:mtext>&#x02009;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>n</mml:mi></mml:mstyle></mml:munderover><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle><mml:mtext>&#x02009;</mml:mtext><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>F</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle></mml:munderover><mml:mo>&#x003A9;</mml:mo></mml:mstyle><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle></mml:msub></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(17)</label></disp-formula>
<p>XGBoost is another gradient boosting algorithm that minimizes the loss function with a second-order Taylor expansion (<xref ref-type="bibr" rid="B22">Raihan et al., 2023</xref>). It is specifically built for high-speed classification and regression tasks, tuning both speed and accuracy. <xref ref-type="disp-formula" rid="EQ18">Equation 18</xref> represents the objective function at the t-th iteration. Here, n is the number of training samples, <inline-formula><mml:math id="M22"><mml:mi>L</mml:mi><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>y</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x000A0;</mml:mtext><mml:msup><mml:mrow><mml:mi>F</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>t</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula> is the loss between the true label and prediction at iteration t, T denotes the number of leaves in the newly added tree, and &#x003B3; and &#x003BB; are regularization hyperparameters that penalize complexity. <xref ref-type="disp-formula" rid="EQ19">Equation 19</xref> represents the regularization component that encourages simpler trees. Here, &#x003B3;T discourages the number of leaves in the tree, discouraging overly complex models, and <inline-formula><mml:math id="M23"><mml:mfrac><mml:mrow><mml:mi>&#x003BB;</mml:mi></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mo>&#x02225;</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x02225;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></inline-formula> is a regularization term applied to the leaf weights w, controlling their magnitude and enhancing generalization.</p>
<disp-formula id="EQ18"><mml:math id="M24"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:msup><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>n</mml:mi></mml:mstyle></mml:munderover><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:msup><mml:mstyle mathvariant="bold-italic"><mml:mi>F</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003B3;</mml:mi><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mfrac><mml:mi>&#x003BB;</mml:mi><mml:mstyle mathvariant="bold-italic"><mml:mn>2</mml:mn></mml:mstyle></mml:mfrac><mml:mo>&#x02225;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle><mml:msup><mml:mo stretchy='false'>&#x02225;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mn>2</mml:mn></mml:mstyle></mml:msup></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(18)</label></disp-formula>
<disp-formula id="EQ19"><mml:math id="M25"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mo>&#x003A9;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>f</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mi>&#x003B3;</mml:mi><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mfrac><mml:mrow><mml:mi>&#x003BB;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:mfrac><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mo>&#x02225;</mml:mo><mml:mi>w</mml:mi><mml:mo>&#x02225;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(19)</label></disp-formula>
<p>SVC identifies a hyperplane that maximizes the margin between two classes. SVC supports linear and non-linear classification with kernel functions. It performs well in high-dimensional spaces and is resistant to overfitting, particularly in situations with a clear margin of separation between classes. <xref ref-type="disp-formula" rid="EQ20">Equation 20</xref> represents the dual optimization objective performed by SVC. Here, &#x003B1;<sub><italic>i</italic></sub> are the Lagrange multipliers to be improved, and <italic>K</italic>(<italic>x</italic><sub><italic>i</italic></sub>, <italic>x</italic><sub><italic>j</italic></sub>) is the kernel function that maps the data into a higher-dimensional space to make it linearly separable. K(x,z) = <italic>e</italic><sup>&#x02212;&#x003B3;||<italic>x</italic> &#x02212; <italic>z</italic>||<sup>2</sup></sup> is the radial basis function (RBF) kernel, where &#x003B3; controls the width of the Gaussian kernel and determines the influence of each training example. This kernelized formulation allows SVC to handle complex, non-linear decision boundaries efficiently while maintaining strong generalization performance.</p>
<disp-formula id="EQ20"><mml:math id="M26"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mo class="qopname">max</mml:mo></mml:mrow><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow></mml:msub><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>n</mml:mi></mml:mstyle></mml:mrow></mml:munderover></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mo>-</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mfrac><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>,</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:munder></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>&#x003B1;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>K</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mo>,</mml:mo></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(20)</label></disp-formula>
<p>A random forest classifier is a machine learning model that employs many decision trees to classify data, making its predictions based on the majority vote of the individual trees. It is known for its accuracy and stability. It performs this by selecting a random subset of features and training the decision trees on a different subset of the data. The final prediction is made through majority voting (<xref ref-type="bibr" rid="B19">Pal, 2005</xref>). <xref ref-type="disp-formula" rid="EQ21">Equation 21</xref> represents the prediction rule. Here, &#x00177;<sub><italic>RF</italic></sub> is the final prediction from the random forest, &#x00177;<sub><italic>tree</italic><sub><italic>k</italic></sub></sub> is the prediction from the k-th decision tree, and mode denotes the majority voting over all trees k = 1&#x02026;K. <xref ref-type="disp-formula" rid="EQ22">Equation 22</xref> represents the sampling correlation formula that is used to find how much the trees in the forest correlate in their predictions. Here, <italic>T</italic>(<italic>x</italic>; <italic>Z</italic>) is the output of a decision tree trained on the data subset Z with random feature sampling &#x003B8;. <italic>Var</italic><sub><italic>z</italic></sub> and <italic>E</italic><sub><italic>z</italic></sub> denote the variance and expectation over the dataset subsets.</p>
<disp-formula id="EQ21"><mml:math id="M27"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>R</mml:mi><mml:mi>F</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>t</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>e</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow></mml:msub></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>K</mml:mi></mml:mstyle></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(21)</label></disp-formula>
<disp-formula id="EQ22"><mml:math id="M28"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:mi>&#x003C1;</mml:mi><mml:mrow><mml:mo>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mfrac><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>V</mml:mi><mml:mi>a</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>r</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>E</mml:mi></mml:mstyle><mml:mo>&#x00398;</mml:mo></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo>;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>Z</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>]</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>V</mml:mi><mml:mi>a</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>r</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>E</mml:mi></mml:mstyle><mml:mo>&#x00398;</mml:mo></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo>;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>Z</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>]</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>E</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>z</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>V</mml:mi><mml:mi>a</mml:mi></mml:mstyle><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>r</mml:mi></mml:mstyle><mml:mo>&#x00398;</mml:mo></mml:msub><mml:mo stretchy='false'>[</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo>;</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>Z</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>]</mml:mo><mml:mo stretchy='false'>]</mml:mo></mml:mrow></mml:mfrac></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(22)</label></disp-formula>
<p>The gradient boosting classifier is an ensemble learning algorithm that constructs multiple decision trees sequentially, where each new tree corrects the errors of the previous ones. It reduces the loss function by optimizing residual errors using gradient descent. Gradient boosting combines weak learners to create a strong predictive model. <xref ref-type="disp-formula" rid="EQ23">Equation 23</xref> denotes the prediction function at iteration m. Here, <italic>F</italic><sub><italic>m</italic></sub>(<italic>x</italic>) is the updated model after the m-th iteration, <italic>h</italic><sub><italic>m</italic></sub>(<italic>x</italic>) is the weak learner added at stage mmm, which fits the negative gradient of the loss function, and &#x003B3;<sub><italic>m</italic></sub> is the step size or weight for the m-th learner.</p>
<disp-formula id="EQ23"><mml:math id="M29"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>F</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>F</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mstyle></mml:mrow></mml:msub><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x0002B;</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mi>&#x003B3;</mml:mi></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>h</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(23)</label></disp-formula>
<p>The KNN classifier is a machine learning algorithm that is non-parametric. It classifies an instance based on the majority class among its k closest neighbors in the feature space. The algorithm depends on distance metrics to compute similarity between points (<xref ref-type="bibr" rid="B7">Cunningham and Delany, 2021</xref>). <xref ref-type="disp-formula" rid="EQ25">Equation 25</xref> represents the prediction rule of KNN. Here, <italic>N</italic><sub><italic>k</italic></sub>(<italic>x</italic>) denotes the set of KNN of instance x, and <italic>y</italic><sub><italic>j</italic></sub> represents the label of the neighbor <italic>x</italic><sub><italic>j</italic></sub>. <xref ref-type="disp-formula" rid="EQ5">Equation 25</xref> provides the formula to calculate the distance between the points. Here, <italic>x</italic><sub><italic>i</italic></sub><italic>x</italic><sub><italic>j</italic></sub> are two points in a p-dimensional space.</p>
<disp-formula id="EQ24"><mml:math id="M30"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mstyle><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>K</mml:mi><mml:mi>N</mml:mi><mml:mi>N</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x02009;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>m</mml:mi><mml:mi>o</mml:mi><mml:mi>d</mml:mi><mml:mi>e</mml:mi></mml:mstyle><mml:mstyle displaystyle="true"><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mo>{</mml:mo><mml:mrow><mml:mo stretchy='false'>&#x0007B;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='true'>&#x0007C;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:msub><mml:mtext>&#x02009;</mml:mtext><mml:mo>&#x02208;</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>N</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mo>}</mml:mo></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow></mml:mstyle></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(24)</label></disp-formula>
<disp-formula id="EQ25"><mml:math id="M31"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mstyle mathvariant="bold-italic"><mml:mi>d</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mo>,</mml:mo></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:msqrt><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover accentunder="false" accent="false"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>p</mml:mi></mml:mstyle></mml:mrow></mml:munderover></mml:mstyle><mml:msup><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mstyle mathvariant="bold"><mml:mo>-</mml:mo></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>2</mml:mn></mml:mstyle></mml:mrow></mml:msup></mml:mrow></mml:msqrt></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(25)</label></disp-formula>
<p>A meta-learner using logistic regression with Lasso regularization is an advanced ensemble learning technique, where logistic regression serves as the final classifier in the meta-learning model. Lasso (L1) regularization is applied to enforce sparsity by shrinking less important coefficients to zero, effectively performing feature selection. <xref ref-type="disp-formula" rid="EQ26">Equation 26</xref> represents the objective function. Here, <italic>p</italic><sub><italic>i</italic></sub> = &#x003C3;(<italic>w</italic><sub>0</sub> &#x0002B; <italic>w</italic><sub>1</sub><italic>x</italic><sub>1</sub> &#x0002B; ... &#x0002B; <italic>w</italic><sub><italic>p</italic></sub><italic>x</italic><sub><italic>p</italic></sub>), Lasso penalty (||<italic>w</italic>||<sub>1</sub>||<italic>w</italic>||<sub>1</sub>) encourages sparsity, &#x00177;<sub><italic>i</italic></sub> is the prediction from base learners, &#x003C3; is the sigmoid function, w is the weight vector, ||<italic>w</italic>||<sub>1</sub> is the L1-norm of the weights, and &#x003BB; is the regularization strength.</p>
<disp-formula id="EQ26"><mml:math id="M32"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:mrow><mml:msub><mml:mrow><mml:mi>min</mml:mi></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle></mml:msub><mml:mtext>&#x02009;</mml:mtext><mml:mo>=</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo>&#x02212;</mml:mo><mml:mstyle displaystyle='true'><mml:munderover><mml:mo>&#x02211;</mml:mo><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>n</mml:mi></mml:mstyle></mml:munderover><mml:mrow><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mi>log</mml:mi><mml:mi>&#x003C3;</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msup><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle></mml:msup><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mo>&#x02212;</mml:mo><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>log</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle><mml:mo>&#x02212;</mml:mo><mml:mi>&#x003C3;</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msup><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>T</mml:mi></mml:mstyle></mml:msup><mml:msub><mml:mstyle mathvariant="bold-italic"><mml:mover accent='true'><mml:mi>y</mml:mi><mml:mo>&#x0005E;</mml:mo></mml:mover></mml:mstyle><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mstyle><mml:mo>+</mml:mo><mml:mtext>&#x02009;</mml:mtext><mml:mi>&#x003BB;</mml:mi><mml:msub><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mrow><mml:mrow><mml:mo>|</mml:mo><mml:mstyle mathvariant="bold-italic"><mml:mi>w</mml:mi></mml:mstyle><mml:mo>|</mml:mo></mml:mrow></mml:mrow><mml:mo>|</mml:mo></mml:mrow></mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mn>1</mml:mn></mml:mstyle></mml:msub></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(26)</label></disp-formula>
</sec>
<sec>
<label>2.2.5</label>
<title>Explainable AI and digital twins</title>
<sec>
<label>2.2.5.1</label>
<title>Grad-CAM</title>
<p>Grad-CAM produces class-discriminative visualizations by combining gradients and convolutional feature maps. For a target class c, the localization heatmap <inline-formula><mml:math id="M34"><mml:msubsup><mml:mrow><mml:mi>L</mml:mi></mml:mrow><mml:mrow><mml:mi>G</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mo>-</mml:mo><mml:mi>C</mml:mi><mml:mi>A</mml:mi><mml:mi>M</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> is computed, as seen in <xref ref-type="disp-formula" rid="EQ28">Equation 28</xref>.</p>
<disp-formula id="EQ27"><mml:math id="M35"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x0221D;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle></mml:mrow></mml:msubsup><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mfrac><mml:mrow><mml:mstyle mathvariant="bold"><mml:mn>1</mml:mn></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>Z</mml:mi></mml:mstyle></mml:mrow></mml:mfrac><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi></mml:mstyle></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:munder></mml:mstyle><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>A</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:msubsup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(27)</label></disp-formula>
<disp-formula id="EQ28"><mml:math id="M36"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>L</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>G</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mo>-</mml:mo><mml:mi>C</mml:mi><mml:mi>A</mml:mi><mml:mi>M</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle></mml:mrow></mml:msubsup><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo></mml:mstyle><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold-italic"><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>L</mml:mi><mml:mi>U</mml:mi></mml:mstyle><mml:mrow><mml:mo stretchy="true">(</mml:mo><mml:mrow><mml:mstyle displaystyle="true"><mml:munder class="msub"><mml:mrow><mml:mo>&#x02211;</mml:mo></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:munder></mml:mstyle><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mo>&#x0221D;</mml:mo></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle></mml:mrow></mml:msubsup><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>A</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>k</mml:mi></mml:mstyle></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="true">)</mml:mo></mml:mrow></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(28)</label></disp-formula>
<p>Here, <italic>A</italic><sup><italic>k</italic></sup> are the activations from the final convolutional layer, <inline-formula><mml:math id="M37"><mml:msubsup><mml:mrow><mml:mo>&#x0221D;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msubsup></mml:math></inline-formula> are the gradient weights, and Z normalizes the spatial dimensions.</p>
</sec>
<sec>
<label>2.2.5.2</label>
<title>Saliency map</title>
<p>Saliency maps detect influential pixels in input images by calculating gradients of class scores with respect to input pixels. For a class c, the saliency map S is derived using <xref ref-type="disp-formula" rid="EQ29">Equation 29</xref>, where <italic>y</italic><sup><italic>c</italic></sup> is the class score and <italic>x</italic><sub><italic>ij</italic></sub> are the input pixels. Such maps highlight regions that most strongly influence predictions, regardless of the model&#x00027;s architecture.</p>
<disp-formula id="EQ29"><mml:math id="M38"><mml:mtable class="eqnarray" columnalign="left"><mml:mtr><mml:mtd><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>S</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mtext>&#x000A0;</mml:mtext><mml:mstyle mathvariant="bold"><mml:mo>=</mml:mo><mml:mtext>&#x000A0;</mml:mtext></mml:mstyle><mml:mstyle displaystyle="true"><mml:munder><mml:mrow><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:munder></mml:mstyle><mml:mstyle displaystyle="true"><mml:mo>|</mml:mo><mml:mfrac><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msup><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>y</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>c</mml:mi></mml:mstyle></mml:mrow></mml:msup></mml:mrow><mml:mrow><mml:mi>&#x02202;</mml:mi><mml:msub><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>x</mml:mi></mml:mstyle></mml:mrow><mml:mrow><mml:mstyle mathvariant="bold-italic"><mml:mi>i</mml:mi><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub></mml:mrow></mml:mfrac><mml:mo>|</mml:mo></mml:mstyle></mml:mtd></mml:mtr></mml:mtable></mml:math><label>(29)</label></disp-formula>
</sec>
<sec>
<label>2.2.5.3</label>
<title>Digital twins</title>
<p>Digital twin is envisioned as an imaginary twin of the diagnostic model that reflects its behavior and decision path through real-time input&#x02013;output mapping. Leveraging Grad-CAM and saliency map visualizations, the digital twin allows clinicians and researchers to understand how the system detects suspicious areas in dermoscopic images. Each component of the pipeline&#x02014;the early activation block, deep feature extraction, feature refinement based on t-SNE, the meta-learning classifier, and the interpretability modules, such as Grad-CAM&#x02014;is a layer of the digital twin. This combination allows the system to reflect in real time and adapt much like a human expert would. This digital representation allows for real-time simulation and verification of the diagnostic process, providing insights into model behavior under different input conditions. It not only adds transparency but also serves as a feedback loop for model optimization by examining performance across different patient populations and image modalities. Embedding this notion in the visualization module guarantees not just correct predictions but also traceability and explainability, fostering trust in clinical environments for AI-supported diagnosis. The conclusion must be an explicit improvement: the digital twin provides interpretability, a real-time personal simulation, and transparency in the outcomes, allowing our model to evolve from a black-box classifier to a self-aware diagnostic assistant that respects the tenets of digital twinning in healthcare AI.</p>
</sec>
</sec>
</sec>
</sec>
<sec sec-type="results" id="s3">
<label>3</label>
<title>Results</title>
<p>The sample distribution of the training and test sets is shown in <xref ref-type="fig" rid="F6">Figure 6</xref>, highlighting any possible class imbalance. The dataset is split in a ratio of 80:20, with 360 images in the test set and 1,440 in the training set. <xref ref-type="fig" rid="F7">Figure 7</xref> shows the count of images in each class. Having a balanced dataset is important for fair model performance since models trained on imbalanced datasets may learn to favor the majority class.</p>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>Count of images of each class after train-test split.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0006.tif">
<alt-text>Bar chart titled "Benign vs Malignant Image Counts," showing categories of benign and malignant. Blue bars represent training images, with benign around 1400 and malignant around 800. Red bars represent test images, both about 200.</alt-text>
</graphic>
</fig>
<fig position="float" id="F7">
<label>Figure 7</label>
<caption><p>Total count of images of each class.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0007.tif">
<alt-text>Bar chart comparing the total number of benign and malignant images. The benign category, shown in blue, has around 1,750 images, while the malignant category, in red, has approximately 1,500 images.</alt-text>
</graphic>
</fig>
<p>The pre-trained model was trained on the ISIC dataset, which contains two classes: benign and malignant, as seen in <xref ref-type="fig" rid="F7">Figure 7</xref>. The model obtained a test accuracy of 90%, indicating that it is very effective in discriminating between the two classes. The plot of training and validation accuracy shown in <xref ref-type="fig" rid="F12">Figure 12</xref> illustrates how accuracy varies as the number of epochs increases. To obtain better feature representations learned by the model, we employed t-SNE, a non-linear dimensionality reduction technique. <xref ref-type="fig" rid="F8">Figures 8</xref>&#x02013;<xref ref-type="fig" rid="F11">11</xref> visualize how different layers separate the benign and malignant classes. The layer that was well clustered was selected for feature extraction, as it demonstrated high discriminative power, as shown in <xref ref-type="fig" rid="F12">Figure 12</xref>.</p>
<fig position="float" id="F8">
<label>Figure 8</label>
<caption><p>t-SNE visualization of the Conv_1.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0008.tif">
<alt-text>t-SNE scatter plot visualizing Conv_ 1 features with two classes: Class 0 in blue and Class 1 in orange. Data points are dispersed across t-SNE Dimensions 1 and 2, showing mixed clustering patterns.</alt-text>
</graphic>
</fig>
<fig position="float" id="F9">
<label>Figure 9</label>
<caption><p>t-SNE visualization of the Conv1_bn.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0009.tif">
<alt-text>t-SNE scatter plot visualizing the representation of two classes, Class 0 in blue and Class 1 in orange. The data points are spread across two dimensions, t-SNE Dimension 1 and t-SNE Dimension 2, with overlapping clusters.</alt-text>
</graphic>
</fig>
<fig position="float" id="F10">
<label>Figure 10</label>
<caption><p>t-SNE visualization of the global_average_pooling2d.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0010.tif">
<alt-text>Scatter plot showing a t-SNE visualization of global average pooling2d features with two classes. Blue dots represent Class 0, and orange dots represent Class 1, dispersed across t-SNE dimensions 1 and 2.</alt-text>
</graphic>
</fig>
<fig position="float" id="F11">
<label>Figure 11</label>
<caption><p>t-SNE visualization of the Conv7.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0011.tif">
<alt-text>Scatter plot showing t-SNE visualization of Conv7 with clusters of blue and orange dots representing two classes. Blue dots represent Class 0 and orange dots represent Class 1, spread across two dimensions labeled t-SNE Dimension 1 and t-SNE Dimension 2.</alt-text>
</graphic>
</fig>
<fig position="float" id="F12">
<label>Figure 12</label>
<caption><p>t-SNE visualization of the Conv1_relu.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0012.tif">
<alt-text>t-SNE scatter plot visualizing Conv1_ relu features, showing two classes: Class 0 in blue and Class 1 in orange. The points are distributed across two dimensions, revealing clustering patterns.</alt-text>
</graphic>
</fig>
<p>Following feature extraction from the most learnable layer, the extracted features were used to train a meta-learning classifier. The t-SNE visualization of the Conv1_relu layer in <xref ref-type="fig" rid="F12">Figure 12</xref> shows a moderate level of feature separation between the benign and malignant cases. <xref ref-type="fig" rid="F13">Figure 13</xref> illustrates the confusion matrices and ROC curves for each individual classifier, and <xref ref-type="fig" rid="F14">Figure 14</xref> shows the combined performance achieved using the proposed meta-learning classifier. Among all, the proposed classifier achieved the highest accuracy, clearly outperforming the individual models and demonstrating the effectiveness of the ensemble approach. <xref ref-type="fig" rid="F15">Figure 15</xref> shows the Line graph showing training and validation accuracy overtwenty epochs. In addition, the t-SNE visualization in <xref ref-type="fig" rid="F16">Figure 16</xref> provides a more optimal clustering of the two classes. From these visualizations, we can see the improved classification performance of the meta-learning classifier compared to the pre-trained model alone. <xref ref-type="fig" rid="F17">Figure 17</xref> shows the meta visualization of skin cancer analysis for the benign class. <xref ref-type="fig" rid="F18">Figure 18</xref> shows the Grad-CAM visualization and saliency map of malignant.</p>
<fig position="float" id="F13">
<label>Figure 13</label>
<caption><p>Classification report and ROC curve of the base models.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0013.tif">
<alt-text>The image contains four sets of confusion matrices paired with ROC curves. Each confusion matrix shows numbers of benign and malignant predictions against true labels, while each corresponding ROC curve displays true positive rates versus false positive rates for benign and malignant classes, indicating different AUC scores. AUC scores vary, with benign and malignant scores visible in each chart.</alt-text>
</graphic>
</fig>
<fig position="float" id="F14">
<label>Figure 14</label>
<caption><p>Classification report and ROC curve of the meta-learning classifier.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0014.tif">
<alt-text>Confusion matrix and ROC curve for a classification model. The confusion matrix shows 349 true positives, 11 false positives, 13 false negatives, and 267 true negatives for benign and malignant classes. The ROC curve displays two lines with an Area Under the Curve (AUC) of 0.96 for both benign and malignant classes, indicating high model performance.</alt-text>
</graphic>
</fig>
<fig position="float" id="F15">
<label>Figure 15</label>
<caption><p>Plot of training and validation accuracy.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0015.tif">
<alt-text>Line graph showing training and validation accuracy over twenty epochs. The red line represents training accuracy, reaching nearly one. The green line denotes validation accuracy, fluctuating between 0.65 and 0.90. The best epoch is marked at twenty.</alt-text>
</graphic>
</fig>
<fig position="float" id="F16">
<label>Figure 16</label>
<caption><p>t-SNE visualization of meta-learning classifier.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0016.tif">
<alt-text>t-SNE visualization of stacking classifier features showing clusters of data points in purple and yellow. The x-axis is labeled t-SNE Component 1, and the y-axis is labeled t-SNE Component 2. A color gradient from purple to yellow represents class labels, ranging from 0.0 to 1.0.</alt-text>
</graphic>
</fig>
<fig position="float" id="F17">
<label>Figure 17</label>
<caption><p>Grad-CAM visualization and saliency map of benign.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0017.tif">
<alt-text>Three images show skin lesion analysis. The first, labeled &#x0201C;Benign,&#x0201D; displays a mole on skin. The second, &#x0201C;Grad-CAM,&#x0201D; highlights key areas in red over the mole. The third, &#x0201C;Saliency Map,&#x0201D; uses blue tones to emphasize the mole&#x00027;s features.</alt-text>
</graphic>
</fig>
<fig position="float" id="F18">
<label>Figure 18</label>
<caption><p>Grad-CAM visualization and saliency map of malignant.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fcomp-07-1646311-g0018.tif">
<alt-text>Three side-by-side images of skin. The first shows a malignant skin lesion. The second is a Grad-CAM visualization highlighting areas of interest in red. The third is a saliency map with attention areas marked in lighter purple.</alt-text>
</graphic>
</fig>
<p>These findings shown in <xref ref-type="table" rid="T2">Table 2</xref> suggest that deep learning models work extremely well for skin cancer detection if trained with the ISIC dataset. The utilization of t-SNE visualization to identify the most informative features, followed by the application of a meta-learning classifier, further improved the performance. The fact that this process worked so well highlights the importance of feature selection in medical image classification using deep learning.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>Comparison with the proposed model.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>Accuracy</bold></th>
<th valign="top" align="center"><bold>Precision</bold></th>
<th valign="top" align="center"><bold>Recall</bold></th>
<th valign="top" align="center"><bold>Sensitivity</bold></th>
<th valign="top" align="center"><bold>Specificity</bold></th>
<th valign="top" align="center"><bold>F1-score</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">MobileNetV2 (<xref ref-type="bibr" rid="B31">Toga&#x000E7;ar et al., 2021</xref>)</td>
<td valign="top" align="center">95.27%</td>
<td valign="top" align="center">95.15%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">95.36%</td>
<td valign="top" align="center">95.15%</td>
<td valign="top" align="center">95.68%</td>
</tr>
<tr>
<td valign="top" align="left">ARDT-DenseNet (<xref ref-type="bibr" rid="B36">Wu et al., 2020</xref>)</td>
<td valign="top" align="center">87.80%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">88.70%</td>
<td valign="top" align="center">87.30%</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">ResNet50, Ensemble Classifiers (<xref ref-type="bibr" rid="B1">Akilandasowmya et al., 2024</xref>)</td>
<td valign="top" align="center">92.04%</td>
<td valign="top" align="center">91.57%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">92.47%</td>
<td valign="top" align="center">93.48%</td>
<td valign="top" align="center">91.49%</td>
</tr>
<tr>
<td valign="top" align="left">InceptionV3 (<xref ref-type="bibr" rid="B10">Gouda et al., 2022</xref>)</td>
<td valign="top" align="center">85.70%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Resnet50-Inception (<xref ref-type="bibr" rid="B10">Gouda et al., 2022</xref>)</td>
<td valign="top" align="center">84.10%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Resnet50 (<xref ref-type="bibr" rid="B10">Gouda et al., 2022</xref>)</td>
<td valign="top" align="center">83.60%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">CNN (<xref ref-type="bibr" rid="B12">Inthiyaz et al., 2023</xref>)</td>
<td valign="top" align="center">87.42%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">ConvNeXtSmall, EfficientNetV2B3 and EfficientNetV2S (<xref ref-type="bibr" rid="B26">Sethanan et al., 2023</xref>)</td>
<td valign="top" align="center">94.20%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">SqueezeNet (<xref ref-type="bibr" rid="B15">Kumar and Vanmathi, 2022</xref>)</td>
<td valign="top" align="center">91.90%</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">-</td>
<td valign="top" align="center">91.40%</td>
<td valign="top" align="center">91.20%</td>
<td valign="top" align="center">-</td>
</tr>
<tr>
<td valign="top" align="left">Proposed model</td>
<td valign="top" align="center">96.25%</td>
<td valign="top" align="center">96.04%</td>
<td valign="top" align="center">95.35%</td>
<td valign="top" align="center">95.35%</td>
<td valign="top" align="center">96.94%</td>
<td valign="top" align="center">95.69%</td>
</tr></tbody>
</table>
</table-wrap>
</sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>Skin cancer is a significant global health problem, and thus there is a need for effective and accurate diagnosis techniques. The suggested model integrates deep feature extraction with a meta-learning classifier to achieve enhanced classification performance. The model applies t-SNE visualization to identify the most informative learnable layers to reduce feature selection and computational complexity. The proposed methodology was tested with respect to standard deep learning-based models and was found to perform more efficiently without losing accuracy. The model displayed 96.25% accuracy, which made it a trustworthy tool for skin cancer classification. Such AI-based testing can provide an attractive alternative to manual diagnosis methods for dermatologists, providing early and accurate detection capabilities. A digital twin platform is integrated to virtually replicate the diagnostic workflow for continuous monitoring, optimization, and model validation within a safe environment. In addition, explainable AI methods are used to provide transparency by identifying the salient image regions that are driving the model&#x00027;s predictions. Future development may involve incorporating a wider range of dermatological data sources and implementing dynamic validation in live clinical settings to further enhance performance and accessibility. In our future work, we hope to broaden the boundaries of the proposed framework by integrating heterogeneous dermatology data sources, including clinical images, patient metadata, and histopathological data, to improve model generalizability and robustness. We will also aim to conduct real-time validations in live clinical environments to understand practicality, adaptability, and end-user&#x00027;s views and feedback on the real-world application of the diagnostic tool. Overall, a deeper investigation into automated feature selection, mobile-embedded device integration, and potential alternatives including continual learning may help develop the scalability and clinical viability of the system.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s5">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s6">
<title>Author contributions</title>
<p>PS: Writing &#x02013; original draft, Methodology. GS: Writing &#x02013; original draft, Software. SV: Writing &#x02013; original draft, Validation. YK: Writing &#x02013; original draft, Visualization. SS: Writing &#x02013; review &#x00026; editing, Funding acquisition, Conceptualization.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s8">
<title>Generative AI statement</title>
<p>The author(s) declare that no Gen AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="s9">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Akilandasowmya</surname> <given-names>G.</given-names></name> <name><surname>Nirmaladevi</surname> <given-names>G.</given-names></name> <name><surname>Suganthi</surname> <given-names>S. U.</given-names></name> <name><surname>Aishwariya</surname> <given-names>A.</given-names></name></person-group> (<year>2024</year>). <article-title>Skin cancer diagnosis: leveraging deep hidden features and ensemble classifiers for early detection and classification</article-title>. <source>Biomed. Signal Process Control</source> <volume>88</volume>:<fpage>105306</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.bspc.2023.105306</pub-id></mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Akter</surname> <given-names>M.</given-names></name> <name><surname>Khatun</surname> <given-names>R.</given-names></name> <name><surname>Talukder</surname> <given-names>M. A.</given-names></name> <name><surname>Islam</surname> <given-names>M. M.</given-names></name> <name><surname>Uddin</surname> <given-names>M. A.</given-names></name> <name><surname>Ahamed</surname> <given-names>M. K. U.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>An integrated deep learning model for skin cancer detection using hybrid feature fusion technique</article-title>. <source>Biomed. Mater. Devices</source> <volume>3</volume>, <fpage>1</fpage>&#x02013;<lpage>15</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s44174-024-00264-3</pub-id></mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alharbi</surname> <given-names>H.</given-names></name> <name><surname>Sampedro</surname> <given-names>G. A.</given-names></name> <name><surname>Juanatas</surname> <given-names>R. A.</given-names></name> <name><surname>Lim</surname> <given-names>S. J.</given-names></name></person-group> (<year>2024</year>). <article-title>Enhanced skin cancer diagnosis: a deep feature extraction-based framework for the multi-classification of skin cancer utilizing dermoscopy images</article-title>. <source>Front. Med.</source> <volume>11</volume>:<fpage>1495576</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmed.2024.1495576</pub-id><pub-id pub-id-type="pmid">39606634</pub-id></mixed-citation></ref>
<ref id="B4">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Amin</surname> <given-names>J.</given-names></name> <name><surname>Sharif</surname> <given-names>A.</given-names></name> <name><surname>Gul</surname> <given-names>N.</given-names></name> <name><surname>Anjum</surname> <given-names>M. A.</given-names></name> <name><surname>Nisar</surname> <given-names>M. W.</given-names></name> <name><surname>Azam</surname> <given-names>F.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Integrated design of deep features fusion for localization and classification of skin cancer</article-title>. <source>Pattern Recognit. Lett.</source> <volume>131</volume>, <fpage>63</fpage>&#x02013;<lpage>70</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.patrec.2019.11.042</pub-id></mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Brancaccio</surname> <given-names>G.</given-names></name> <name><surname>Balato</surname> <given-names>A.</given-names></name> <name><surname>Malvehy</surname> <given-names>J.</given-names></name> <name><surname>Puig</surname> <given-names>S.</given-names></name> <name><surname>Argenziano</surname> <given-names>G.</given-names></name> <name><surname>Kittler</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Artificial intelligence in skin cancer diagnosis: a reality check</article-title>. <source>J. Invest. Dermatol.</source> <volume>144</volume>, <fpage>492</fpage>&#x02013;<lpage>499</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jid.2023.10.004</pub-id><pub-id pub-id-type="pmid">37978982</pub-id></mixed-citation></ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Claret</surname> <given-names>S. A.</given-names></name> <name><surname>Dharmian</surname> <given-names>J. P.</given-names></name> <name><surname>Manokar</surname> <given-names>A. M.</given-names></name></person-group> (<year>2024</year>). <article-title>Artificial intelligence-driven enhanced skin cancer diagnosis: leveraging convolutional neural networks with discrete wavelet transformation</article-title>. <source>Egypt J. Med. Hum. Genet.</source> <volume>25</volume>:<fpage>50</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s43042-024-00522-5</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cunningham</surname> <given-names>P.</given-names></name> <name><surname>Delany</surname> <given-names>S.J.</given-names></name></person-group> (<year>2021</year>). <article-title>K-nearest neighbour classifiers-a tutorial</article-title>. <source>ACM Comput. Surv.</source> <volume>54</volume>, <fpage>1</fpage>&#x02013;<lpage>25</lpage>. doi: <pub-id pub-id-type="doi">10.1145/3459665</pub-id></mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ekmekyapar</surname> <given-names>T.</given-names></name> <name><surname>Ta&#x0015F;ci</surname> <given-names>B.</given-names></name></person-group> (<year>2023</year>). <article-title>Exemplar MobileNetV2-based artificial intelligence for robust and accurate diagnosis of multiple sclerosis</article-title>. <source>Diagnostics</source> <volume>13</volume>:<fpage>3030</fpage>. doi: <pub-id pub-id-type="doi">10.3390/diagnostics13193030</pub-id><pub-id pub-id-type="pmid">37835771</pub-id></mixed-citation></ref>
<ref id="B9">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Georgiou</surname> <given-names>T.</given-names></name> <name><surname>Liu</surname> <given-names>Y.</given-names></name> <name><surname>Chen</surname> <given-names>W.</given-names></name> <name><surname>Lew</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>A survey of traditional and deep learning-based feature descriptors for high dimensional data in computer vision</article-title>. <source>Int. J. Multimed. Inf. Retr.</source> <volume>9</volume>, <fpage>135</fpage>&#x02013;<lpage>170</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s13735-019-00183-w</pub-id></mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gouda</surname> <given-names>W.</given-names></name> <name><surname>Sama</surname> <given-names>N. U.</given-names></name> <name><surname>Al-Waakid</surname> <given-names>G.</given-names></name> <name><surname>Humayun</surname> <given-names>M.</given-names></name> <name><surname>Jhanjhi</surname> <given-names>N. Z.</given-names></name></person-group> (<year>2022</year>). <article-title>Detection of skin cancer based on skin lesion images using deep learning</article-title>. <source>Healthcare</source> <volume>10</volume>:<fpage>1183</fpage>. doi: <pub-id pub-id-type="doi">10.3390/healthcare10071183</pub-id><pub-id pub-id-type="pmid">35885710</pub-id></mixed-citation></ref>
<ref id="B11">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Himel</surname> <given-names>G. M. S.</given-names></name> <name><surname>Islam</surname> <given-names>M. M.</given-names></name> <name><surname>Al-Aff</surname> <given-names>K. A.</given-names></name> <name><surname>Karim</surname> <given-names>S. I.</given-names></name> <name><surname>Sikder</surname> <given-names>M. K. U.</given-names></name></person-group> (<year>2024</year>). <article-title>Skin cancer segmentation and classification using vision transformer for automatic analysis in dermatoscopy-based noninvasive digital system</article-title>. <source>Int. J. Biomed. Imaging</source> <volume>2024</volume>:<fpage>3022192</fpage>. doi: <pub-id pub-id-type="doi">10.1155/2024/3022192</pub-id><pub-id pub-id-type="pmid">38344227</pub-id></mixed-citation></ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Inthiyaz</surname> <given-names>S.</given-names></name> <name><surname>Altahan</surname> <given-names>B. R.</given-names></name> <name><surname>Ahammad</surname> <given-names>S. H.</given-names></name> <name><surname>Rajesh</surname> <given-names>V.</given-names></name> <name><surname>Kalangi</surname> <given-names>R. R.</given-names></name> <name><surname>Smirani</surname> <given-names>L. K.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Skin disease detection using deep learning</article-title>. <source>Adv. Eng. Softw.</source> <volume>175</volume>:<fpage>103361</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.advengsoft.2022.103361</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Itoo</surname> <given-names>N. N.</given-names></name> <name><surname>Garg</surname> <given-names>V. K.</given-names></name></person-group> (<year>2022</year>). <article-title>&#x0201C;Heart disease prediction using a stacked ensemble of supervised machine learning classifiers,&#x0201D;</article-title> in <source>Proceedings of the 2022 International Mobile and Embedded Technology Conference (MECON)</source> (<publisher-loc>Noida</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>599</fpage>&#x02013;<lpage>604</lpage>. doi: <pub-id pub-id-type="doi">10.1109/MECON53876.2022.9751883</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kanchana</surname> <given-names>K.</given-names></name> <name><surname>Kavitha</surname> <given-names>S.</given-names></name> <name><surname>Anoop</surname> <given-names>K. J.</given-names></name> <name><surname>Chinthamani</surname> <given-names>B.</given-names></name></person-group> (<year>2024</year>). <article-title>Enhancing skin cancer classification using efficient net B0-B7 through convolutional neural networks and transfer learning with patient-specific data</article-title>. <source>Asian Pac. J. Cancer Prev.</source> <volume>25</volume>:<fpage>1795</fpage>. doi: <pub-id pub-id-type="doi">10.31557/APJCP.2024.25.5.1795</pub-id><pub-id pub-id-type="pmid">38809652</pub-id></mixed-citation></ref>
<ref id="B15">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kumar</surname> <given-names>K. A.</given-names></name> <name><surname>Vanmathi</surname> <given-names>C.</given-names></name></person-group> (<year>2022</year>). <article-title>Optimization driven model and segmentation network for skin cancer detection</article-title>. <source>Comput. Electr. Eng.</source> <volume>103</volume>:<fpage>108359</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compeleceng.2022.108359</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Mahmud</surname> <given-names>F.</given-names></name> <name><surname>Mahfiz</surname> <given-names>M. M.</given-names></name> <name><surname>Kabir</surname> <given-names>M. Z. I.</given-names></name> <name><surname>Abdullah</surname> <given-names>Y.</given-names></name></person-group> (<year>2023</year>). <article-title>&#x0201C;An interpretable deep learning approach for skin cancer categorization,&#x0201D;</article-title> in <source>2023 26th International Conference on Computer and Information Technology (IC-CIT)</source> (<publisher-loc>Cox&#x00027;s Bazar</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x02013;<lpage>6</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ICCIT60459.2023.10508527</pub-id></mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Melarkode</surname> <given-names>N.</given-names></name> <name><surname>Srinivasan</surname> <given-names>K.</given-names></name> <name><surname>Qaisar</surname> <given-names>S. M.</given-names></name> <name><surname>Plawiak</surname> <given-names>P.</given-names></name></person-group> (<year>2023</year>). <article-title>AI-powered diagnosis of skin cancer: a contemporary review, open challenges and future research directions</article-title>. <source>Cancers</source> <volume>15</volume>:<fpage>1183</fpage>. doi: <pub-id pub-id-type="doi">10.3390/cancers15041183</pub-id><pub-id pub-id-type="pmid">36831525</pub-id></mixed-citation></ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Naqvi</surname> <given-names>M.</given-names></name> <name><surname>Gilani</surname> <given-names>S. Q.</given-names></name> <name><surname>Syed</surname> <given-names>T.</given-names></name> <name><surname>Marques</surname> <given-names>O.</given-names></name> <name><surname>Kim</surname> <given-names>H.C.</given-names></name></person-group> (<year>2023</year>). <article-title>Skin cancer detection using deep learning-a review</article-title>. <source>Diagnostics</source> <volume>13</volume>:<fpage>1911</fpage>. doi: <pub-id pub-id-type="doi">10.3390/diagnostics13111911</pub-id><pub-id pub-id-type="pmid">37296763</pub-id></mixed-citation></ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pal</surname> <given-names>M.</given-names></name></person-group> (<year>2005</year>). <article-title>Random Forest classifier for remote sensing classification</article-title>. <source>Int. J. Remote Sens.</source> <volume>26</volume>, <fpage>217</fpage>&#x02013;<lpage>222</lpage>. doi: <pub-id pub-id-type="doi">10.1080/01431160412331269698</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Prokhorenkova</surname> <given-names>L.</given-names></name> <name><surname>Gusev</surname> <given-names>G.</given-names></name> <name><surname>Vorobev</surname> <given-names>A.</given-names></name> <name><surname>Dorogush</surname> <given-names>A. V.</given-names></name> <name><surname>Gulin</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;CatBoost: unbiased boosting with categorical features,&#x0201D;</article-title> in <source>Advances in Neural Information Processing Systems (NeurIPS)</source> (<publisher-loc>Cambridge, MA</publisher-loc>: <publisher-name>The MIT Press</publisher-name>).</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ragab</surname> <given-names>M.</given-names></name> <name><surname>Alshehri</surname> <given-names>S.</given-names></name> <name><surname>Azim</surname> <given-names>G. A.</given-names></name> <name><surname>Aldawsari</surname> <given-names>H. M.</given-names></name> <name><surname>Noor</surname> <given-names>A.</given-names></name> <name><surname>Alyami</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>COVID-19 identification system using transfer learning technique with MobileNetV2 and chest X-ray images</article-title>. <source>Front. Public Health</source> <volume>10</volume>:<fpage>819156</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpubh.2022.819156</pub-id><pub-id pub-id-type="pmid">35309201</pub-id></mixed-citation></ref>
<ref id="B22">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raihan</surname> <given-names>M. J.</given-names></name> <name><surname>Khan</surname> <given-names>M. A. M.</given-names></name> <name><surname>Kee</surname> <given-names>S. H.</given-names></name> <name><surname>Nahid</surname> <given-names>A.A.</given-names></name></person-group> (<year>2023</year>). <article-title>Detection of the chronic kidney disease using XGBoost classifier and explaining the influence of the attributes on the model using SHAP</article-title>. <source>Sci. Rep.</source> <volume>13</volume>:<fpage>6263</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-023-33525-0</pub-id><pub-id pub-id-type="pmid">37069256</pub-id></mixed-citation></ref>
<ref id="B23">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Salinas</surname> <given-names>M. P.</given-names></name> <name><surname>Sep&#x000FA;lveda</surname> <given-names>J.</given-names></name> <name><surname>Hidalgo</surname> <given-names>L.</given-names></name> <name><surname>Peirano</surname> <given-names>D.</given-names></name> <name><surname>Morel</surname> <given-names>M.</given-names></name> <name><surname>Uribe</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>A systematic review and meta-analysis of artificial intelligence versus clinicians for skin cancer diagnosis</article-title>. <source>NPJ Digit. Med.</source> <volume>7</volume>:<fpage>125</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41746-024-01103-x</pub-id><pub-id pub-id-type="pmid">38744955</pub-id></mixed-citation></ref>
<ref id="B24">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Sandler</surname> <given-names>M.</given-names></name> <name><surname>Howard</surname> <given-names>A.</given-names></name> <name><surname>Zhu</surname> <given-names>M.</given-names></name> <name><surname>Zhmoginov</surname> <given-names>A.</given-names></name> <name><surname>Chen</surname> <given-names>L. C.</given-names></name></person-group> (<year>2018</year>). <article-title>&#x0201C;Mobilenetv2: inverted residuals and linear bottlenecks,&#x0201D;</article-title> in <source>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source> (<publisher-loc>Salt Lake City, UT</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>4510</fpage>&#x02013;<lpage>4523</lpage>. doi: <pub-id pub-id-type="doi">10.1109/CVPR.2018.00474</pub-id><pub-id pub-id-type="pmid">39300076</pub-id></mixed-citation></ref>
<ref id="B25">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sawal</surname> <given-names>S.</given-names></name> <name><surname>Kathait</surname> <given-names>S.S.</given-names></name></person-group> (<year>2023</year>). <article-title>MobileNetV2: transfer learning for elephant detection</article-title>. <source>Int. J. Comput. Appl.</source> <volume>975</volume>, <fpage>59</fpage>&#x02013;<lpage>65</lpage>. doi: <pub-id pub-id-type="doi">10.5120/ijca2025924403</pub-id><pub-id pub-id-type="pmid">7911558</pub-id></mixed-citation></ref>
<ref id="B26">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sethanan</surname> <given-names>K.</given-names></name> <name><surname>Pitakaso</surname> <given-names>R.</given-names></name> <name><surname>Srichok</surname> <given-names>T.</given-names></name> <name><surname>Khonjun</surname> <given-names>S.</given-names></name> <name><surname>Thannipat</surname> <given-names>P.</given-names></name> <name><surname>Wanram</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Double AMIS-ensemble deep learning for skin cancer classification</article-title>. <source>Expert Syst. Appl.</source> <volume>234</volume>:<fpage>121047</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.eswa.2023.121047</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Surya</surname> <given-names>A.</given-names></name> <name><surname>Shah</surname> <given-names>A.</given-names></name> <name><surname>Kabore</surname> <given-names>J.</given-names></name> <name><surname>Sasikumar</surname> <given-names>S.</given-names></name></person-group> (<year>2023</year>). <article-title>Enhanced breast cancer tumor classification using MobileNetV2: a detailed exploration on image intensity, error mitigation, and streamlit-driven real-time deployment</article-title>. <source>arXiv</source>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.2312.03020</pub-id></mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tai</surname> <given-names>C. E. A.</given-names></name> <name><surname>Janes</surname> <given-names>E.</given-names></name> <name><surname>Czarnecki</surname> <given-names>C.</given-names></name> <name><surname>Wong</surname> <given-names>A.</given-names></name></person-group> (<year>2024</year>). <article-title>Double-condensing attention condenser: leveraging attention in deep learning to detect skin cancer from skin lesion images</article-title>. <source>Sensors</source> <volume>24</volume>:<fpage>7231</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s24227231</pub-id><pub-id pub-id-type="pmid">39599008</pub-id></mixed-citation></ref>
<ref id="B29">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thompson</surname> <given-names>N. C.</given-names></name> <name><surname>Greenewald</surname> <given-names>K.</given-names></name> <name><surname>Lee</surname> <given-names>K.</given-names></name> <name><surname>Manso</surname> <given-names>G. F.</given-names></name></person-group> (<year>2020</year>). <article-title>The computational limits of deep learning</article-title>. <source>arXiv</source>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.2007.05558</pub-id></mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tika Adilah</surname> <given-names>M.</given-names></name> <name><surname>Kristiyanti</surname> <given-names>D. A.</given-names></name></person-group> (<year>2023</year>). <article-title>Implementation of transfer learning MobileNetV2 architecture for identification of potato leaf disease</article-title>. <source>J. Theor. Appl. Inf. Technol.</source> <volume>101</volume>, <fpage>6273</fpage>&#x02013;<lpage>6285</lpage>.</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Toga&#x000E7;ar</surname> <given-names>M.</given-names></name> <name><surname>C&#x000F6;mert</surname> <given-names>Z.</given-names></name> <name><surname>Ergen</surname> <given-names>B.</given-names></name></person-group> (<year>2021</year>). <article-title>Intelligent skin cancer detection applyingautoencoder, MobileNetV2 and spiking neural networks</article-title>. <source>Chaos Solitons Fractals</source> <volume>144</volume>:<fpage>110714</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.chaos.2021.110714</pub-id></mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Van der Maaten</surname> <given-names>L.</given-names></name> <name><surname>Hinton</surname> <given-names>G.</given-names></name></person-group> (<year>2008</year>). <article-title>Visualizing data using t-SNE</article-title>. <source>J. Mach. Learn Res.</source> <volume>9</volume>, <fpage>2579</fpage>&#x02013;<lpage>2605</lpage>.</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Villa-Pulgarin</surname> <given-names>J. P.</given-names></name> <name><surname>Ruales-Torres</surname> <given-names>A. A.</given-names></name> <name><surname>Arias-Garzon</surname> <given-names>D.</given-names></name> <name><surname>Bravo-Ortiz</surname> <given-names>M. A.</given-names></name> <name><surname>Arteaga-Arteaga</surname> <given-names>H. B.</given-names></name> <name><surname>Mora-Rubio</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Optimized convolutional neural network models for skin lesion classification</article-title>. <source>Comput. Mater. Continua.</source> <volume>70</volume>, <fpage>2131</fpage>&#x02013;<lpage>2148</lpage>. doi: <pub-id pub-id-type="doi">10.32604/cmc.2022.019529</pub-id></mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>M.</given-names></name> <name><surname>Gao</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name></person-group> (<year>2025</year>). <article-title>Recent global patterns in skin cancer incidence, mortality, and prevalence</article-title>. <source>Chin. Med. J.</source> <volume>138</volume>, <fpage>185</fpage>&#x02013;<lpage>192</lpage>. doi: <pub-id pub-id-type="doi">10.1097/CM9.0000000000003416</pub-id><pub-id pub-id-type="pmid">39682020</pub-id></mixed-citation></ref>
<ref id="B35">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wei</surname> <given-names>M. L.</given-names></name> <name><surname>Tada</surname> <given-names>M.</given-names></name> <name><surname>So</surname> <given-names>A.</given-names></name> <name><surname>Torres</surname> <given-names>R.</given-names></name></person-group> (<year>2024</year>). <article-title>Artificial intelligence and skin cancer</article-title>. <source>Front. Med.</source> <volume>11</volume>:<fpage>1331895</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmed.2024.1331895</pub-id><pub-id pub-id-type="pmid">38566925</pub-id></mixed-citation></ref>
<ref id="B36">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Hu</surname> <given-names>W.</given-names></name> <name><surname>Wen</surname> <given-names>Y.</given-names></name> <name><surname>Tu</surname> <given-names>W.</given-names></name> <name><surname>Liu</surname> <given-names>X.</given-names></name></person-group> (<year>2020</year>). <article-title>Skin lesion classification using densely connected convolutional networks with attention residual learning</article-title>. <source>Sensors</source> <volume>20</volume>:<fpage>7080</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s20247080</pub-id><pub-id pub-id-type="pmid">33321864</pub-id></mixed-citation></ref>
<ref id="B37">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname> <given-names>L.</given-names></name> <name><surname>Mohammadi</surname> <given-names>M.</given-names></name></person-group> (<year>2024</year>). <article-title>Brain tumor diagnosis from MRI based on Mobilenetv2 optimized by contracted fox optimization algorithm</article-title>. <source>Heliyon</source> <volume>10</volume>:<fpage>e23866</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.heliyon.2023.e23866</pub-id><pub-id pub-id-type="pmid">38223737</pub-id></mixed-citation></ref>
<ref id="B38">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yuan</surname> <given-names>W.</given-names></name> <name><surname>Du</surname> <given-names>Z.</given-names></name> <name><surname>Han</surname> <given-names>S.</given-names></name></person-group> (<year>2024</year>). <article-title>Semi-supervised skin cancer diagnosis based on self-feedback threshold focal learning</article-title>. <source>Discov. Oncol.</source> <volume>15</volume>:<fpage>180</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s12672-024-01043-8</pub-id><pub-id pub-id-type="pmid">38776027</pub-id></mixed-citation></ref>
<ref id="B39">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>N.</given-names></name> <name><surname>Cai</surname> <given-names>Y. X.</given-names></name> <name><surname>Wang</surname> <given-names>Y. Y.</given-names></name> <name><surname>Tian</surname> <given-names>Y. T.</given-names></name> <name><surname>Wang</surname> <given-names>X. L.</given-names></name> <name><surname>Badami</surname> <given-names>B.</given-names></name></person-group> (<year>2020</year>). <article-title>Skin cancer diagnosis based on optimized convolutional neural network</article-title>. <source>Artif. Intell. Med.</source> <volume>102</volume>:<fpage>101756</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.artmed.2019.101756</pub-id><pub-id pub-id-type="pmid">31980095</pub-id></mixed-citation></ref>
<ref id="B40">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname> <given-names>Q.</given-names></name> <name><surname>Zhuang</surname> <given-names>H.</given-names></name> <name><surname>Zhao</surname> <given-names>M.</given-names></name> <name><surname>Xu</surname> <given-names>S.</given-names></name> <name><surname>Meng</surname> <given-names>R.</given-names></name></person-group> (<year>2024</year>). <article-title>A study on expression recognition based on improved mobilenetV2 network</article-title>. <source>Sci. Rep</source>. <volume>14</volume>:<fpage>8121</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-58736-x</pub-id><pub-id pub-id-type="pmid">38582772</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn id="fn0001" fn-type="custom" custom-type="edited-by">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1645388/overview">Kannimuthu Subramanian</ext-link>, Karpagam College of Engineering, India</p>
</fn>
<fn id="fn0002" fn-type="custom" custom-type="reviewed-by">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3069407/overview">Danilo Pelusi</ext-link>, University of Teramo, Italy; <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3106819/overview">Surya Sumarni Hussein</ext-link>, Universiti Teknologi MARA, Malaysia</p>
</fn>
</fn-group>
</back>
</article>