<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2025.1668839</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>PotatoLeafNet: two-stage convolutional neural networks for effective Potato Leaf disease identification and classification</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Bhavani</surname><given-names>Girigula Durga</given-names></name>
<xref ref-type="aff" rid="aff1"></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3313496"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Chalapathi</surname><given-names>Mukkoti Maruthi Venkata</given-names></name>
<xref ref-type="aff" rid="aff1"></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3129410"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
</contrib>
</contrib-group>
<aff id="aff1"><institution>School of Computer Science and Engineering (SCOPE), VIT-AP University</institution>, <city>Amaravati, Andhra Pradesh</city>, <country country="in">India</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Mukkoti Maruthi Venkata Chalapathi, <email xlink:href="mailto:mmv.chalapathi@gmail.com">mmv.chalapathi@gmail.com</email>; <email xlink:href="mailto:mmv.chalapthi@vitap.ac.in">mmv.chalapthi@vitap.ac.in</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-12">
<day>12</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>8</volume>
<elocation-id>1668839</elocation-id>
<history>
<date date-type="received">
<day>21</day>
<month>07</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>26</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>03</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Bhavani and Chalapathi.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Bhavani and Chalapathi</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-12">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Potato foliar diseases, particularly early and late blight, pose a serious threat to yield and food security, yet reliable visual recognition remains challenging due to cultivar heterogeneity, variable symptom expression, and acquisition noise in field-like imagery. To address these issues, we propose PotatoLeafNet, a two-stage deep learning framework that combines a fixed-sequence image-augmentation pipeline with a compact, task-optimized 11-layer convolutional neural network (CNN) using 3&#x202F;&#x00D7;&#x202F;3 kernels for robust, data-efficient classification of potato leaf conditions (Healthy, Early Blight, Late Blight).</p>
</sec>
<sec>
<title>Methods</title>
<p>We construct a dataset of 4,072 labeled potato leaf images from the PlantVillage-Potato subset and standardize all inputs to 224&#x202F;&#x00D7;&#x202F;224 RGB tensors with pixel intensities normalized to [0,1]. A balanced, fixed-order augmentation policy&#x2014;comprising rotation, translation, shear, zoom, horizontal flipping, brightness adjustment, and channel jitter&#x2014;is applied exclusively to the training split, increasing it to 6,000 images (2,000 per class) while keeping the validation and test sets free of synthetic samples. The second stage consists of an 11-layer CNN implemented in TensorFlow/Keras and trained with categorical cross-entropy loss and the Adam optimizer under a unified training and evaluation protocol. Performance is benchmarked against strong CNN and hybrid baselines, including ResNet-50&#x202F;+&#x202F;VGG-16, VGG-16&#x202F;+&#x202F;MobileNetV2, MobileNetV2, and Inception-V3.</p>
</sec>
<sec>
<title>Results</title>
<p>On the PlantVillage-Potato test set, PotatoLeafNet achieves 98.52% accuracy, 98.67% macro-precision, 99.67% macro-recall, 99.16% macro-F1, and 1.00 macro-AUC, outperforming all baseline models under identical preprocessing and training conditions. In particular, PotatoLeafNet surpasses ResNet-50 + VGG-16 (97.10% accuracy, AUC 0.98), VGG-16&#x202F;+&#x202F;MobileNetV2 (94.80% accuracy, AUC 0.93), MobileNetV2 (93.20% accuracy, AUC 0.92), and Inception-V3 (92.50% accuracy, AUC 0.91). Short 10-epoch runs yield stable convergence (training accuracy 88.22%, validation accuracy 86.91%, test accuracy 88.15%), indicating efficient learning from the augmented distribution.</p>
</sec>
<sec>
<title>Discussion</title>
<p>The results demonstrate that explicitly coupling a fixed sequential augmentation stage with a lightweight 3&#x00D7;3-kernel CNN enables high tri-class accuracy, strong recall for disease classes, and improved generalization relative to deeper or fused architectures, without incurring substantial computational cost. By emphasizing disease-relevant structure while limiting overfitting, PotatoLeafNet provides a practical and resource-efficient solution for automated screening of potato leaf health in real-world agronomic settings, supporting timely and data-driven disease management.</p>
</sec>
</abstract>
<kwd-group>
<kwd>potato leaf diseases</kwd>
<kwd>convolutional neural networks</kwd>
<kwd>dual CNN</kwd>
<kwd>sequential image augmentation</kwd>
<kwd>early blight</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="11"/>
<table-count count="7"/>
<equation-count count="11"/>
<ref-count count="49"/>
<page-count count="18"/>
<word-count count="11263"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>AI in Food, Agriculture and Water</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>The potato (<italic>Solanum tuberosum</italic>) is a staple crop and a vital source of calories and micronutrients for millions of people worldwide (<xref ref-type="bibr" rid="ref43">Sunjoyo and Nugroho, 2022</xref>). Yields, however, are highly vulnerable to foliar pathologies most notably early blight and late blight which inflict substantial economic losses and threaten local food security when outbreaks go undetected or unmanaged (<xref ref-type="bibr" rid="ref18">Hou et al., 2021</xref>). Early, reliable diagnosis is therefore essential for timely intervention, yet traditional field scouting by experts is labor-intensive, subjective, and difficult to scale across heterogeneous environments and planting cycles. These practical constraints motivate automated systems that deliver accurate, consistent, and rapid decisions directly from visual evidence. Convolutional neural networks (CNNs) have reshaped image understanding by learning hierarchical representations from data and have already shown strong performance in plant disease recognition tasks (<xref ref-type="bibr" rid="ref46">Tugrul et al., 2022</xref>). Despite this promise, potato leaf disease classification remains challenging in real deployments. Symptoms vary with cultivar, phenological stage, and stress conditions; image capture occurs under fluctuating illumination, background clutter, motion blur, and sensor noise; and publicly available datasets are often limited in size and balance across classes. Moreover, many prior approaches emphasize single-disease detection rather than precise multi-class discrimination among healthy leaves and the major disease categories (early blight and late blight) required for agronomic decision-making (<xref ref-type="bibr" rid="ref4">Alhammad et al., 2025</xref>). These factors collectively degrade generalization and complicate robust deployment on resource-constrained devices.</p>
<p>To address these limitations, we propose PotatoLeafNet, a two-stage convolutional framework for potato leaf disease detection and classification that explicitly couples data diversification with a compact, task-optimized classifier. Stage 1 performs sequential image augmentation including rotations, scalings, flips, and related geometric and photometric transforms to expand the training distribution and encode invariances that mirror field variability (<xref ref-type="bibr" rid="ref30">Potato Leaf Disease Dataset, 2025</xref>). By structuring augmentation as a dedicated stage, the pipeline intentionally exposes the learner to controlled perturbations that emulate acquisition noise and viewpoint change, thereby improving robustness without inflating model capacity. Stage 2 is a lightweight CNN tailored for potato leaves: convolutional blocks with ReLU activations and 3&#x202F;&#x00D7;&#x202F;3 kernels extract localized texture and lesion-edge cues; max-pooling progressively reduces spatial resolution while preserving salient patterns; global average pooling compacts feature maps to mitigate overfitting; and a fully connected head (a 128-unit ReLU layer followed by a softmax) produces calibrated class probabilities for healthy, early blight, and late blight. This design emphasizes parameter efficiency and computational tractability while retaining discriminative power under real-world noise. Our evaluation plan reflects these deployment goals. We train and test PotatoLeafNet on a diversified collection of potato leaf images spanning healthy, early blight, and late blight categories (<xref ref-type="bibr" rid="ref37">Sangar and Rajasekar, 2025</xref>). Performance is assessed using standard metrics accuracy, precision, recall, and F1-score to quantify both overall correctness and class-wise reliability. We further benchmark against contemporary CNN-based methods to examine accuracy efficiency trade-offs and to determine whether an explicit augmentation stage coupled with a compact classifier offers practical advantages over monolithic architectures. In addition, we analyze error modes to illuminate failure cases (e.g., tiny lesions with blurred boundaries, confounding background textures), informing future improvements to both model and data regimen. The significant contributions to the research are:</p><list list-type="bullet">
<list-item>
<p>Introduced a novel two-stage convolutional neural network architecture, PotatoLeafNet, specifically optimized for high accuracy in detecting and classifying potato leaf diseases, addressing limitations in existing models</p>
</list-item>
<list-item>
<p>Implemented advanced sequential image augmentation techniques within a CNN framework to significantly enhance the model&#x2019;s ability to generalize across diverse and unseen environmental conditions, a step beyond traditional augmentation practices.</p>
</list-item>
<list-item>
<p>Conducted a rigorous evaluation of the PotatoLeafNet model using an extensive dataset that includes a balanced representation of Healthy, Early Blight, and Late Blight potato leaf images, ensuring robust testing against varied disease manifestations.</p>
</list-item>
<list-item>
<p>Demonstrated superior performance of the PotatoLeafNet model through a comparative analysis with existing state-of-the-art models, highlighting advancements in accuracy and computational efficiency.</p>
</list-item>
</list>
<p>The study introduces a two-stage CNN-based potato leaf disease detection and classification method. Deep learning and image augmentation increase illness detection using this method.</p>
</sec>
<sec id="sec2">
<label>2</label>
<title>Basic preliminaries and related research work</title>
<p>Leaf diseases are a common problem in plants and crops, and they can cause significant damage to both the yield and quality of the harvest various factors cause leaf maladies including fungi, bacteria, viruses, and environmental stressors. Depending on the specific disease and the plant species affected, the symptoms of leaf diseases can vary greatly. Some common symptoms of leaf diseases include discoloration, spotting, wilting, deformity, and defoliation (<xref ref-type="bibr" rid="ref1">Afzaal et al., 2021</xref>). The entire plant may sometimes be involved, leading to stunted growth, reduced yield, or even death. Detecting and classifying diseases is significant for the control of the conditions. It can be done using various techniques such as visual inspection, laboratory analysis, and remote sensing. Recently, methods for automatically identifying and classifying leaf diseases using images of plant leaves have also been developed utilizing machine learning and computer vision techniques.</p>
<sec id="sec3">
<label>2.1</label>
<title>Types of potato leaf diseases</title>
<p>Potato foliage is vulnerable to a spectrum of pathogens with markedly different epidemiologies and symptomatology. Late blight (<italic>Phytophthora infestans</italic>), a fast-spreading oomycete disease, remains the most devastating, initiating water-soaked lesions that rapidly coalesce into necrotic brown areas with a characteristic grayish, downy sporulation under humid conditions, and frequently extending to tubers (<xref ref-type="bibr" rid="ref19">Jafar et al., 2024</xref>). Early blight (Alternaria solani) typically emerges on older leaves as discrete dark lesions that enlarge with concentric &#x201C;target-spot&#x201D; rings, progressing to chlorosis and premature defoliation in warm, humid environments (<xref ref-type="bibr" rid="ref29">Potato Disease Types, 2025</xref>). Viral diseases such as Potato virus Y (PVY) and Potato leafroll virus (PLRV), both primarily aphid-borne, induce mosaic mottling, leaf curling, and canopy yellowing (PVY), or the diagnostic upward rolling and brittle texture of leaves (PLRV), with attendant losses in yield and tuber quality that depend on cultivar and viral strain (<xref ref-type="bibr" rid="ref25">Liu and Wang, 2021</xref>). Soil and vascular-invading fungi drive wilt syndromes early dying dominated by Verticillium dahliae and late dying associated with Fusarium spp. leading to progressive wilting, chlorosis, and necrosis that culminate in significant productivity declines. Bacterial threats such as bacterial ring rot cause leaf yellowing, wilting, and vascular browning with corky ring formation, and can persist in soils and on equipment, complicating eradication efforts. While multiple leaf diseases impair crop performance, early and late blight are generally the most consequential for field management decisions; representative phenotypes for Early blight, Late blight, and Healthy leaves are shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Sample images <bold>(a)</bold> potato early blight, <bold>(b)</bold> potato late blight, <bold>(c)</bold> potato healthy leaf.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Three leaves are shown. Leaf a is yellowed with dark spots, indicating disease. Leaf b has a large brown patch, also showing signs of disease. Leaf c is healthy and green.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Literature review on potato leaf disease detection and classification</title>
<p>This potato leaf disease detection and categorization literature review summarizes current methodologies. <xref ref-type="bibr" rid="ref32">Radwan et al. (2025)</xref> developed a weather-driven pipeline for early and late blight using K-means, PCA, copula analysis and multiple classifiers, with binary Greylag Goose Optimization for feature selection. On a 4,000-record meteorological dataset, the best MLP with selected features reached 98.3% accuracy. This tabular risk-forecasting setup complements image-based screening. <xref ref-type="bibr" rid="ref9">Chen and Liu (2025)</xref> introduced CBSNet with Channel Reconstruction Multi-Scale Convolution and Spatial Triple Attention, plus a Bat&#x2013;Lion training strategy for robustness. On a self-built potato leaf image set, CBSNet achieved 92.04% accuracy and 91.58% precision, extracting tiny lesions and blurred edges effectively. <xref ref-type="bibr" rid="ref10">Dey et al. (2025)</xref> proposed a lightweight CNN tailored for real-time classification, reducing depth and parameters to 204,227 while preserving accuracy on high-resolution potato leaf images. The model attained 98.6% test accuracy and class-wise precision of 0.99 (early blight), 0.98 (late blight), 1.00 (healthy), outperforming VGG16, AlexNet, and ResNet50. <xref ref-type="bibr" rid="ref41">Sinamenye et al. (2025)</xref> fused EfficientNetV2-B3 with a Vision Transformer to couple local convolutional features with global context. Trained on the Potato Leaf Disease Dataset reflecting field variability, the hybrid reached 85.06% accuracy, improving prior results by 11.43 points. <xref ref-type="bibr" rid="ref36">Salihu et al. (2025)</xref> built a CNN trained with Adam, using scaling, augmentation, and normalization over a curated set of healthy, early blight, and late blight images. The model achieved 96.88% accuracy, with class metrics including precision 0.76, recall 0.93, F1 0.84 for healthy and near-perfect scores for blight classes. <xref ref-type="bibr" rid="ref3">Ala&#x2019;a (2025)</xref> extracted generalized Jones polynomial texture features and classified with SVM on Plant Village potato images. The GJP-SVM pipeline preprocessing, feature extraction, dimensionality reduction, classification reached 98.45% accuracy, showing strong performance from hand-crafted descriptors. <xref ref-type="bibr" rid="ref49">Zhang et al. (2025)</xref> benchmarked VGG16, MobileNetV1, ResNet50, and ViT, then proposed VGG16S with global average pooling, CBAM attention, and Leaky ReLU to shrink parameters to 15&#x202F;M. After response-surface hyperparameter tuning, VGG16S achieved 97.87% test accuracy and generalized well on public sets. <xref ref-type="bibr" rid="ref20">Kaur et al. (2025)</xref> presented PotConvNet, a compact CNN trained on two potato image datasets with resizing, normalization, augmentation, and fixed splits. Reported accuracies were 99.76% (Dataset 1) and 97.78% (Dataset 2), validated by F1, precision, recall, Cohen&#x2019;s kappa, and ROC AUC. <xref ref-type="bibr" rid="ref27">Nur et al. (2025)</xref> optimized Inception V3 via transfer learning and targeted fine-tuning of terminal layers on a domain-specific potato leaf set. The approach yielded 97.78% accuracy with precision 98%, recall 98%, F1 98%, offering strong performance with practical efficiency. <xref ref-type="bibr" rid="ref40">Shah et al. (2025)</xref> introduced PLDC-Net, using EfficientNet-B1 as a backbone, fine-tuned with dense layers and an SVM output head; data balancing and augmentation were emphasized. On an unseen test set, the model achieved 98.39% average accuracy, providing a reliable transfer-learning baseline for multi-disease identification. We diagnose and categorize potato plant diseases. Various studies on diagnosing and categorizing potato plant diseases may be found in the literature on potato leaf disease classification and detection (<xref ref-type="bibr" rid="ref12">Fuentes et al., 2017</xref>). CNN and other deep learning approaches have shown promise for automating the detection and classification process, reducing the need for human expertise-several CNN architectures, transfer learning, feature extraction, and ensemble methods to improve accuracy and robustness (<xref ref-type="bibr" rid="ref13">Geetharamani and Pandian, 2019</xref>). The study (<xref ref-type="bibr" rid="ref2">Ahmed et al., 2025</xref>) suggested a deep CNN model to identify outstanding and ailing foliage across crops. They trained their model using the Plant Village dataset, which includes photos of diseased and healthy leaves and the backgrounds of 38 distinct crop kinds. However, they did not zero in on potato crop illnesses, and the data used to prepare the algorithm in the United States and Switzerland missed Pakistan-endemic infections on potato leaves.</p>
<p>Despite having little data, the scientists used deep learning, specifically CNNs, to identify potato illnesses (<xref ref-type="bibr" rid="ref22">Lee et al., 2020</xref>). A CNN model was created (<xref ref-type="bibr" rid="ref6">Awal et al., 2019</xref>) to distinguish between healthy potato leaves and those affected by early or late blight. They also used the regionally targeted Plant Village dataset in their research (<xref ref-type="bibr" rid="ref21">Khalifa et al., 2021</xref>). In this research, we looked at how well deep learning methods and convolutional neural networks, in particular, might do at identifying diseases on potato leaves. The authors trained a CNN network using a collection of photos of diseased potato leaves. The success of the suggested method in illness detection demonstrates the promise of deep learning for this application area. According to <xref ref-type="bibr" rid="ref14">Ghosal et al. (2019)</xref>, the CNN model has the ability to differentiate between various plant classes. This study, <xref ref-type="bibr" rid="ref34">Rathod et al. (2020)</xref> used deep learning to detect potato leaf blight early. The authors trained a CNN architecture to interpret potato leaf images. The model&#x2019;s early blight detection highlights deep learning&#x2019;s potential for potato leaf diseases. The authors examined deep learning and transfer learning for potato disease diagnosis (<xref ref-type="bibr" rid="ref24">Liang et al., 2019</xref>). Using potato leaf images, the authors updated VGG16, a pre-trained CNN model. Pre-trained CNN models with transfer learning were useful in potato disease detection. A network for identifying and assessing plant diseases was demonstrated in <xref ref-type="bibr" rid="ref11">Ferentinos (2018)</xref>. To distinguish between healthy and diseased plants from photographs of their leaves (<xref ref-type="bibr" rid="ref35">Rozaqi et al., 2020</xref>; <xref ref-type="bibr" rid="ref38">Sanjeev et al., 2020</xref>) looked at many deep-learning architectures. These included AlexNet, Overfeat, AlexNetOWTBn, VGG, and GoogLeNet. The authors applied transfer learning to the PlantVillage dataset to identify local agricultural diseases. We developed a CNN model to detect potato plants with early, late, or robust blight. We trained the model using PlantVillage, disease data. FFNNs can distinguish between early, late, and healthy foliage (<xref ref-type="bibr" rid="ref8">Barman et al., 2020</xref>). They trained and tested their system using PlantVillage. Using a self-built CNN (SBCNN) model, <xref ref-type="bibr" rid="ref45">Tiwari et al. (2020)</xref> classified potato leaves as early, late, or healthy. The regional PlantVillage dataset improved their model&#x2019;s accuracy. They did not utilize experimental data to validate their model. <xref ref-type="bibr" rid="ref15">Gupta et al. (2019)</xref> extracted and classified features using KNN, SVM, a neural network, and a pre-trained VGG19 model using KNN, SVM, and a neural network. PlantVillage has trained the computer to identify early and late blight symptoms on potato foliage. Research demonstrates that CNNs and other forms of deep learning effectively identify and categorize diseases in potato leaves. To further improve the performance of deep learning models, even with minimal training data, practitioners have turned to methods including data augmentation, transfer learning, and fine-tuning pre-trained models. These findings show that deep learning may improve potato disease detection and classification, which is crucial for the crop&#x2019;s long-term health.</p>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>Literature on potato leaf disease detection and classification using augmentation and deep learning models</title>
<p>The study, <xref ref-type="bibr" rid="ref7">Bappi et al. (2025)</xref> provided a novel deep-learning algorithm for potato leaf tissue disease detection using augmentation approaches. Scaling, flipping, and rotating the training dataset enhanced the model&#x2019;s accuracy. The research (<xref ref-type="bibr" rid="ref33">Rahman et al., 2021</xref>) examined how different kinds of enhancement may affect deep Convolutional Neural Networks (CNNs) ability to spot illnesses in potato leaves. In this research (<xref ref-type="bibr" rid="ref28">Plant Village Dataset, 2024</xref>), we applied deep learning models and data augmentation to improve our ability to identify diseases in potato leaves. The authors in the research work developed a deep learning-based method that uses data augmentation techniques to detect potato diseases. They used augmentation methods, including scaling, flipping, and rotating, to upsurge the size of the training dataset. In potato disease identification, training a CNN model on the expanded dataset resulted in high accuracy. <xref ref-type="table" rid="tab1">Table 1</xref> summarizes augmentation and deep learning studies on potato leaf disease detection and classification. These studies demonstrate the scope of current potato leaf disease identification and categorization efforts. While typical machine learning methods have shown promise, recent research has demonstrated that deep learning, particularly CNNs, may boost accuracy and automation. The proposed study on two-stage PotatoLeafNet CNN architectures will examine their ability to accurately identify and classify potato leaf diseases.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Summary of the literature on potato leaf disease detection and classification using augmentation and deep learning models.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Reference no</th>
<th align="left" valign="top">Approach</th>
<th align="left" valign="top">Data augmentation methods</th>
<th align="left" valign="top">Deep learning model</th>
<th align="left" valign="top">Key findings</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">
<xref ref-type="bibr" rid="ref21">Khalifa et al. (2021)</xref>
</td>
<td align="left" valign="middle">Deep learning-based approach</td>
<td align="left" valign="middle">Random cropping, flipping, and rotation</td>
<td align="left" valign="middle">CNN</td>
<td align="left" valign="middle">Achieved 94% accuracy, demonstrating robustness in varied conditions</td>
</tr>
<tr>
<td align="left" valign="middle">
<xref ref-type="bibr" rid="ref14">Ghosal et al. (2019)</xref>
</td>
<td align="left" valign="middle">Deep learning with augmentation techniques</td>
<td align="left" valign="middle">Scaling, rotation, and noise addition</td>
<td align="left" valign="middle">CNN</td>
<td align="left" valign="middle">Improved accuracy by 2% over non-augmented models</td>
</tr>
<tr>
<td align="left" valign="middle">
<xref ref-type="bibr" rid="ref34">Rathod et al. (2020)</xref>
</td>
<td align="left" valign="middle">Deep learning with data augmentation</td>
<td align="left" valign="middle">Color adjustment, zooming, and shifting</td>
<td align="left" valign="middle">CNN</td>
<td align="left" valign="middle">Enhanced model stability and a 5% increase in detection rate</td>
</tr>
<tr>
<td align="left" valign="middle">
<xref ref-type="bibr" rid="ref30">Potato Leaf Disease Dataset (2025)</xref>
</td>
<td align="left" valign="middle">Deep learning-based approach</td>
<td align="left" valign="middle">Extensive geometric and photometric transformations</td>
<td align="left" valign="middle">CNN</td>
<td align="left" valign="middle">Matched state-of-the-art accuracy, highlighting efficiency in processing large datasets</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec6">
<label>3</label>
<title>PotatoLeafNet: two-stage deep learning approach for accurate potato leaf disease detection and classification</title>
<p>Challenges in deep learning approaches for potato leaf disease identification include inaccurate disease recognition, disease variations, high false rates, inadequate training samples, imbalanced classes, slow convergence, and improved accuracy. Deep learning methods have been extensively researched to identify and categorize potato leaf diseases. Early identification and treatment of potato leaves are crucial, but the lack of agricultural expertise in rural areas can be time-consuming and hindered. Acquiring such datasets remains a difficult task. <xref ref-type="fig" rid="fig2">Figure 2</xref> represents the Flowchart for the proposed PotatoLeafNet two-stage CNN models for Potato Leaf Disease Detection and Classification.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Flowchart for the proposed PotatoLeafNet two-stage CNN models for potato leaf disease detection and classification.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">A flowchart showing the two-stage PotatoLeafNet pipeline, including image preprocessing, sequential image augmentation, dataset splitting into training, validation, and testing sets, CNN model training, prediction on test data, and performance evaluation.</alt-text>
</graphic>
</fig>
<sec id="sec7">
<label>3.1</label>
<title>Potato leaf disease dataset description</title>
<p>Plant Village Dataset provides high-quality photos of different potato leaves (<xref ref-type="bibr" rid="ref26">Mishra and Srivastava, 2019</xref>). Healthy, Early and Late Blight were photographed. Because of its availability, researchers have used the Plant Village dataset to simulate potato leaf diseases in the literature. This region-specific dataset includes few training and validation pictures and uneven class distribution. We need a fresh and comprehensive potato leaf dataset to address these research gaps. We curate the new dataset as the Potato Leaf Disease Dataset. Early Blight, containing 1,628 potato images, is the most critical disease affecting potatoes. The subsequent severe risk Late Blight contains 1,424 leaf images. We will examine 1,020 leaf images from the Healthy Next class for model training and testing. The dataset contains a complete 4,072 potato leaf images with three classes. The ratio between training, validation, and testing is 80:10:10. <xref ref-type="fig" rid="fig3">Figure 3</xref> displays the potato leaf images from each of the three categories. <xref ref-type="fig" rid="fig3">Figure 3</xref> presents the distribution of images across three classes of potato leaves: Early Blight, Late Blight, and Healthy. The Early Blight class has the largest number of images, just under 1,800, indicating a higher prevalence or focus on this category within the dataset. Late Blight follows closely, with a count near 1,600 images. The Healthy class has the fewest images, slightly above 1,400, suggesting a lesser representation in the dataset. This visual distribution highlights an imbalanced dataset which may be used for training a machine learning model to classify the health status of potato leaves.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>The potato leaf images from each of the 3 categories (original distribution).</p>
</caption>
<graphic xlink:href="frai-08-1668839-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Bar chart titled "Potato Leaf image classes" showing three orange bars representing different classes: Early Blight (around 1600), Late Blight (around 1400), and Healthy (around 800). The y-axis measures counts from 0 to 1800.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec8">
<label>3.2</label>
<title>Image processing and sequential image augmentation</title>
<p>Pre-processing was applied to all images to enhance lesion visibility, suppress background clutter, and standardize inputs prior to learning. Specifically, we performed contrast normalization to mitigate illumination variability, foreground&#x2013;background separation to isolate the lamina, and spatial normalization to a common resolution. This stage improves the signal-to-noise ratio presented to the network and, in turn, the reliability of feature extraction for downstream classification (<xref ref-type="bibr" rid="ref17">Hernandez-Valencia et al., 2020</xref>). To reduce storage and I/O overhead without compromising diagnostically salient content, we employed lossless and hybrid compression. Lossless codecs Huffman coding and run-length encoding (RLE) preserve the exact pixel values while exploiting redundancy to shrink file size (<xref ref-type="bibr" rid="ref48">Yao et al., 2020</xref>). In the hybrid scheme, regions containing disease cues (lesion edges, texture) are preserved losslessly, whereas visually noncritical background is compressed lossily, striking a balance between fidelity and efficiency for large-scale training and deployment (<xref ref-type="bibr" rid="ref44">TensorFlow Sequential Data Augmentation, 2025</xref>). (Compression is decoupled from resizing, it reduces bytes on disk/transfer, not spatial resolution.) leaf images captured in RGB are converted to grayscale (<xref ref-type="bibr" rid="ref16">Gurucharan, 2020</xref>). Edge of Caution to recognize the edges in a leaf image and alleviate the irritation, unambiguous evidence is utilized (<xref ref-type="bibr" rid="ref31">Powers, 2020</xref>). The external designs in leaf images are equal in how they are perceived from the edge. When the upper shape is taken as (p, q), the breadth and the level are (r, s), and these four centers do not settle the bobbing (<xref ref-type="bibr" rid="ref23">Li et al., 2022</xref>). Each member of the upright hopping square is still a work in progress. The return on investment region is removed using the primary RGB leaf image&#x2019;s coordinates (p&#x202F;+&#x202F;r, q&#x202F;+&#x202F;s). Finally, the dreaded leaf symbol may be put to rest.</p>
<sec id="sec9">
<label>3.2.1</label>
<title>Sequential image augmentation</title>
<p>Sequential image augmentation can be incorporated into a sequential model in TensorFlow-Keras by using the &#x201C;tf.keras.layers.experimental.pre-processing&#x201D; module. This module provides various layers that can be added to the sequential model to apply different image augmentation techniques. These layers can be sequentially added to the model to apply random horizontal flipping, random rotation within a specific range, random zooming, and rescaling of pixel values. By adding the image augmentation pipeline as the first layer of the CNN sequential model, the subsequent layers can learn from augmented data, enabling the model to generalize better and handle variations in the input images.</p>
<p>The &#x2018;image_augmentation&#x2019; sequential model used in this research represents an augmentation pipeline using various pre-processing layers.</p><list list-type="bullet">
<list-item>
<p><bold>Random Rotation</bold>: This layer randomly rotates the input image by a maximum of 0.2 radians (approximately 11.5 degrees) in a counterclockwise direction. It introduces variability by simulating different object orientations in the image.</p>
</list-item>
<list-item>
<p><bold>Random Zoom</bold>: This layer applies random zooming to the input image, ranging from 0 to 20% of the original size. It helps capture different scales or perspectives of the object.</p>
</list-item>
<list-item>
<p><bold>Random Height</bold>: This layer randomly changes the height of the input image by scaling it between 80 and 120% of the original height. It adds variability by modifying the image&#x2019;s aspect ratio.</p>
</list-item>
<list-item>
<p><bold>Random Width</bold>: This layer randomly changes the width of the input image by scaling it between 80 and 120% of the original width. Like &#x2018;Random Height&#x2019; it introduces variability by modifying the aspect ratio.</p>
</list-item>
<list-item>
<p><bold>Rescaling:</bold> This layer rescales the pixel values of the input image to a range of [0, 1]. In this case, it divides each pixel value by 255, assuming the input image has an 8-bit color depth. Rescaling is a typical pre-processing step to ensure numerical stability and convergence during model training.</p>
</list-item>
</list>
<p>The &#x2018;image_augmentation&#x2019; model can be used as a pre-processing step in your overall CNN model pipeline. It applies random transformations to the input images during training, enhancing the diversity and robustness of the data.</p>
</sec>
</sec>
<sec id="sec10">
<label>3.3</label>
<title>Augmented data split into training, validation, and testing</title>
<p>The potato leaf disease dataset was divided into training, validation, and testing sets using 80, 10, and 10% split ratios. Sequential image augmentation procedures on the training set reduced overfitting and increased dataset variation. Rescaling, rotating, modifying shear and zoom ranges, flipping horizontally, adjusting brightness, and moving channels were these tactics. CNN model predictions were improved using Adam optimization with forward and backpropagation. Thus, CNN model output accuracy was ensured. The validation and testing sets contained 20% of the training set, which included images of early, robust, and late blight. The PotatoLeafNet model categorized practice pictures and predicted class labels on the training dataset.</p>
</sec>
<sec id="sec11">
<label>3.4</label>
<title>PotatoLeafNet&#x2014;potato leaf-based CNN for potato leaf disease detection and classification</title>
<p>The existing literature on deep learning approaches reveals several challenges, including misdiagnosis of potato leaf identification, variations in potato leaves due to different varieties, and environmental factors. Early detection and management of potato diseases are crucial, but the process is time-consuming, and access to agricultural expertise is limited in rural areas (<xref ref-type="bibr" rid="ref5">Alzakari et al., 2025</xref>). CNNs have shown remarkable progress in image-based recognition, eliminating the need for extensive image pre-processing and enabling automatic feature selection (<xref ref-type="bibr" rid="ref47">Weng et al., 2024</xref>). However, the availability of large datasets specifically for potato leaf challenges remains a significant obstacle.</p>
<sec id="sec12">
<label>3.4.1</label>
<title>Convolutional neural network (CNN) model</title>
<p>CNNs were developed to process the data represented in grid-like structures like images. The pixels in an image are arranged in a grid, and the value of each pixel determines its hue and luminance. Likewise, each neuron in a CNN processes information within its receptive field. Like how the human brain processes visual information, CNN layers detect simpler patterns first, then more complex ones as the layer progresses.</p>
<p>Convolutional neural networks have input, hidden, and output layers. Convolution, normalization, pooling, and fully-connected layers lie between the output and input layers. The convolutional layer&#x2019;s filters create classification feature maps. Image processing uses ReLU. This paper proposes an improved fine-grained robust PotatoLeafNet model for classifying potato leaf diseases. To minimize the size of the leaf picture and create several images, image pre-processing and sequential image augmentation methods are utilized at the first level. A CNN learning model using a CNN has been established at the next level to identify sick leaves in the images. The PotatoLeafNet model for potato leaf disease prediction is shown in <xref ref-type="fig" rid="fig4">Figure 4</xref>.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Architecture of the proposed PotatoLeafNet model for potato leaf detection and classification.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">A schematic representation of the PotatoLeafNet CNN architecture, consisting of sequential image augmentation followed by multiple convolutional and max-pooling layers, global average pooling, fully connected layers, and a softmax output for potato leaf disease classification.</alt-text>
</graphic>
</fig>
<p>CNN models have numerous convolutional, pooling, and fully linked layers. Because of its high complexity, a neural network can develop hierarchical representations of the input data, which are crucial for precise categorization. Section 3.6 presents a detailed pseudo code for the proposed CNN model for Potato Leaf detection and classification. The Convolutional Neural Network model consists of multiple layers, including Conv2D, Batch Normalization, Max Pooling, and Activation functions. Here is a description of the model architecture:<list list-type="order">
<list-item>
<p><bold>Conv2D layers</bold>: The convolutional operation distinguishes a CNN from other neural networks. The basic form of convolution consists of two functions that take real numbers as arguments. To explain convolution, we can pretend that it is possible to track where a car is using a laser that gives an output: <italic>x(t)</italic>, where x is the car&#x2019;s position in time step <italic>t</italic>. Several measurements can be taken to reduce possible noise during the measurements, and an average value of them is used as the measurement value. Later measurements have greater value than the older ones. Therefore a weight function, w(a), is used, where a represents how old a measurement is. The weight function w must be a valid density function. If these weighted average measurements are performed every time step, it can be described with a function, s known as the Convolution function.<disp-formula id="E1">
<mml:math id="M1">
<mml:mi>s</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo>&#x222B;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mi>w</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mi mathvariant="italic">da</mml:mi>
</mml:math>
<label>(1)</label>
</disp-formula></p>
</list-item>
</list></p>
<p>In CNN terminology, the first argument in the convolution function is called the input, and the second is called the kernel; what is returned is called the <italic>feature map</italic>.</p><disp-formula id="E2">
<mml:math id="M2">
<mml:mi>s</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>w</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(2)</label>
</disp-formula>
<p>For the example with the car above to be realistic, the data cannot be collected in each time step when the amount had become too large, but in regular intervals, for example, every second or minute. In such a case, the time variable t would only be of integer type; likewise, the variables x and w, then the mathematical discrete convolution, can be defined as.</p><disp-formula id="E3">
<mml:math id="M3">
<mml:mi>s</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>&#x2217;</mml:mo>
<mml:mi>w</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo>&#x221E;</mml:mo>
</mml:mrow>
<mml:mo>&#x221E;</mml:mo>
</mml:msubsup>
<mml:mi>x</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mi>w</mml:mi>
<mml:mspace width="0.25em"/>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>a</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(3)</label>
</disp-formula>
<p>The model includes 5 Conv2D layers. Conv2D performs convolution operations on the input image to extract features. Each Conv2D layer consists of a set of learnable filters that scan the input image and produce feature maps. These filters capture different patterns and features at different scales.<list list-type="order">
<list-item>
<p><bold>Batch Normalization layers</bold>: The batch normalized activation is<disp-formula id="E4">
<mml:math id="M4">
<mml:msub>
<mml:munder accentunder="true">
<mml:mi>x</mml:mi>
<mml:mo stretchy="true">_</mml:mo>
</mml:munder>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>&#x03BC;</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
</mml:mrow>
<mml:msqrt>
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mi>B</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo>+</mml:mo>
<mml:mi>e</mml:mi>
</mml:mrow>
</mml:msqrt>
</mml:mfrac>
</mml:math>
<label>(4)</label>
</disp-formula></p>
</list-item>
</list>Where <inline-formula>
<mml:math id="M5">
<mml:msub>
<mml:mi>&#x03BC;</mml:mi>
<mml:mi>B</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>m</mml:mi>
</mml:mfrac>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>m</mml:mi>
</mml:msubsup>
<mml:mi mathvariant="italic">xi</mml:mi>
</mml:math>
</inline-formula> is the batch mean, and <inline-formula>
<mml:math id="M6">
<mml:msup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mi>B</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>m</mml:mi>
</mml:mfrac>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>m</mml:mi>
</mml:msubsup>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi mathvariant="italic">xi</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi mathvariant="italic">&#x03BC;B</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:math>
</inline-formula> is the batch variance. Batch Normalization is applied after each Conv2D layer. Adjusting the mean and variance helps normalize the previous layer&#x2019;s output. It helps stabilize the training process and improve the model&#x2019;s overall performance.<list list-type="order">
<list-item>
<p><bold>Max Pooling layers</bold>: Max Pooling precedes each Conv2D layer. The maximum value inside a pool size lower feature map spatial dimension. Max Pooling helps down-sampling the feature maps and extracting the most essential features while reducing computational complexity.<disp-formula id="E5">
<mml:math id="M7">
<mml:msubsup>
<mml:mi>h</mml:mi>
<mml:mi mathvariant="italic">xy</mml:mi>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mo>max</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>.</mml:mo>
<mml:mi>s</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>&#x2026;</mml:mo>
<mml:mn>..</mml:mn>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mspace width="0.25em"/>
<mml:msup>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msup>
<mml:mspace width="0.25em"/>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(5)</label>
</disp-formula></p>
</list-item>
<list-item>
<p><bold>Activation functions</bold>: 7 activation functions are used in the model&#x2019;s layers. Activation functions allow the model to learn complicated patterns and make nonlinear judgments. CNNs often use ReLU, sigmoid, and tanh activation functions to feed one layer&#x2019;s output into the next.<disp-formula id="E6">
<mml:math id="M8">
<mml:mtext mathvariant="italic">ReLU</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
<mml:mspace width="0.25em"/>
</mml:math>
<label>(6)</label>
</disp-formula></p>
</list-item>
</list></p>
<p>Combining Conv2D layers, Batch Normalization, Max Pooling, and Activation functions helps the CNN model extract and learn intricate features from the input data effectively. It allows the model to capture the information for accurate classification or detection tasks. The model uses convolutional operations, sequential image augmentation, and global average pooling to accurately and efficiently classify potato leaf diseases. <xref ref-type="table" rid="tab2">Table 2</xref> shows the proposed PotatoLeafNet model architecture summary, and <xref ref-type="fig" rid="fig4">Figure 4</xref> represents the proposed PotatoLeafNet model for Potato Leaf detection and classification.</p><list list-type="bullet">
<list-item>
<p><bold>Sequential Image Augmentation</bold>: The input images are enhanced by this layer using various image augmentation methods, including random flipping, rotation, zooming, and rescaling. It transforms the pictures to prepare them for better generalization.</p>
</list-item>
<list-item>
<p><bold>Conv2D (60 filters, 3&#x202F;&#x00D7;&#x202F;3, ReLU):</bold> This convolutional layer performs convolutions on the input images using 60 filters of size 3&#x202F;&#x00D7;&#x202F;3 and applies the ReLU activation function. It extracts 60 different features from the input images, resulting in an output shape of (None, 222, 222, 60).</p>
</list-item>
<list-item>
<p><bold>MaxPooling2D (pool size: 2&#x202F;&#x00D7;&#x202F;2):</bold> By taking the highest value inside each 2&#x202F;&#x00D7;&#x202F;2 region, this max pooling layer decreases the spatial dimensions of the feature maps by a factor of 2 (<xref ref-type="bibr" rid="ref45">Tiwari et al., 2020</xref>). It helps in reducing the spatial dimensions and capturing the most salient features, resulting in an output shape of (None, 111, 111, 60).</p>
</list-item>
<list-item>
<p><bold>GlobalAveragePooling2D</bold>: This layer performs global average pooling, reducing the spatial dimensions to a single value per channel. It summarizes spatial information and retains essential features. Resulting in an output shape of (None, 60).</p>
</list-item>
<list-item>
<p><bold>Dense (128&#x202F;units, ReLU):</bold> This fully connected dense layer with 128&#x202F;units applies the ReLU activation function. It introduces non-linearity and learns high-level representations based on the extracted features from previous layers. Resulting in an output shape of (None, 128).</p>
</list-item>
<list-item>
<p><bold>Dense (total_classes units, Softmax):</bold> The softmax activation function is used in the last dense layer, which has units equal to the entire number of classes in the classification job. It generates class probabilities, which show the chance that each input picture belongs to a specific class of illness.</p>
</list-item>
</list>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Summary of the proposed PotatoLeafNet model architecture.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Layer</th>
<th align="left" valign="top">Output shape</th>
<th align="center" valign="top">Number of parameters</th>
<th align="left" valign="top">Unique Configuration</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">Sequential Image Augmentation</td>
<td align="left" valign="middle">(None, 224, 224, 3)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Custom augmentation settings for potato leaf images</td>
</tr>
<tr>
<td align="left" valign="middle">Conv2D (60 filters, 3&#x202F;&#x00D7;&#x202F;3, ReLU)</td>
<td align="left" valign="middle">(None, 222, 222, 60)</td>
<td align="center" valign="middle">1,740</td>
<td align="left" valign="middle">Optimized for initial feature extraction</td>
</tr>
<tr>
<td align="left" valign="middle">MaxPooling2D (pool size: 2&#x00D7;2)</td>
<td align="left" valign="middle">(None, 111, 111, 60)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Reduces dimensionality, retains critical spatial features</td>
</tr>
<tr>
<td align="left" valign="middle">Conv2D (60 filters, 3&#x202F;&#x00D7;&#x202F;3, ReLU)</td>
<td align="left" valign="middle">(None, 109, 109, 60)</td>
<td align="center" valign="middle">32,460</td>
<td align="left" valign="middle">Additional depth to capture complex features</td>
</tr>
<tr>
<td align="left" valign="middle">MaxPooling2D (pool size: 2&#x00D7;2)</td>
<td align="left" valign="middle">(None, 54, 54, 60)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Further reduces spatial dimensions, focuses on feature pooling</td>
</tr>
<tr>
<td align="left" valign="middle">Conv2D (60 filters, 3&#x202F;&#x00D7;&#x202F;3, ReLU)</td>
<td align="left" valign="middle">(None, 52, 52, 60)</td>
<td align="center" valign="middle">32,460</td>
<td align="left" valign="middle">Increases model&#x2019;s capacity to learn detailed features</td>
</tr>
<tr>
<td align="left" valign="middle">MaxPooling2D (pool size: 2&#x00D7;2)</td>
<td align="left" valign="middle">(None, 26, 26, 60)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Enhances the abstraction level of the features</td>
</tr>
<tr>
<td align="left" valign="middle">Conv2D (60 filters, 3&#x202F;&#x00D7;&#x202F;3, ReLU)</td>
<td align="left" valign="middle">(None, 24, 24, 60)</td>
<td align="center" valign="middle">32,460</td>
<td align="left" valign="middle">Prepares for high-level reasoning by the network</td>
</tr>
<tr>
<td align="left" valign="middle">MaxPooling2D (pool size: 2&#x00D7;2)</td>
<td align="left" valign="middle">(None, 12, 12, 60)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Last pooling step to compact features before classification</td>
</tr>
<tr>
<td align="left" valign="middle">GlobalAveragePooling2D</td>
<td align="left" valign="middle">(None, 60)</td>
<td align="center" valign="middle">0</td>
<td align="left" valign="middle">Reduces each feature map to a single number to minimize overfitting</td>
</tr>
<tr>
<td align="left" valign="middle">Dense (128&#x202F;units, ReLU)</td>
<td align="left" valign="middle">(None, 128)</td>
<td align="center" valign="middle">7,808</td>
<td align="left" valign="middle">Dense layer to combine features into higher-level attributes</td>
</tr>
<tr>
<td align="left" valign="middle">Dense (total_classes units, Softmax)</td>
<td align="left" valign="middle">(None, total_classes)</td>
<td align="center" valign="middle">total_classes</td>
<td align="left" valign="middle">Tailored for the specific number of disease classes</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>These layers form the PotatoLeafNet model, which combines sequential image augmentation, convolutional layers for feature extraction, pooling layers for spatial dimension reduction, global average pooling for summarization, and fully connected layers for classification. The model is trained to classify potato leaf disease images into their respective classes.</p>
</sec>
</sec>
<sec id="sec13">
<label>3.5</label>
<title>Performance measure</title>
<p>Multiple metrics are used to evaluate the success of a network. Using different task metrics helps represent the network&#x2019;s ability to solve a given problem. The evaluation metrics can use true positive (TP), false positive (FP), true negative (TN), and false negative (FN).</p>
<p><bold>Classification Accuracy:</bold> is determined by the ratio of correct prediction to total predictions.</p><disp-formula id="E7">
<mml:math id="M9">
<mml:mtext mathvariant="italic">Accu</mml:mtext>
<mml:mi mathvariant="normal">r</mml:mi>
<mml:mi mathvariant="italic">acy</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mtext mathvariant="italic">Number of Correct Predictions</mml:mtext>
<mml:mtext mathvariant="italic">Total number of Predictions</mml:mtext>
</mml:mfrac>
</mml:math>
<label>(7)</label>
</disp-formula>
<p><bold>Precision:</bold> Precision determines with what precision the network places images in the positive category. Precision is calculated as follows:</p><disp-formula id="E8">
<mml:math id="M10">
<mml:mtext mathvariant="italic">Precision</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(8)</label>
</disp-formula>
<p><bold>Recall:</bold> Recall indicates how many positive images the network recorded. The recall is calculated as follows:</p><disp-formula id="E9">
<mml:math id="M11">
<mml:mtext mathvariant="italic">Recall</mml:mtext>
<mml:mspace width="0.25em"/>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(9)</label>
</disp-formula>
<p><bold>F1-Score</bold>: F1-Score is a combination of Precision and Recall. The calculation is as follows:</p><disp-formula id="E10">
<mml:math id="M12">
<mml:mi>F</mml:mi>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext mathvariant="italic">Score</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mn>2</mml:mn>
<mml:mo>&#x2217;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="italic">Precision</mml:mtext>
<mml:mo>&#x2217;</mml:mo>
<mml:mtext mathvariant="italic">Recall</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="italic">Precision</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="italic">Recall</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(10)</label>
</disp-formula>
</sec>
<sec id="sec14">
<label>3.6</label>
<title>Algorithm of the proposed PotatoLeafNet model for potato leaf detection and classification</title>
<p>PotatoLeafNet model for potato leaf detection and classification shown in <xref ref-type="sec" rid="sec14">Algorithm 1</xref>.<statement id="algo1" content-type="algorithm">
<label>ALGORITHM 1</label>
<p><preformat><bold>Input: Potato Leaf Disease Dataset</bold>
<bold>Output: Disease Detection and Classification of Potato Leaves</bold><bold>Step1:</bold> Acquire the Potato images with Late Blight, Early Blight, and Healthy
<bold>Step2:</bold> Loading the data (X_train,y_train), (X_test,y_test)=image.load_data()
<bold>Step 3:</bold> first stage of PotatoLeafNet for sequential image augmentation model for image augmentation with 6 layers. Each layer is performing RandomFlip, RandomRotation(0.2), RandomZoom(0.2), Rescaling(1./255)
<bold>Step 4</bold>: Put the correct labels on the pictures of the potato leaf images.
<bold>Step 5</bold>: Sort photos into categories using the available class labels from the training and testing datasets.
<bold>Step 6:</bold> Initialize the parameters image size, epochs, batch size, and train and test image labels.
<bold>Step 7:</bold> The Second stage of the PotatoLeafNet Model uses 4 blocks containing Conv2D, Max Pool2D, and GloalAveragePooling2D, followed by Desne layers. The total number of layers is 11.
<bold>Step 8:</bold> Evaluate the trained model using a separate testing dataset.
&#x2003;&#x2003;&#x2003;Calculate the test loss and accuracy of the model.
<bold>Step 9:</bold> Check the accuracy of the proposed models, and see how they stack up against the rest of the CNN models out there. Make predictions on new data
predictions = model.predict(new_images).</preformat>
</p>
</statement></p>
</sec>
</sec>
<sec sec-type="results|discussion" id="sec15">
<label>4</label>
<title>Results and discussion</title>
<p>All experiments were implemented in Python using TensorFlow and Keras, optimizing a categorical cross-entropy objective with Adam and a learning-rate schedule; runs were executed on a server equipped with an NVIDIA P100 GPU, an Intel i5 CPU, and 8&#x202F;GB RAM. The evaluation centered on four aims: reliably tri-classifying potato leaf images into Early Blight, Late Blight, and Healthy categories, quantifying the effect of a fixed sequential image-augmentation pipeline during training on the PotatoLeafNet&#x2019;s performance; benchmarking PotatoLeafNet against contemporary convolutional baselines; and situating the empirical findings within prior deep-learning studies on potato leaf disease identification.</p>
<sec id="sec16">
<label>4.1</label>
<title>Datasets description</title>
<p>We curated a diverse, high-quality corpus of potato leaf images spanning Healthy, Early Blight, and Late Blight classes. Training uses the PlantVillage Potato subset, a widely used, fully open benchmark for potato leaf disease recognition, to mitigate its limitations and class imbalance, we additionally compiled a complementary Potato Leaf Disease Dataset with 4,072 images 1,628 Early Blight, 1,424 Late Blight, and 1,020 Healthy. To assess real-world generalization beyond PlantVillage, we conduct cross-dataset validation: models trained on PlantVillage are evaluated, without further tuning, on PlantDoc (<xref ref-type="bibr" rid="ref42">Singh et al., 2020</xref>) (in-situ scenes with variable lighting, occlusion, and background clutter) and on <xref ref-type="bibr" rid="ref39">Shabrina et al. (2023)</xref> field collection (uncontrolled conditions; seven potato classes remapped to {Healthy, Early Blight, Late Blight} for comparability). We report Accuracy, Macro-F1, per-class Precision/Recall, Matthews Correlation Coefficient (MCC), and Expected Calibration Error (ECE), and provide confusion matrices and Grad-CAM overlays. Finally, a few-shot field-adaptation ablation (10% labeled field images) quantifies domain shift and the benefit of lightweight adaptation.</p>
</sec>
<sec id="sec17">
<label>4.2</label>
<title>Data pre-processing and sequential image augmentation</title>
<p>All images were prepared for CNN training by converting them to RGB float tensors and resizing uniformly to 224&#x202F;&#x00D7;&#x202F;224 pixels. Pixel intensities were normalized to [0,1] to stabilize optimization. The dataset comprised 4,072 potato-leaf images across three classes Healthy, Early Blight, and Late Blight. To enhance the robustness and generalizability of PotatoLeafNet, we applied a fixed-order (sequential) image-augmentation pipeline in Keras on the training split only, thereby increasing appearance diversity while preserving label integrity and class balance. The augmentation sequence consisted of rotation (&#x00B1;25&#x00B0;), width shift (&#x00B1;0.10), height shift (&#x00B1;0.10), shear (0.20), random zoom (up to 0.20), horizontal flip, brightness jitter (0.5&#x2013;1.0), and channel shift (0.05). Applying this policy expanded the training corpus to 6,000 images, balanced as 2,000 per class, which mitigated class imbalance and improved generalization across diverse disease manifestations. <xref ref-type="fig" rid="fig5">Figure 5</xref> illustrates representative pre-processed images at the target 224&#x202F;&#x00D7;&#x202F;224 resolution.</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>The pre-processed images with a resolution of 224&#x00D7;224.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Six close-up images of leaves showing various stages of damage. The leaves display different degrees of browning and yellowing. Some leaves have visible spots and decay, while others appear healthier with minimal discoloration. The background is a neutral gray.</alt-text>
</graphic>
</fig>
<sec id="sec18">
<label>4.2.1</label>
<title>Evaluation protocol</title>
<p>We access robustness with Repeated K-Stratified K-Fold (<italic>k</italic>&#x202F;=&#x202F;5, <italic>r</italic>&#x202F;=&#x202F;3; total <italic>N</italic>&#x202F;=&#x202F;15 fits). For fold f, models are trained on <inline-formula>
<mml:math id="M13">
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>f</mml:mi>
<mml:mtext mathvariant="italic">train</mml:mtext>
</mml:msubsup>
</mml:math>
</inline-formula>, validated on <inline-formula>
<mml:math id="M14">
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>f</mml:mi>
<mml:mi mathvariant="italic">val</mml:mi>
</mml:msubsup>
</mml:math>
</inline-formula> (early stopping, best checkpoint), and evaluated on <inline-formula>
<mml:math id="M15">
<mml:msubsup>
<mml:mi>D</mml:mi>
<mml:mi>f</mml:mi>
<mml:mtext mathvariant="italic">test</mml:mtext>
</mml:msubsup>
</mml:math>
</inline-formula>. We report Accuracy, Macro-precision, Macro- Recall, Macro-F1 as <inline-formula>
<mml:math id="M16">
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>&#x00B1;</mml:mo>
<mml:mi>&#x03C3;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> over all N runs, with 95% cls via the t-distribution:</p><disp-formula id="E11">
<mml:math id="M17">
<mml:mtable columnalign="left" displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mspace width="0.33em"/>
</mml:mrow>
</mml:mfrac>
<mml:mspace width="0.33em"/>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:msubsup>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>&#x03C3;</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfrac>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:munderover>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mi>C</mml:mi>
<mml:msub>
<mml:mi>I</mml:mi>
<mml:mrow>
<mml:mn>95</mml:mn>
<mml:mo>%</mml:mo>
</mml:mrow>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>&#x00B1;</mml:mo>
<mml:msub>
<mml:mi>t</mml:mi>
<mml:mrow>
<mml:mn>0.975</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>N</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfrac>
<mml:mi>&#x03C3;</mml:mi>
<mml:mrow>
<mml:mo>&#x221A;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:mtd>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(11)</label>
</disp-formula>
<p>For pairwise model comparisons on identical folds we apply paired t-tests and Wilcoxon signed-rank tests, prediction-level differences are examined with the McNemar test. We report Hedges&#x2019; g and cliff&#x2019;s <inline-formula>
<mml:math id="M18">
<mml:mi>&#x03B4;</mml:mi>
</mml:math>
</inline-formula> as effect sizes and apply Holm-Bonferroni to control family-wise error. To quantify optimization stochasticity, we additionally train each model on the canonical split with 5 distinct random seeds and report mean &#x00B1; SD Protocol details Input Size 224&#x002A;2,224, identical preprocessing/normalization across models, no augmentation on validation/test, Stratification by class (Healthy, Early Blight, Late Blight), fixed fold indices shared by all models, deterministic settings (global seed, seeded data loaders, CuDNN deterministic).</p>
</sec>
</sec>
<sec id="sec19">
<label>4.3</label>
<title>PotatoLeafNet performance on potato leaf disease dataset</title>
<p><xref ref-type="table" rid="tab3">Table 3</xref> illustrates PotatoLeafNet configurations. The dataset shows the PotatoLeafNet model, which changes internal parameters to improve performance during training. The model learns to extract important traits and properly characterize Late, early blight, and Healthy over several epochs. Each epoch&#x2019;s accuracy and loss statistics show model performance. Accuracy is the percentage of correctly predicted instances concerning actual. <xref ref-type="fig" rid="fig6">Figures 6</xref>&#x2013;<xref ref-type="fig" rid="fig9">9</xref> demonstrate the accuracy of PotatoLeafNet architectures&#x2019; potato leaf disease detection and classification.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Parameters used in the PotatoLeafNet model.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">S. No</th>
<th align="left" valign="top">Parameter used</th>
<th align="center" valign="top">Value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Training epochs</td>
<td align="center" valign="top">10 and 100</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Optimizer</td>
<td align="center" valign="top">Adam</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">Batch size</td>
<td align="center" valign="top">32</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="left" valign="top">Drop out</td>
<td align="center" valign="top">0.25</td>
</tr>
<tr>
<td align="left" valign="top">5</td>
<td align="left" valign="top">Image size</td>
<td align="center" valign="top">224&#x202F;&#x00D7;&#x202F;224</td>
</tr>
<tr>
<td align="left" valign="top">6</td>
<td align="left" valign="top">Kernel size</td>
<td align="center" valign="top">3</td>
</tr>
<tr>
<td align="left" valign="top">7</td>
<td align="left" valign="top">Data shuffle</td>
<td align="center" valign="top">True for every 1,000 images</td>
</tr>
<tr>
<td align="left" valign="top">8</td>
<td align="left" valign="top">No of classes</td>
<td align="center" valign="top">3</td>
</tr>
<tr>
<td align="left" valign="top">9</td>
<td align="left" valign="top">Callback</td>
<td align="center" valign="top">True on model checkpoint</td>
</tr>
<tr>
<td align="left" valign="top">10</td>
<td align="left" valign="top">Loss function</td>
<td align="center" valign="top">Cross entropy</td>
</tr>
<tr>
<td align="left" valign="top">11</td>
<td align="left" valign="top">Learning rate</td>
<td align="center" valign="top">0.001</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>The performance of PotatoLeafNet on potato leaf disease with sequential image augmentation for 100 epochs.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g006.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two line graphs depicting model performance over 100 epochs. The left graph shows model accuracy, with both training and validation accuracy increasing rapidly initially and stabilizing around 0.97. The right graph shows model loss, with both training and validation loss decreasing to about 0.1, indicating effective model training.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Performance of PotatoLeafNet on potato leaf disease without sequential image augmentation for 100 epochs.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g007.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two line graphs show model accuracy and loss over 100 epochs. The left graph displays accuracy, with training and validation lines converging near 1.0. The right graph illustrates loss, with both lines declining and stabilizing around zero.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>Performance of PotatoLeafNet on potato leaf disease with sequential image augmentation for 10 epochs.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g008.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two line graphs display model performance over epochs. The left graph shows model accuracy, with the training line increasing from 0.4 to 0.9 and validation rising slightly higher. The right graph shows model loss decreasing from 1.1 to 0.3 for both training and validation, indicating improved performance.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig9">
<label>Figure 9</label>
<caption>
<p>Performance of PotatoLeafNet on potato leaf disease without sequential image augmentation for 10 epochs.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g009.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Two line graphs showing model accuracy over epochs. The left graph depicts training and validation accuracy lines, both increasing with some fluctuation, reaching around 0.9. The right graph similarly displays increasing accuracy over epochs, with the validation line spiking late. Both graphs have epochs on the x-axis and accuracy on the y-axis.</alt-text>
</graphic>
</fig>
<p>The two-stage CNN model evaluates PotatoLeafNet on potato leaf disease data. Potato leaf disease datasets employ six-layer sequential image enhancement. A freshly developed and fine-tuned CNN model analyzes the training dataset&#x2019;s accuracy. The PotatoLeafNet model was optimized with sequential picture augmentation. PotatoLeafNet model training requires a huge sample.</p>
<p><xref ref-type="fig" rid="fig6">Figure 6</xref> summarizes the learning dynamics of PotatoLeafNet trained for 100 epochs with sequential image augmentation on the potato leaf disease dataset. The training accuracy rapidly increases and saturates at 98.92% with a final training loss of 0.0356, while the validation accuracy stabilizes around 97.53% with a closely aligned validation loss curve. The small gap between training and validation accuracies, together with the monotonically decreasing and non-diverging loss trajectories, indicates that the augmented model generalizes well beyond the training set. On the held-out test set, PotatoLeafNet attains 98.52% accuracy, confirming that sequential image augmentation provides effective regularization and supports highly reliable classification of healthy, early blight, and late blight leaves.</p>
<p><xref ref-type="fig" rid="fig7">Figure 7</xref> shows the behavior of a single-stage PotatoLeafNet trained for 100 epochs without sequential image augmentation. In this setting, the model reaches 96.17% training accuracy with a training loss of 0.2250 and achieves 96.52% validation accuracy, but the validation accuracy and loss curves exhibit noticeably larger oscillations than in <xref ref-type="fig" rid="fig6">Figure 6</xref>. Test accuracy is reduced to 96.01%, i.e., 2.51 percentage points below the augmented two-stage PotatoLeafNet, and the substantially higher training loss further reflects less stable optimization. Comparing <xref ref-type="fig" rid="fig6">Figures 6</xref>, <xref ref-type="fig" rid="fig7">7</xref> demonstrates that sequential image augmentation not only increases training/validation/test accuracy by +2.75/+1.01/+2.51 points, respectively, but also yields smoother validation trajectories and lower loss, highlighting its role as an effective regularizer.</p>
<p><xref ref-type="fig" rid="fig8">Figure 8</xref> depicts PotatoLeafNet trained for only 10 epochs with sequential image augmentation. Even under this short training regime, the model already reaches 88.22% training accuracy with a training loss of 0.3535, while the validation accuracy rises to 86.91% and the validation loss decreases steadily. The corresponding test accuracy of 88.15% confirms that the augmented model generalizes well even before full convergence. These dynamics indicate that augmentation quickly exposes the network to diverse views of each class, enabling the model to acquire discriminative features early in training and to maintain a small and stable train&#x2013;validation gap.</p>
<p><xref ref-type="fig" rid="fig9">Figure 9</xref> presents the same 10-epoch training schedule without sequential image augmentation. In this baseline configuration, the model attains 87.82% training accuracy and 0.3410 training loss, with validation and test accuracies of 85.82 and 86.91%, respectively. Compared with <xref ref-type="fig" rid="fig8">Figure 8</xref>, both validation and test accuracies are consistently lower and the gap between training and validation curves is slightly larger, suggesting mild overfitting when the network is trained on a less diverse set of images. The corresponding loss curve also shows a less smooth descent, pointing to reduced robustness of the optimization process. Together, <xref ref-type="fig" rid="fig8">Figures 8</xref>, <xref ref-type="fig" rid="fig9">9</xref> illustrate that, even at an early training stage, sequential image augmentation improves generalization and stabilizes the learning dynamics of PotatoLeafNet.</p>
<sec id="sec20">
<label>4.3.1</label>
<title>Model efficiency and parameters</title>
<p>To contextualize deployment cost alongside accuracy (<xref ref-type="table" rid="tab4">Table 4</xref>), we benchmarked five models under a unified protocol and report tuning strategy, wall-clock training time, parameter count, and FP32 memory footprint (4 bytes per parameter). The proposed PotatoLeafNet used manual tuning with ReduceLROnPlateau and ModelCheckpoint, a fixed sequential augmentation policy, and Adam (lr&#x202F;=&#x202F;1e-3); it trained in 1.2&#x202F;h, contains 16.5 M parameters, and occupies 66&#x202F;MB of memory (moderate complexity). The ResNet-50&#x202F;+&#x202F;VGG-16 fusion, using transfer learning with fine-tuning, trained in 1.5&#x202F;h and comprises 164.00 M parameters (656&#x202F;MB) [or 38.30&#x202F;M, 153&#x202F;MB, if reported without the ImageNet classifier]; VGG-16&#x202F;+&#x202F;MobileNetV2 with grid-search tuning trained in 1.8&#x202F;h and totals 141.90 M parameters (567.6&#x202F;MB) [or 16.97&#x202F;M, 67.9&#x202F;MB, without top]. MobileNetV2, tuned via random search, trained in 0.8&#x202F;h, has 3.54 M parameters (14&#x202F;MB) and low complexity. Inception-V3, fine-tuned via standard transfer learning, trained in 1.2&#x202F;h and includes 23.85 M parameters (&#x2248;95&#x202F;MB) with moderate complexity. These results show that PotatoLeafNet is far smaller than fusion baselines and within an order of magnitude of MobileNetV2; post-training INT8 quantization typically reduces memory by 4&#x202F;&#x00D7;&#x202F;(e.g., PotatoLeafNet to 16.5&#x202F;MB, MobileNetV2 to 3.5&#x202F;MB), improving feasibility for real-time mobile/web deployment.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Comparison of model efficiency and parameter complexity.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model</th>
<th align="center" valign="top">Parameter tuning method</th>
<th align="center" valign="top">Training time (h)</th>
<th align="center" valign="top">Parameter count (Millions)</th>
<th align="left" valign="top">Hyperparameter complexity</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Proposed PotatoLeafNet</td>
<td align="left" valign="top">Manual tuning with ReduceLROnPlateau + ModelCheckpoint; fixed sequential augmentation; Adam (lr&#x202F;=&#x202F;1e-3)</td>
<td align="char" valign="top" char=".">1.2</td>
<td align="char" valign="top" char=".">16.5&#x202F;M</td>
<td align="left" valign="top">Moderate</td>
</tr>
<tr>
<td align="left" valign="top">ResNet-50&#x202F;+&#x202F;VGG-16</td>
<td align="left" valign="top">Transfer Learning + Fine-Tuning</td>
<td align="char" valign="top" char=".">1.5</td>
<td align="char" valign="top" char=".">138.36&#x202F;M</td>
<td align="left" valign="top">High</td>
</tr>
<tr>
<td align="left" valign="top">VGG-16&#x202F;+&#x202F;MobileNetV2</td>
<td align="left" valign="top">Transfer Learning + Grid Search</td>
<td align="char" valign="top" char=".">1.8</td>
<td align="char" valign="top" char=".">141.67&#x202F;M</td>
<td align="left" valign="top">High</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV2</td>
<td align="left" valign="top">Random Search</td>
<td align="char" valign="top" char=".">0.8</td>
<td align="char" valign="top" char=".">3.54&#x202F;M</td>
<td align="left" valign="top">Low</td>
</tr>
<tr>
<td align="left" valign="top">Inception-V3</td>
<td align="left" valign="top">Neural Architecture Search (NAS)</td>
<td align="char" valign="top" char=".">1.2</td>
<td align="char" valign="top" char=".">23.85&#x202F;M</td>
<td align="left" valign="top">Moderate</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="sec21">
<label>4.4</label>
<title>Comparison of accuracy between proposed method and existing studies</title>
<p><xref ref-type="table" rid="tab5">Table 5</xref> benchmarks the proposed PotatoLeafNet against recent potato-disease studies spanning handcrafted descriptors with classical classifiers (Ala&#x2019;a), transfer-learned CNNs (<xref ref-type="bibr" rid="ref27">Nur et al., 2025</xref>; <xref ref-type="bibr" rid="ref40">Shah et al., 2025</xref>), compact bespoke CNNs (<xref ref-type="bibr" rid="ref20">Kaur et al., 2025</xref>; <xref ref-type="bibr" rid="ref36">Salihu et al., 2025</xref>), hybrid CNN&#x2013;Transformer designs (<xref ref-type="bibr" rid="ref41">Sinamenye et al., 2025</xref>; <xref ref-type="bibr" rid="ref49">Zhang et al., 2025</xref>), and a non-image tabular risk-forecasting approach (<xref ref-type="bibr" rid="ref32">Radwan et al., 2025</xref>). Despite heterogeneity in data sources and class definitions, PotatoLeafNet attains 98.52% accuracy on PlantVillage (Healthy/Early/Late), placing it among the top performers while using a compact 11-layer 3&#x202F;&#x00D7;&#x202F;3 convolutional stack and a fixed sequential photometric augmentation policy. Notably, several comparators optimize for different modalities (e.g., meteorological risk factors) or field-like imagery; therefore, results are indicative rather than strictly commensurate, and cross-dataset validation remains essential for assessing real-world robustness.</p>
<table-wrap position="float" id="tab5">
<label>Table 5</label>
<caption>
<p>Comparison analysis of PotatoLeafNet with existing studies on potato leaf disease dataset.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Study [Ref]</th>
<th align="left" valign="top">Year</th>
<th align="left" valign="top">Model/approach</th>
<th align="left" valign="top">Dataset (classes)</th>
<th align="center" valign="top">Reported<break/>accuracy (%)</th>
<th align="left" valign="top">Core techniques/notes</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref32">Radwan et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">MLP with binary Greylag Goose Optimization for feature selection, Compared with LR, SVM, KNN, Gradient Boosting</td>
<td align="left" valign="top">4,000<break/>meteorological records for early and late blight risk</td>
<td align="char" valign="top" char=".">98.30</td>
<td align="left" valign="top">K-means, PCA, copula analysis for structure, tabular risk forecasting complementary to image-based screening</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref9">Chen and Liu (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">CBSNet with Channel Reconstruction Multi-Scale convolution Spatial Triple Attention, Bat-Lion strategy</td>
<td align="left" valign="top">Self-built photo leaf set (Healthy, Early blight, Late blight)</td>
<td align="char" valign="top" char=".">92.04</td>
<td align="left" valign="top">Target tiny lesions, blurred edges, and noise, attention driven multi-scale feature extraction</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref41">Sinamenye et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">Hybrid EfficientNetV2-B3 with vision Transformer</td>
<td align="left" valign="top">Potato Leaf Disease Dataset (field-like variability)</td>
<td align="char" valign="top" char=".">85.06</td>
<td align="left" valign="top">Combines local convolutional features and global transformer context for generalization</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref36">Salihu et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">CNN trained with Adam</td>
<td align="left" valign="top">Curated set (Healthy, Early blight, Late blight)</td>
<td align="char" valign="top" char=".">96.88</td>
<td align="left" valign="top">Scaling augmentation, normalization, confusion-matrix based evaluation</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref3">Ala&#x2019;a (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">Generalized Jones polynomial features with SVM Classifier</td>
<td align="left" valign="top">Plant Village (Potato)</td>
<td align="char" valign="top" char=".">98.45</td>
<td align="left" valign="top">Pipeline: Preprocessing, GJP feature extraction, dimensionality reduction, SVM, Strong handcrafted descriptor baseline</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref49">Zhang et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">VGG16S (GAP with CBAM and Leaky ReLU, about 15 M parameters)</td>
<td align="left" valign="top">Early blight and viral disease set with augmentation</td>
<td align="char" valign="top" char=".">97.87</td>
<td align="left" valign="top">Response-Surface hyperparameter tuning, ablations and public dataset tests reported</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref20">Kaur et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">PotConvNet (Compact CNN)</td>
<td align="left" valign="top">Two Potato datasets with defined splits</td>
<td align="char" valign="top" char=".">97.78</td>
<td align="left" valign="top">Resizing, normalization, augmentation, high accuracy on Dataset 1 and Strong cross-dataset results</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref27">Nur et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">Inceptionv3 with transfer learning and targeted fine tunning</td>
<td align="left" valign="top">Domain-Specific potato leaf set</td>
<td align="char" valign="top" char=".">97.78</td>
<td align="left" valign="top">Fine-tuned terminal layers, efficient and practical</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref40">Shah et al. (2025)</xref>
</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">PLDC-Net with EfficientNet-B1 backbone and SVM Classifier</td>
<td align="left" valign="top">Balanced multi-disease image set from online sources</td>
<td align="char" valign="top" char=".">98.39</td>
<td align="left" valign="top">Emphasizes data balancing and robust augmentation, evaluated on unseen images</td>
</tr>
<tr>
<td align="left" valign="top">Proposed model</td>
<td align="left" valign="top">2025</td>
<td align="left" valign="top">PotatoLeafNet</td>
<td align="left" valign="top">Plant Village (Potato)</td>
<td align="char" valign="top" char=".">98.52</td>
<td align="left" valign="top">Photometric transforms applied before learning; same policy across classes.11-layer convolutional stack with 3&#x202F;&#x00D7;&#x202F;3 kernels (PotatoLeafNet)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec22">
<label>4.5</label>
<title>Comparative performance</title>
<p>Under the same training&#x2013;evaluation protocol, the Proposed model achieves 98.52% accuracy, 98.67% precision, 99.67% recall, 99.16% F1-score, and 1.00 AUC. Relative to ResNet-50&#x202F;+&#x202F;VGG-16 (97.10, 95.00, 94.00, 94.00%, 0.98), this corresponds to absolute gains of +1.42&#x202F;pp. accuracy, +3.67&#x202F;pp. precision, +5.67&#x202F;pp. recall, +5.16&#x202F;pp&#x202F;F1, and +0.02 AUC <xref ref-type="table" rid="tab6">Table 6</xref>. Against VGG-16&#x202F;+&#x202F;MobileNetV2 (94.80, 92.00, 91.00, 91.00%, 0.93), the gains are +3.72&#x202F;pp, +6.67&#x202F;pp, +8.67&#x202F;pp, +8.16&#x202F;pp, and +0.07. Versus MobileNetV2 (93.20, 91.00, 90.00, 90.00%, 0.92), the improvements are +5.32&#x202F;pp, +7.67&#x202F;pp, +9.67&#x202F;pp, +9.16&#x202F;pp, and +0.08; and versus Inception-V3 (92.50, 90.00, 89.00, 89.00%, 0.91), they are +6.02&#x202F;pp, +8.67&#x202F;pp, +10.67&#x202F;pp, +10.16&#x202F;pp, and +0.09. The consistent, largest margin in recall indicates the Proposed model substantially reduces false negatives critical for early disease screening while simultaneously delivering the best precision, F1, and AUC among all baselines.</p>
<table-wrap position="float" id="tab6">
<label>Table 6</label>
<caption>
<p>Comparative performance of the proposed model and deep learning.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model</th>
<th align="center" valign="top">Accuracy (%)</th>
<th align="left" valign="top">Precision (%)</th>
<th align="center" valign="top">Recall (%)</th>
<th align="center" valign="top">F1-score (%)</th>
<th align="center" valign="top">AUC</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Proposed model</td>
<td align="char" valign="top" char=".">98.52</td>
<td align="char" valign="top" char=".">98.67</td>
<td align="char" valign="top" char=".">99.67</td>
<td align="char" valign="top" char=".">99.16</td>
<td align="char" valign="top" char=".">1.00</td>
</tr>
<tr>
<td align="left" valign="top">ResNet-50&#x202F;+&#x202F;VGG-16</td>
<td align="char" valign="top" char=".">97.1</td>
<td align="char" valign="top" char=".">0.95</td>
<td align="char" valign="top" char=".">0.94</td>
<td align="char" valign="top" char=".">0.94</td>
<td align="char" valign="top" char=".">0.98</td>
</tr>
<tr>
<td align="left" valign="top">VGG-16&#x202F;+&#x202F;MobileNetV2</td>
<td align="char" valign="top" char=".">94.8</td>
<td align="char" valign="top" char=".">0.92</td>
<td align="char" valign="top" char=".">0.91</td>
<td align="char" valign="top" char=".">0.91</td>
<td align="char" valign="top" char=".">0.93</td>
</tr>
<tr>
<td align="left" valign="top">MobileNetV2</td>
<td align="char" valign="top" char=".">93.2</td>
<td align="char" valign="top" char=".">0.91</td>
<td align="char" valign="top" char=".">0.90</td>
<td align="char" valign="top" char=".">0.90</td>
<td align="char" valign="top" char=".">0.92</td>
</tr>
<tr>
<td align="left" valign="top">Inception-V3</td>
<td align="char" valign="top" char=".">92.5</td>
<td align="char" valign="top" char=".">0.90</td>
<td align="char" valign="top" char=".">0.89</td>
<td align="char" valign="top" char=".">0.89</td>
<td align="char" valign="top" char=".">0.91</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec23">
<label>4.6</label>
<title>PotatoLeafNet model performance on correctly predicted images</title>
<p>The classification accuracy of the model on correctly labeled images reflects the strength of the proposed CNN framework in distinguishing between diseased and non-diseased leaf samples. A correctly identified instance refers to an image that the model assigns the appropriate label to, whether the leaf is affected or unaffected. This aspect of performance was evaluated using standard accuracy-based metrics. The model consistently delivered accurate predictions across all categories, showcasing its reliability in handling both training and unseen test samples. Its ability to differentiate between various visual patterns linked to disease manifestations underlines its robustness and generalizability. The successful identification of all leaf conditions confirms the framework&#x2019;s precision and operational reliability in practical settings. <xref ref-type="fig" rid="fig10">Figure 10</xref> provides a visual representation of the model&#x2019;s performance in identifying each class correctly, further validating its strength in class-wise prediction and its potential for real-world application in automated plant disease assessment systems.</p>
<fig position="float" id="fig10">
<label>Figure 10</label>
<caption>
<p>PotatoLeafNet Model performance on correctly predicted images.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g010.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Grid of sixteen leaves, each labeled with true and predicted disease states: "Late Blight", "Early Blight", or "Healthy". Some leaves show brown spots or discoloration, indicating blight, while others appear healthy.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec24">
<label>4.7</label>
<title>Performance measures on individual diseases prediction and classification</title>
<p>Collectively, these metrics provide a comprehensive view of class-wise prediction performance. Training the PotatoLeafNet model for 100 epochs yielded strong results: precision of 98.00% for Early blight, 99.00% for Late blight, and 99.00% for Healthy leaves; recall of 100.00, 99.00, and 100.00%, respectively; and F1-scores of 99.00, 99.00, and 99.50% for Early blight, Late blight, and Healthy leaves, respectively, corresponding to a macro-averaged F1 of 99.16%. Taken together, these indicators suggest that PotatoLeafNet accurately predicts and classifies potato leaf disease categories. At the same time, <xref ref-type="table" rid="tab5">Table 5</xref> indicates that Early and Late Blight can be less reliably predicted under challenging conditions such as out-of-distribution inputs, limited representativeness in the training data, noisy or ambiguous images, and potential overfitting whereas Healthy leaves are generally classified more accurately. <xref ref-type="fig" rid="fig11">Figure 11</xref> presents the confusion matrix for a small held-out test subset after 100 training epochs (Early Blight&#x202F;=&#x202F;4, Late Blight&#x202F;=&#x202F;6, Healthy&#x202F;=&#x202F;2). All instances were correctly identified, with no false positives or negatives, corresponding to 99% accuracy and precision/recall of &#x003E;0.98 for each class on that subset. However, given the limited sample size, these perfect results should be interpreted cautiously and validated on larger, more diverse datasets to confirm generalization (<xref ref-type="table" rid="tab7">Table 7</xref>).</p>
<fig position="float" id="fig11">
<label>Figure 11</label>
<caption>
<p>Confusion matrix for the PotatoLeafNet model on predicted individual disease classes after 100 epochs.</p>
</caption>
<graphic xlink:href="frai-08-1668839-g011.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Confusion matrix illustrating classification results. The actual classes on the y-axis are Healthy, Early Blight, and Late Blight. The predicted classes on the x-axis are Healthy, Early Blight, and Late Blight. It shows 2 correct predictions for Healthy, 4 for Early Blight, and 6 for Late Blight, with zero incorrect predictions.</alt-text>
</graphic>
</fig>
<table-wrap position="float" id="tab7">
<label>Table 7</label>
<caption>
<p>Performance measures on individual potato leaf diseases classes.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">S. No</th>
<th align="center" valign="top">Class</th>
<th align="center" valign="top">Precision (%)</th>
<th align="center" valign="top">Recall (%)</th>
<th align="center" valign="top">F1-Score (%)</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Early blight</td>
<td align="char" valign="top" char=".">98.00</td>
<td align="char" valign="top" char=".">100.00</td>
<td align="char" valign="top" char=".">99.00</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Late blight</td>
<td align="char" valign="top" char=".">99.00</td>
<td align="char" valign="top" char=".">99.00</td>
<td align="char" valign="top" char=".">99.00</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="left" valign="top">Healthy leaves</td>
<td align="char" valign="top" char=".">99.00</td>
<td align="char" valign="top" char=".">100.00</td>
<td align="char" valign="top" char=".">99.50</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec25">
<label>4.8</label>
<title>Discussion</title>
<p>Potato production underpins global food security, yet yields and quality are threatened by diverse foliar diseases whose early diagnosis is complicated by cultivar heterogeneity, variable symptom expression, and environmental noise, making rapid and accurate detection essential. To address this need, we propose PotatoLeafNet, a two-stage convolutional framework for automated identification of potato leaf conditions. In the first stage, a fixed sequential image-augmentation pipeline expands intra-class variability and mitigates overfitting; in the second, an 11-layer CNN with 3&#x202F;&#x00D7;&#x202F;3 kernels learns discriminative morphological and textural representations from the augmented images. Evaluated on the enhanced dataset, PotatoLeafNet achieved an overall accuracy of 98.92%, with complementary performance measures confirming its ability to correctly categorize samples. In comparative analyses, the approach outperformed representative state-of-the-art baselines and consistently predicted Late Blight, Early Blight, and Healthy classes with high reliability. By enabling precise differentiation among these categories, PotatoLeafNet facilitates timely intervention and supports evidence-based disease-management strategies in real-world agronomic settings.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="sec26">
<label>5</label>
<title>Conclusion</title>
<p>This work introduced PotatoLeafNet, a two-stage convolutional framework that couples a fixed sequential image-augmentation pipeline with an 11-layer, 3&#x202F;&#x00D7;&#x202F;3-kernel CNN to deliver reliable detection and classification of potato leaf conditions (Healthy, Early Blight, Late Blight). Trained on an augmented and class-balanced dataset, PotatoLeafNet achieved 98.92% overall accuracy under the 100-epoch setting and maintained strong agreement between training, validation, and independent test splits; even in a constrained 10-epoch regime it sustained competitive generalization (training 88.22%, validation 86.91%, test 88.15%). In head-to-head comparisons on the same dataset, PotatoLeafNet consistently outperformed representative CNN baselines, indicating that the combination of sequential augmentation and a compact convolutional stack yields discriminative, disease-relevant representations without sacrificing computational efficiency. Practically, these attributes make PotatoLeafNet a strong candidate for field deployment in resource-limited settings (e.g., mobile or edge devices), where rapid, accurate triage can enable timely intervention, reduce losses, and support data-driven integrated pest management. While the results are robust, two limitations merit attention. First, performance was established on curated images; domain shift in truly in-situ imagery (lighting variation, occlusion, mixed infections, cultivar differences) can degrade accuracy. Second, the present evaluation emphasizes aggregate metrics; class-wise calibration, error analysis, and explainability are essential before widescale adoption. Addressing these gaps will strengthen external validity and user trust. Future work will expand training with diverse, field-acquired datasets and explicitly address domain shift through domain-generalization techniques such as style transfer and test-time adaptation; provide fine-grained diagnostics including per-class precision and recall, confusion matrices, and confidence calibration together with explainability analyses (Grad-CAM/saliency) to verify that decisions focus on pathognomonic regions; conduct ablation studies to isolate the contribution of each augmentation transform and architectural component; examine robustness under label noise and data drift across seasons and geographies; and prototype a lightweight, on-device inference stack employing batching and quantization to validate throughput and latency in real agronomic workflows, thereby advancing PotatoLeafNet from a high-performing classifier to a deployable decision-support tool.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec27">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="sec28">
<title>Author contributions</title>
<p>GB: Resources, Validation, Writing &#x2013; original draft, Methodology, Data curation. MC: Funding acquisition, Writing &#x2013; review &#x0026; editing, Supervision, Validation, Conceptualization.</p>
</sec>
<sec sec-type="COI-statement" id="sec29">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec30">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec31">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Afzaal</surname><given-names>H.</given-names></name> <name><surname>Farooq</surname><given-names>M. S.</given-names></name> <name><surname>Rehman</surname><given-names>A. U.</given-names></name> <name><surname>Sultana</surname><given-names>S.</given-names></name> <name><surname>Zafar</surname><given-names>A.</given-names></name> <name><surname>Habib</surname><given-names>M. A.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Detection of a potato disease (early blight) using artificial intelligence</article-title>. <source>Remote Sens</source> <volume>13</volume>:<fpage>411</fpage>. doi: <pub-id pub-id-type="doi">10.3390/rs13030411</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ahmed</surname><given-names>M. M.</given-names></name> <name><surname>Aftab</surname><given-names>R. S.</given-names></name> <name><surname>Hamim</surname><given-names>S. A.</given-names></name> <name><surname>Abdullah-Al-Jubair</surname><given-names>M.</given-names></name> <name><surname>Nandi</surname><given-names>D.</given-names></name></person-group> (<year>2025</year>). &#x201C;<article-title>Harnessing convolutional neural networks for potato leaf disease detection: a proposed model</article-title>&#x201D; in <source>Machine vision in plant leaf disease detection for sustainable agriculture</source>. eds. <person-group person-group-type="editor"><name><surname>Mridha</surname><given-names>M. F.</given-names></name> <name><surname>Dey</surname><given-names>N.</given-names></name></person-group> (<publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer Nature Singapore</publisher-name>), <fpage>91</fpage>&#x2013;<lpage>103</lpage>.</mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ala&#x2019;a</surname><given-names>R.</given-names></name></person-group> (<year>2025</year>). <article-title>Potato leaves disease classification based on generalized Jones polynomials image features</article-title>. <source>MethodsX</source> <volume>14</volume>:<fpage>103421</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.mex.2025.103421</pub-id>, <pub-id pub-id-type="pmid">40567945</pub-id></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alhammad</surname><given-names>S. M.</given-names></name> <name><surname>Khafaga</surname><given-names>D. S.</given-names></name> <name><surname>El-Hady</surname><given-names>W. M.</given-names></name> <name><surname>Samy</surname><given-names>F. M.</given-names></name> <name><surname>Hosny</surname><given-names>K. M.</given-names></name></person-group> (<year>2025</year>). <article-title>Deep learning and explainable AI for classification of potato leaf diseases</article-title>. <source>Front. Artif. Intell.</source> <volume>7</volume>:<fpage>1449329</fpage>. doi: <pub-id pub-id-type="doi">10.3389/frai.2024.1449329</pub-id>, <pub-id pub-id-type="pmid">39963448</pub-id></mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Alzakari</surname><given-names>S. A.</given-names></name> <name><surname>Alhussan</surname><given-names>A. A.</given-names></name> <name><surname>Qenawy</surname><given-names>A. S. T.</given-names></name> <name><surname>Elshewey</surname><given-names>A. M.</given-names></name></person-group> (<year>2025</year>). <article-title>Early detection of potato disease using an enhanced convolutional neural network-long short-term memory deep learning model</article-title>. <source>Potato Res.</source> <volume>68</volume>, <fpage>695</fpage>&#x2013;<lpage>713</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11540-024-09760-x</pub-id>, <pub-id pub-id-type="pmid">41328310</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Awal</surname><given-names>M. A.</given-names></name> <name><surname>Roy</surname><given-names>K.</given-names></name> <name><surname>Rahman</surname><given-names>M. M.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Potato leaf disease recognition using deep learning</article-title>&#x201D; in <source>Proceedings of the 2019 international conference on robotics, electrical and signal processing techniques (ICREST)</source>. (<publisher-loc>Piscataway</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>120</fpage>&#x2013;<lpage>125</lpage>.</mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bappi</surname><given-names>I.</given-names></name> <name><surname>Richter</surname><given-names>D. J.</given-names></name> <name><surname>Kim</surname><given-names>K.</given-names></name></person-group> (<year>2025</year>). <article-title>Assessing the effectiveness of augmentation techniques in enhancing plant leaf disease classification</article-title>. <source>Smart Media J.</source> <volume>14</volume>, <fpage>17</fpage>&#x2013;<lpage>25</lpage>. doi: <pub-id pub-id-type="doi">10.30693/SMJ.2025.14.1.17</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Barman</surname><given-names>U.</given-names></name> <name><surname>Choudhury</surname><given-names>R. D.</given-names></name> <name><surname>Sahu</surname><given-names>D.</given-names></name> <name><surname>Borah</surname><given-names>S.</given-names></name> <name><surname>Das</surname><given-names>R.</given-names></name> <name><surname>Roy</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2020</year>). &#x201C;<article-title>Comparative assessment of deep learning to detect the leaf diseases of potato based on data augmentation</article-title>&#x201D; in <source>Proceedings of the 2020 international conference on computational performance evaluation (ComPE); 2020 Jul 2&#x2013;4</source>, <publisher-loc>Shillong, India</publisher-loc>: <publisher-name>IEEE</publisher-name>. <fpage>682</fpage>&#x2013;<lpage>687</lpage>.</mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname><given-names>Y.</given-names></name> <name><surname>Liu</surname><given-names>W.</given-names></name></person-group> (<year>2025</year>). <article-title>CBSNet: an effective method for potato leaf disease classification</article-title>. <source>Plants</source> <volume>14</volume>:<fpage>632</fpage>. doi: <pub-id pub-id-type="doi">10.3390/plants14050632</pub-id>, <pub-id pub-id-type="pmid">40094555</pub-id></mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dey</surname><given-names>T. K.</given-names></name> <name><surname>Pradhan</surname><given-names>J.</given-names></name> <name><surname>Khan</surname><given-names>D. A.</given-names></name></person-group> (<year>2025</year>). <article-title>Optimized potato leaf disease detection with an enhanced convolutional neural network</article-title>. <source>IETE J. Res.</source> <volume>71</volume>, <fpage>1777</fpage>&#x2013;<lpage>1790</lpage>. doi: <pub-id pub-id-type="doi">10.1080/03772063.2025.2467761</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ferentinos</surname><given-names>K. P.</given-names></name></person-group> (<year>2018</year>). <article-title>Deep learning models for plant disease detection and diagnosis</article-title>. <source>Comput. Electron. Agric.</source> <volume>145</volume>, <fpage>311</fpage>&#x2013;<lpage>318</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2018.01.009</pub-id></mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fuentes</surname><given-names>A.</given-names></name> <name><surname>Yoon</surname><given-names>S.</given-names></name> <name><surname>Kim</surname><given-names>S.</given-names></name> <name><surname>Park</surname><given-names>D.</given-names></name></person-group> (<year>2017</year>). <article-title>A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition</article-title>. <source>Sensors</source> <volume>17</volume>:<fpage>2022</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s17092022</pub-id>, <pub-id pub-id-type="pmid">28869539</pub-id></mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Geetharamani</surname><given-names>G.</given-names></name> <name><surname>Pandian</surname><given-names>A.</given-names></name></person-group> (<year>2019</year>). <article-title>Identification of plant leaf diseases using a nine-layer deep convolutional neural network</article-title>. <source>Comput. Electr. Eng.</source> <volume>76</volume>, <fpage>323</fpage>&#x2013;<lpage>338</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compeleceng.2019.04.011</pub-id></mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Ghosal</surname><given-names>S.</given-names></name> <name><surname>Bandyopadhyay</surname><given-names>S.</given-names></name> <name><surname>Sahidullah</surname><given-names>M.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Deep learning models for early blight disease detection in potato leaves</article-title>&#x201D; in <source>Proceedings of the 2019 2nd international conference on advanced computational and communication paradigms (ICACCP)</source>. (<publisher-loc>Piscataway</publisher-loc>: <publisher-name>IEEE</publisher-name>), <fpage>1</fpage>&#x2013;<lpage>6</lpage>.</mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gupta</surname><given-names>A.</given-names></name> <name><surname>Gupta</surname><given-names>R.</given-names></name> <name><surname>Gupta</surname><given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Deep learning-based potato leaf disease identification using augmentation techniques</article-title>. <source>J. Agric. Sci.</source> <volume>7</volume>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>.</mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Gurucharan</surname><given-names>M. K</given-names></name></person-group>. (<year>2020</year>). Basic CNN architecture: Explaining 5 layers of convolutional neural network. UpGrad Blog. Available online at: <ext-link xlink:href="https://www.upgrad.com/blog/basic-cnn-architecture" ext-link-type="uri">https://www.upgrad.com/blog/basic-cnn-architecture</ext-link></mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hernandez-Valencia</surname><given-names>E.</given-names></name> <name><surname>Ramirez-Pedraza</surname><given-names>A.</given-names></name> <name><surname>Morales-Sandoval</surname><given-names>M.</given-names></name> <name><surname>Sossa-Azuela</surname><given-names>J. H.</given-names></name> <name><surname>Castro-Espinoza</surname><given-names>F.</given-names></name> <name><surname>Aceves-Fernandez</surname><given-names>M. A.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Lossless compression for multispectral images of agricultural products based on 3D DCT</article-title>. <source>J. Appl. Res. Technol.</source> <volume>18</volume>, <fpage>301</fpage>&#x2013;<lpage>309</lpage>.</mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hou</surname><given-names>C.</given-names></name> <name><surname>Zhuang</surname><given-names>J.</given-names></name> <name><surname>Tang</surname><given-names>Y.</given-names></name> <name><surname>He</surname><given-names>Y.</given-names></name> <name><surname>Miao</surname><given-names>A.</given-names></name> <name><surname>Huang</surname><given-names>H.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Recognition of early blight and late blight diseases on potato leaves based on graph cut segmentation</article-title>. <source>J. Agric. Food Res.</source> <volume>5</volume>:<fpage>100154</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jafr.2021.100154</pub-id></mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jafar</surname><given-names>A.</given-names></name> <name><surname>Bibi</surname><given-names>N.</given-names></name> <name><surname>Naqvi</surname><given-names>R. A.</given-names></name> <name><surname>Sadeghi-Niaraki</surname><given-names>A.</given-names></name> <name><surname>Jeong</surname><given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Revolutionizing agriculture with artificial intelligence: plant disease detection methods, applications, and their limitations</article-title>. <source>Front. Plant Sci.</source> <volume>15</volume>:<fpage>1356260</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2024.1356260</pub-id>, <pub-id pub-id-type="pmid">38545388</pub-id></mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kaur</surname><given-names>K.</given-names></name> <name><surname>Kaur</surname><given-names>H.</given-names></name> <name><surname>Singh</surname><given-names>M. L.</given-names></name> <name><surname>Singh</surname><given-names>R.</given-names></name></person-group> (<year>2025</year>). <article-title>PotConvNet: an automated deep convolutional neural network-based framework for identification of potato leaf diseases</article-title>. <source>Potato Res.</source> <volume>68</volume>, <fpage>1</fpage>&#x2013;<lpage>36</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11540-025-09764-9</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Khalifa</surname><given-names>N. E. M.</given-names></name> <name><surname>Taha</surname><given-names>M. H. N.</given-names></name> <name><surname>Hassanien</surname><given-names>A. E.</given-names></name> <name><surname>Elghamrawy</surname><given-names>S.</given-names></name> <name><surname>Ghazal</surname><given-names>M.</given-names></name> <name><surname>Chetouani</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2021</year>). &#x201C;<article-title>Artificial intelligence in potato leaf disease classification: a deep learning approach</article-title>&#x201D; in <source>Machine learning and big data analytics paradigms: analysis, applications and challenges</source>. eds. <person-group person-group-type="editor"><name><surname>Singh</surname><given-names>A.</given-names></name> <name><surname>Kumar</surname><given-names>A.</given-names></name> <name><surname>Sharma</surname><given-names>S.</given-names></name></person-group> (<publisher-loc>Berlin/Heidelberg</publisher-loc>: <publisher-name>Springer</publisher-name>), <fpage>63</fpage>&#x2013;<lpage>79</lpage>.</mixed-citation></ref>
<ref id="ref22"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Lee</surname><given-names>T. Y.</given-names></name> <name><surname>Chan</surname><given-names>C. S.</given-names></name> <name><surname>Mayo</surname><given-names>S. J.</given-names></name> <name><surname>Remagnino</surname><given-names>P.</given-names></name> <name><surname>Ahmed</surname><given-names>F.</given-names></name> <name><surname>Lim</surname><given-names>K. S.</given-names></name> <etal/></person-group>. (<year>2020</year>). &#x201C;<article-title>Health detection for potato leaf with convolutional neural network</article-title>&#x201D; in <source>Proceedings of the 2020 Indo&#x2013;Taiwan 2nd international conference on computing, analytics and networks (Indo-Taiwan ICAN); 2020 Feb 7&#x2013;15</source>, <publisher-loc>Rajpura, India</publisher-loc>: <publisher-name>IEEE</publisher-name>. <fpage>289</fpage>&#x2013;<lpage>293</lpage>.</mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>X.</given-names></name> <name><surname>Zhou</surname><given-names>Y.</given-names></name> <name><surname>Liu</surname><given-names>J.</given-names></name> <name><surname>Wang</surname><given-names>L.</given-names></name> <name><surname>Zhang</surname><given-names>J.</given-names></name> <name><surname>Fan</surname><given-names>X.</given-names></name></person-group> (<year>2022</year>). <article-title>The detection method of potato foliage diseases in complex background based on instance segmentation and semantic segmentation</article-title>. <source>Front. Plant Sci.</source> <volume>13</volume>:<fpage>899754</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2022.899754</pub-id>, <pub-id pub-id-type="pmid">35865287</pub-id></mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liang</surname><given-names>Q.</given-names></name> <name><surname>Xiang</surname><given-names>S.</given-names></name> <name><surname>Hu</surname><given-names>Y.</given-names></name> <name><surname>Coppola</surname><given-names>G.</given-names></name> <name><surname>Zhang</surname><given-names>D.</given-names></name> <name><surname>Sun</surname><given-names>W.</given-names></name></person-group> (<year>2019</year>). <article-title>PD2SE-net: computer-assisted plant disease diagnosis and severity estimation network</article-title>. <source>Comput. Electron. Agric.</source> <volume>157</volume>, <fpage>518</fpage>&#x2013;<lpage>529</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2019.01.034</pub-id></mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>J.</given-names></name> <name><surname>Wang</surname><given-names>X.</given-names></name></person-group> (<year>2021</year>). <article-title>Plant diseases and pest's detection based on deep learning: a review</article-title>. <source>Plant Methods</source> <volume>17</volume>:<fpage>22</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s13007-021-00722-9</pub-id>, <pub-id pub-id-type="pmid">33627131</pub-id></mixed-citation></ref>
<ref id="ref26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mishra</surname><given-names>S. K.</given-names></name> <name><surname>Srivastava</surname><given-names>S.</given-names></name></person-group> (<year>2019</year>). <article-title>Computer vision-based automated identification and classification of mango leaf diseases</article-title>. <source>J. Imaging</source> <volume>5</volume>:<fpage>38</fpage>. doi: <pub-id pub-id-type="doi">10.1504/IJGW.2023.134911</pub-id></mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nur</surname><given-names>K. N. A.</given-names></name> <name><surname>Addyna</surname><given-names>N. I.</given-names></name> <name><surname>Windarto</surname><given-names>A. P.</given-names></name> <name><surname>Wanto</surname><given-names>A.</given-names></name> <name><surname>Poningsih</surname><given-names>P.</given-names></name></person-group> (<year>2025</year>). <article-title>Optimization of the InceptionV3 architecture for potato leaf disease classification</article-title>. <source>JITK (Jurnal Ilmu Pengetahuan dan Teknologi Komputer)</source> <volume>10</volume>, <fpage>849</fpage>&#x2013;<lpage>858</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.jitk.2025.07.003</pub-id></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll1">Plant Village Dataset</collab></person-group> (<year>2024</year>). Kaggle [dataset]. Available online at: <ext-link xlink:href="https://www.kaggle.com/datasets/mohitsingh1804/plantvillage" ext-link-type="uri">https://www.kaggle.com/datasets/mohitsingh1804/plantvillage</ext-link> (Accessed April 29, 2024).</mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll2">Potato Disease Types</collab></person-group>. (<year>2025</year>). AHDB Potatoes. Available online at: <ext-link xlink:href="https://potatoes.ahdb.org.uk/knowledge-library/potato-disease-identification" ext-link-type="uri">https://potatoes.ahdb.org.uk/knowledge-library/potato-disease-identification</ext-link></mixed-citation></ref>
<ref id="ref30"><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll3">Potato Leaf Disease Dataset</collab></person-group> (<year>2025</year>). Kaggle dataset 2024. Available online at: <ext-link xlink:href="https://www.kaggle.com/datasets/rizwan123456789/potato-disease-leaf-datasetpld" ext-link-type="uri">https://www.kaggle.com/datasets/rizwan123456789/potato-disease-leaf-datasetpld</ext-link></mixed-citation></ref>
<ref id="ref31"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Powers</surname><given-names>D. M. W.</given-names></name></person-group> (<year>2020</year>). <article-title>Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation</article-title>. <source>arXiv preprint arXiv:2010.16061</source>. doi: <pub-id pub-id-type="doi">10.48550/arXiv.2010.16061</pub-id></mixed-citation></ref>
<ref id="ref32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Radwan</surname><given-names>M.</given-names></name> <name><surname>Alhussan</surname><given-names>A. A.</given-names></name> <name><surname>Ibrahim</surname><given-names>A.</given-names></name> <name><surname>Tawfeek</surname><given-names>S. M.</given-names></name></person-group> (<year>2025</year>). <article-title>Potato leaf disease classification using optimized machine learning models and feature selection techniques</article-title>. <source>Potato Res.</source> <volume>68</volume>, <fpage>897</fpage>&#x2013;<lpage>921</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11540-024-09763-8</pub-id></mixed-citation></ref>
<ref id="ref33"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rahman</surname><given-names>M. M.</given-names></name> <name><surname>Islam</surname><given-names>S. M. S.</given-names></name> <name><surname>Hossain</surname><given-names>M. E.</given-names></name> <name><surname>Ahmed</surname><given-names>K.</given-names></name></person-group> (<year>2021</year>). <article-title>Enhanced potato leaf disease classification using deep learning and data augmentation</article-title>. <source>Comput. Electron. Agric.</source> <volume>186</volume>:<fpage>106233</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2021.106233</pub-id></mixed-citation></ref>
<ref id="ref34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rathod</surname><given-names>R.</given-names></name> <name><surname>Shah</surname><given-names>R.</given-names></name> <name><surname>Patel</surname><given-names>S.</given-names></name> <name><surname>Jani</surname><given-names>A.</given-names></name> <name><surname>Mehta</surname><given-names>K.</given-names></name> <name><surname>Desai</surname><given-names>V.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Potato disease detection using deep learning and transfer learning</article-title>. <source>Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol.</source> <volume>6</volume>, <fpage>495</fpage>&#x2013;<lpage>500</lpage>.</mixed-citation></ref>
<ref id="ref35"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Rozaqi</surname><given-names>A. J.</given-names></name> <name><surname>Prasetyo</surname><given-names>E.</given-names></name> <name><surname>Nugroho</surname><given-names>H. A.</given-names></name> <name><surname>Setiawan</surname><given-names>N. A.</given-names></name> <name><surname>Santoso</surname><given-names>B.</given-names></name> <name><surname>Hidayat</surname><given-names>R.</given-names></name> <etal/></person-group>. (<year>2020</year>). &#x201C;<article-title>Identification of disease in potato leaves using convolutional neural network (CNN) algorithm</article-title>&#x201D; in <source>Proceedings of the 2020 3rd international conference on information and communications technology (ICOIACT); 2020 Nov 24&#x2013;25</source>, <publisher-loc>Yogyakarta, Indonesia</publisher-loc>: <publisher-name>IEEE</publisher-name>. <fpage>72</fpage>&#x2013;<lpage>76</lpage>.</mixed-citation></ref>
<ref id="ref36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Salihu</surname><given-names>S. A.</given-names></name> <name><surname>Adebayo</surname><given-names>S. O.</given-names></name> <name><surname>Abikoye</surname><given-names>O. C.</given-names></name> <name><surname>Usman-Hamza</surname><given-names>F. E.</given-names></name> <name><surname>Mabayoje</surname><given-names>M. A.</given-names></name> <name><surname>Brahma</surname><given-names>B.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Detection and classification of potato leaves diseases using convolutional neural network and Adam optimizer</article-title>. <source>Procedia Comput. Sci.</source> <volume>258</volume>, <fpage>2</fpage>&#x2013;<lpage>17</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.procs.2025.07.001</pub-id></mixed-citation></ref>
<ref id="ref37"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sangar</surname><given-names>G.</given-names></name> <name><surname>Rajasekar</surname><given-names>V.</given-names></name></person-group> (<year>2025</year>). <article-title>Optimized classification of potato leaf disease using EfficientNet-LITE and KE-SVM in diverse environments</article-title>. <source>Front. Plant Sci.</source> <volume>16</volume>:<fpage>1499909</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2025.1499909</pub-id>, <pub-id pub-id-type="pmid">40385236</pub-id></mixed-citation></ref>
<ref id="ref38"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sanjeev</surname><given-names>K.</given-names></name> <name><surname>Ramesh</surname><given-names>B.</given-names></name> <name><surname>Kumar</surname><given-names>V.</given-names></name> <name><surname>Singh</surname><given-names>R.</given-names></name> <name><surname>Sharma</surname><given-names>P.</given-names></name> <name><surname>Verma</surname><given-names>A.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Early prediction of potato leaf diseases using ANN classifier</article-title>. <source>Orient J Comput Sci Technol.</source> <volume>13</volume>, <fpage>2</fpage>&#x2013;<lpage>4</lpage>. doi: <pub-id pub-id-type="doi">10.13005/ojcst13.0203.11</pub-id></mixed-citation></ref>
<ref id="ref39"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shabrina</surname><given-names>N. H.</given-names></name> <name><surname>Indarti</surname><given-names>S.</given-names></name> <name><surname>Maharani</surname><given-names>R.</given-names></name> <name><surname>Kristiyanti</surname><given-names>D. A.</given-names></name> <name><surname>Irmawati</surname></name> <name><surname>Prastomo</surname><given-names>N.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>A novel dataset of potato leaf disease in uncontrolled environment</article-title>. <source>Data Brief</source> <volume>52</volume>:<fpage>109955</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.dib.2023.109955</pub-id>, <pub-id pub-id-type="pmid">38125373</pub-id></mixed-citation></ref>
<ref id="ref40"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shah</surname><given-names>S. K.</given-names></name> <name><surname>Su&#x2019;ud</surname><given-names>M. B. M.</given-names></name> <name><surname>Khan</surname><given-names>A.</given-names></name> <name><surname>Alam</surname><given-names>M. M.</given-names></name> <name><surname>Ayaz</surname><given-names>M.</given-names></name></person-group> (<year>2025</year>). <article-title>PLDC-net: potato leaf disease classification network based on an efficient convolutional neural network</article-title>. <source>Eng. Rep.</source> <volume>7</volume>:<fpage>e70178</fpage>. doi: <pub-id pub-id-type="doi">10.1002/eng2.70178</pub-id></mixed-citation></ref>
<ref id="ref41"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sinamenye</surname><given-names>J. H.</given-names></name> <name><surname>Chatterjee</surname><given-names>A.</given-names></name> <name><surname>Shrestha</surname><given-names>R.</given-names></name></person-group> (<year>2025</year>). <article-title>Potato plant disease detection: leveraging hybrid deep learning models</article-title>. <source>BMC Plant Biol.</source> <volume>25</volume>:<fpage>647</fpage>. doi: <pub-id pub-id-type="doi">10.1186/s12870-025-06679-4</pub-id>, <pub-id pub-id-type="pmid">40380088</pub-id></mixed-citation></ref>
<ref id="ref42"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Singh</surname><given-names>D.</given-names></name> <name><surname>Jain</surname><given-names>N.</given-names></name> <name><surname>Jain</surname><given-names>P.</given-names></name> <name><surname>Kayal</surname><given-names>P.</given-names></name> <name><surname>Kumawat</surname><given-names>S.</given-names></name> <name><surname>Batra</surname><given-names>N.</given-names></name></person-group> (<year>2020</year>). &#x201C;<article-title>PlantDoc: a dataset for visual plant disease detection</article-title>&#x201D; in <source>Proceedings of the 7th ACM IKDD CoDS and 25th COMAD (CoDS COMAD 2020)</source> (<publisher-loc>New York, NY, USA</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name>), <fpage>249</fpage>&#x2013;<lpage>253</lpage>.</mixed-citation></ref>
<ref id="ref43"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Sunjoyo</surname><given-names>N.</given-names></name> <name><surname>Nugroho</surname><given-names>N.</given-names></name></person-group> (<year>2022</year>). Agriculture and food. World Bank. Available online at: <ext-link xlink:href="https://www.worldbank.org/en/topic/agriculture" ext-link-type="uri">https://www.worldbank.org/en/topic/agriculture</ext-link></mixed-citation></ref>
<ref id="ref44"><mixed-citation publication-type="other"><person-group person-group-type="author"><collab id="coll4">TensorFlow Sequential Data Augmentation</collab></person-group> (<year>2025</year>). TensorFlow Tutorials. Available online at: <ext-link xlink:href="https://www.tensorflow.org/tutorials/images/data_augmentation" ext-link-type="uri">https://www.tensorflow.org/tutorials/images/data_augmentation</ext-link></mixed-citation></ref>
<ref id="ref45"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Tiwari</surname><given-names>D.</given-names></name> <name><surname>Singh</surname><given-names>A.</given-names></name> <name><surname>Sharma</surname><given-names>R.</given-names></name> <name><surname>Verma</surname><given-names>S.</given-names></name> <name><surname>Gupta</surname><given-names>P.</given-names></name> <name><surname>Mishra</surname><given-names>R.</given-names></name> <etal/></person-group>. (<year>2020</year>). &#x201C;<article-title>Potato leaf diseases detection using deep learning</article-title>&#x201D; in <source>Proceedings of the 2020 4th international conference on intelligent computing and control systems (ICICCS); 2020 May 13&#x2013;15</source>, <publisher-loc>Madurai, India</publisher-loc>: <publisher-name>IEEE</publisher-name>. <fpage>461</fpage>&#x2013;<lpage>466</lpage>.</mixed-citation></ref>
<ref id="ref46"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tugrul</surname><given-names>B.</given-names></name> <name><surname>Elfatimi</surname><given-names>E.</given-names></name> <name><surname>Eryigit</surname><given-names>R.</given-names></name></person-group> (<year>2022</year>). <article-title>Convolutional neural networks in detection of plant leaf diseases: a review</article-title>. <source>Agriculture</source> <volume>12</volume>:<fpage>1192</fpage>. doi: <pub-id pub-id-type="doi">10.3390/agriculture12081192</pub-id></mixed-citation></ref>
<ref id="ref47"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Weng</surname><given-names>L.</given-names></name> <name><surname>Tang</surname><given-names>Z.</given-names></name> <name><surname>Sardar</surname><given-names>M. F.</given-names></name> <name><surname>Yu</surname><given-names>Y.</given-names></name> <name><surname>Ai</surname><given-names>K.</given-names></name> <name><surname>Liang</surname><given-names>S.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Unveiling the frontiers of potato disease research through bibliometric analysis</article-title>. <source>Front. Microbiol.</source> <volume>15</volume>:<fpage>1430066</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fmicb.2024.1430066</pub-id>, <pub-id pub-id-type="pmid">39027102</pub-id></mixed-citation></ref>
<ref id="ref48"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Yao</surname><given-names>X.</given-names></name> <name><surname>Guan</surname><given-names>Z.</given-names></name> <name><surname>Zhou</surname><given-names>Y.</given-names></name> <name><surname>Tang</surname><given-names>J.</given-names></name> <name><surname>Hu</surname><given-names>Y.</given-names></name> <name><surname>Yang</surname><given-names>B.</given-names></name> <etal/></person-group>. (<year>2020</year>). &#x201C;<article-title>Hybrid compression algorithm for remote sensing images based on JPEG and fractal compression</article-title>&#x201D; in <source>Proceedings of the 12th international conference on measuring technology and mechatronics automation (ICMTMA)</source>, <publisher-loc>Guangzhou, China</publisher-loc>: <publisher-name>IEEE</publisher-name>. <fpage>128</fpage>&#x2013;<lpage>132</lpage>.</mixed-citation></ref>
<ref id="ref49"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>C.</given-names></name> <name><surname>Wang</surname><given-names>S.</given-names></name> <name><surname>Wang</surname><given-names>C.</given-names></name> <name><surname>Wang</surname><given-names>H.</given-names></name> <name><surname>Du</surname><given-names>Y.</given-names></name> <name><surname>Zong</surname><given-names>Z.</given-names></name></person-group> (<year>2025</year>). <article-title>Research on a potato leaf disease diagnosis system based on deep learning</article-title>. <source>Agriculture</source> <volume>15</volume>:<fpage>424</fpage>. doi: <pub-id pub-id-type="doi">10.3390/agriculture15040424</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/99920/overview">Aalt-Jan Van Dijk</ext-link>, University of Amsterdam, Netherlands</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3145629/overview">Chih-Chin Lai</ext-link>, National University of Kaohsiung, Taiwan</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3163924/overview">Meenakshi Sharma</ext-link>, Galgotias University, India</p>
</fn>
</fn-group>
</back>
</article>