<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2025.1540642</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Plant Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Automated severity level estimation of wheat rust using an EfficientNet-CBAM hybrid model</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Nigam</surname>
<given-names>Sapna</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2066046/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Jain</surname>
<given-names>Rajni</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Singh</surname>
<given-names>Vaibhav Kumar</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2712274/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Singh</surname>
<given-names>Ashish Kumar</given-names>
</name>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Krishna</surname>
<given-names>Hari</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff5">
<sup>5</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<uri xlink:href="https://loop.frontiersin.org/people/559718/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Division of Computer Applications, Indian Council of Agricultural Research (ICAR)-Indian Agricultural Statistics Research Institute</institution>, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Division of Technology and Sustainable Agriculture, Indian Council of Agricultural Research (ICAR)-National Institute of Agricultural Economics and Policy Research</institution>, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Division of Plant Pathology, Wheat Pathology Laboratory, Indian Council of Agricultural Research (ICAR)-Indian Agricultural Research Institute</institution>, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Computer Services Centre, Indian Institute of Technology Delhi</institution>, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<aff id="aff5">
<sup>5</sup>
<institution>Division of Genetics, Indian Council of Agricultural Research (ICAR)-Indian Agricultural Research Institute</institution>, <addr-line>New Delhi</addr-line>, <country>India</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Qinhu Wang, Northwest A&amp;F University, China</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Peisen Yuan, Nanjing Agricultural University, China</p>
<p>Xinli Zhou, Southwest University of Science and Technology, China</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Rajni Jain, <email xlink:href="mailto:Rajni.jain@icar.gov.in">Rajni.jain@icar.gov.in</email>; <email xlink:href="mailto:rajnijain67@gmail.com">rajnijain67@gmail.com</email>; Vaibhav Kumar Singh, <email xlink:href="mailto:Dr.singhvaibhav@gmail.com">Dr.singhvaibhav@gmail.com</email>
</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>23</day>
<month>05</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>16</volume>
<elocation-id>1540642</elocation-id>
<history>
<date date-type="received">
<day>06</day>
<month>12</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>04</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 Nigam, Jain, Singh, Singh and Krishna</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>Nigam, Jain, Singh, Singh and Krishna</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Wheat rust is a severe fungal disease that significantly impacts wheat crops, resulting in substantial losses in quality and quantity, often exceeding 50%. Timely and early accurate estimation of disease severity in fields is critical for effective disease management. Early identification of Rust at low severity levels can facilitate prompt implementation of control measures, potentially saving crops. This paper introduces an automated wheat rust severity stage estimation model utilizing the EfficientNet architecture and attention mechanism. The convolutional Block Attention Module was integrated into EfficientNet-B0 in place of the SE module to enhance feature extraction by simultaneously considering channel and spatial information. The proposed hybrid approach aims to identify rust disease severity accurately. The model is trained on an image dataset comprising three major rust types&#x2014;stripe, stem, leaf, and healthy plants captured under real-life field conditions. Each disease is categorized into four severity stages: healthy, low, medium, and high. Experimental results demonstrate that the proposed model achieves impressive performance, with a training accuracy of 99.51% and a testing accuracy of 96.68%. Moreover, comparative analysis against state-of-the-art CNN models highlights the superior performance of our approach. An Android application was also designed and developed to facilitate real-time classification of plant disease severity. This system incorporates a severity model optimized for enhanced classification accuracy and rapid recognition, ensuring efficient performance.</p>
</abstract>
<kwd-group>
<kwd>wheat rust</kwd>
<kwd>EfficientNet architecture</kwd>
<kwd>attention mechanism</kwd>
<kwd>disease severity estimation</kwd>
<kwd>transfer learning</kwd>
</kwd-group>
<counts>
<fig-count count="12"/>
<table-count count="6"/>
<equation-count count="6"/>
<ref-count count="48"/>
<page-count count="17"/>
<word-count count="6147"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Plant Pathogen Interactions</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Wheat ranks among the primary staple crops globally, with over half of its production dedicated to human consumption, livestock feed, and processing. However, wheat-producing nations face formidable challenges from plant diseases and pests, jeopardizing agricultural sustainability and profitability. Among these, rusts&#x2014;comprising stripe, leaf, and stem Rust&#x2014;stand out as particularly menacing fungal diseases, prevalent across almost all wheat-growing regions (<xref ref-type="bibr" rid="B27">Lu et al., 2017</xref>; <xref ref-type="bibr" rid="B30">Nigam and Jain, 2020</xref>). Left unchecked, rusts can mutate into virulent strains, leading to catastrophic crop failures. Conventional disease identification and severity assessment methods rely on manual visual inspection, which is fraught with inefficiencies, subjectivity, and labor intensiveness (<xref ref-type="bibr" rid="B4">Bock et&#xa0;al., 2010</xref>; <xref ref-type="bibr" rid="B29">Mohanty et al., 2016</xref>; <xref ref-type="bibr" rid="B1">Arnal Barbedo, 2019</xref>; <xref ref-type="bibr" rid="B2">Atila et&#xa0;al., 2021</xref>). Recent advances in computer vision, artificial intelligence (AI), and deep learning offer promising opportunities to automate disease detection and severity assessment through image analysis. In the existing literature, remarkable achievements in AI-based plant disease classification (<xref ref-type="bibr" rid="B22">LeCun et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B11">Fuentes et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B1">Arnal Barbedo, 2019</xref>; <xref ref-type="bibr" rid="B7">Chin et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B33">Pavithra et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B8">Dheeraj and Chand, 2022</xref>, <xref ref-type="bibr" rid="B9">2024</xref>) underscore the potential of these technologies in integrated disease management.</p>
<p>However, while significant progress has been made in plant disease detection, much of the research has primarily focused on disease type classification, leaving a critical gap in the accurate quantification of disease severity. This gap limits experts&#x2019; ability to recommend optimal pesticide applications, compromising disease control efficacy and environmental sustainability. Thus, there is a growing demand for automated disease severity classification using AI-driven approaches (<xref ref-type="bibr" rid="B10">Esgario et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B47">Wspanialy and Moussa, 2020</xref>; <xref ref-type="bibr" rid="B17">Hu et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B32">Nigam et&#xa0;al., 2021</xref>). Disease severity, a crucial parameter for assessing the intensity of plant diseases, is traditionally quantified by comparing the diseased area of a plant part (such as leaves, fruits, or stems) to the total area of the affected part, based on standardized severity grading systems (<xref ref-type="bibr" rid="B3">Bock et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B25">Liu et al., 2020</xref>). For wheat stripe rust, severity evaluation is essential for effective monitoring, but it has primarily been carried out through visual observation, a method that requires experienced assessors and is both time-consuming and prone to errors (<xref ref-type="bibr" rid="B19">Jiang et&#xa0;al., 2022</xref>). Accurately estimating lesion areas according to severity standards is challenging, further complicating the process.</p>
<p>In contrast, disease incidence, which only requires determining whether a plant part is diseased or not, is easier to assess but does not provide a precise estimate of severity. The relationship between disease incidence and severity is influenced by factors such as lesion distribution, wheat plant resistance <italic>to Puccinia striiformis (Pst</italic>), and overall incidence levels (<xref ref-type="bibr" rid="B6">Chen et al., 2014</xref>), limiting the practical utility of incidence-based severity estimation methods. Recent machine learning advancements have provided some solutions for severity estimation. For instance, <xref ref-type="bibr" rid="B45">Wang et&#xa0;al. (2017)</xref> developed a model to predict disease severity at early, medium, and final stages, achieving notable accuracy. Similarly, <xref ref-type="bibr" rid="B24">Liang et&#xa0;al. (2019)</xref> introduced the PD<sup>2</sup>SE-Net model for horticultural crops, while <xref ref-type="bibr" rid="B48">Zhao et&#xa0;al. (2021)</xref> proposed SevNet, which uses ResNet and CBAM to classify tomato disease severity with impressive accuracies of 97.59% and 95.37%, respectively.</p>
<p>Despite these advances, research on cereal crop severity estimation remains limited due to the scarcity of image datasets. Notable exceptions include the BLSNet model for rice (<xref ref-type="bibr" rid="B5">Chen et&#xa0;al., 2021</xref>) and models for maize common rust severity prediction by <xref ref-type="bibr" rid="B37">Sibiya and Sumbwanyambe (2021)</xref> and <xref ref-type="bibr" rid="B12">Haque et&#xa0;al. (2022)</xref>. Particularly underexplored is the classification of wheat yellow rust severity, with only one model&#x2014;Yellow-Rust-Xception&#x2014;proposed for differentiating yellow rust stages, achieving a modest 91% accuracy (<xref ref-type="bibr" rid="B13">Hayit et&#xa0;al., 2021</xref>). Also, <xref ref-type="bibr" rid="B19">Jiang et&#xa0;al., 2022</xref> and <xref ref-type="bibr" rid="B20">Jiang et&#xa0;al., 2023</xref> developed the machine learning models for severity assessment in wheat stripe rust. However, no deep learning-based model is developed in literature to estimate the severity of all three wheat rusts. This highlights the pressing need for further research and development to enhance the precision and reliability of disease severity assessments, particularly in wheat crops.</p>
<p>Moreover, challenges persist regarding the availability of public image databases, predominantly comprising lab-captured images rather than real-world field scenarios (<xref ref-type="bibr" rid="B28">Mi et al., 2020</xref>; <xref ref-type="bibr" rid="B31">Nigam et al., 2023</xref>). Hence, addressing these limitations is crucial for robust automated disease severity detection systems trained on datasets collected from natural field conditions.</p>
<p>Therefore, this paper focuses on the critical task of wheat disease severity classification, addressing the challenges associated with identifying and categorizing disease symptoms, understanding their impact on crop health, and exploring effective management strategies. The study emphasizes early detection of low-severity stages to mitigate crop loss and support sustainable agricultural practices. The main contributions of this research are summarized as follows:</p>
<list list-type="bullet">
<list-item>
<p>A comprehensive Wheat Disease Severity Dataset (WheatSev) was created, comprising 5,438 real-field images of wheat crops affected by stripe rust, leaf rust, and stem Rust across various growth stages.</p>
</list-item>
<list-item>
<p>A convolutional block attention module (CBAM) was integrated into the EfficientNet B0 architecture to classify wheat disease severity into three levels: low, medium, and high. The CBAM-EfficientNet model demonstrated superior classification performance compared to several established architectures, including VGGNet19, ResNet152, MobileNetV2, DenseNet169, InceptionV3, and the original EfficientNet B0.</p>
</list-item>
<list-item>
<p>The proposed model significantly improved classification performance in terms of accuracy, recall, precision, and F1 score by leveraging the combined strengths of EfficientNet B0 and CBAM layers. This approach effectively addressed technical challenges such as vanishing gradients and computational complexity while enhancing the robustness of the model.</p>
</list-item>
<list-item>
<p>Robust data augmentation techniques were employed to increase data diversity, mitigating the risk of overfitting and ensuring the model&#x2019;s reliability in classifying diverse real-world samples.</p>
</list-item>
<list-item>
<p>The model&#x2019;s efficiency was validated through extensive hyperparameter tuning, comparative analyses with state-of-the-art architectures, and Grad-CAM visualizations. Experimental results underscored the effectiveness of the CBAM-EfficientNet B0 model for accurate wheat disease severity estimation.</p>
</list-item>
</list>
</sec>
<sec id="s2" sec-type="materials|methods">
<label>2</label>
<title>Materials and methods</title>
<p>This section delineates the tools, techniques, and procedures employed in the present study, including the acquisition of the image dataset and the proposed framework incorporating an attention module within the EfficientNet architecture.</p>
<sec id="s2_1">
<label>2.1</label>
<title>Image dataset</title>
<p>Images depicting wheat rusts at various severity level stages were captured within the fields of ICAR-Indian Agricultural Research Institute, New Delhi, India, spanning three consecutive crop seasons. Image acquisition took place during sunny conditions, typically between 11:00 am to 1:00 pm, at ten-day intervals following the initial onset of disease symptoms. This timing ensured consistent leaf growth stages across all captured images.</p>
<p>A handheld mobile camera with a 20-megapixel resolution and a 25mm wide-angle lens was utilized for image acquisition. The deliberate use of mobile devices instead of professional cameras aimed to mirror the tools commonly employed by farmers in similar scenarios. The severity level estimation dataset encompasses images of three major wheat diseases, categorized into three severity levels by plant pathologists: low, medium, high and healthy (<xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1</bold>
</xref>). Images were captured by directing the camera lens toward regions of the leaves exhibiting disease symptoms at various growth stages. Subsequently, pathologists labeled the severity levels based on the percentage of disease symptoms present. Images categorized as low severity level exhibited severity symptoms ranging from 0 to 25 percent, those classified as medium severity level ranged from 25 to 50 percent, and high severity level images displayed more than 50 percent disease symptoms. The images with no disease symptoms were considered healthy as shown in <xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1</bold>
</xref>.</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Wheat rusts images at severity levels: Low, Medium, High and Healthy.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g001.tif"/>
</fig>
<p>The primary emphasis of this study lies in low-level severity image classification to facilitate timely disease detection and mitigate crop loss. Initially, the severity stage estimation dataset comprised 5438 images distributed across ten classes, including three rust severity stages for each disease class and a class representing healthy leaves (refer to <xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>). To enhance classification performance and achieve balance among disease classes, the original dataset underwent augmentation, resulting in a total of 10252 images, with 1000 images allocated to each disease-infected class.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Description of image dataset for severity stage estimation.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">Disease</th>
<th valign="top" align="left">Class Name</th>
<th valign="top" align="left"># Original images</th>
<th valign="top" align="left">Severity level (%)</th>
<th valign="top" align="left">#Augmented images</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">No Disease</td>
<td valign="top" align="left">Healthy</td>
<td valign="top" align="left">1252</td>
<td valign="top" align="left">0</td>
<td valign="top" align="left">1252</td>
</tr>
<tr>
<td valign="top" rowspan="3" align="left">Yellow Rust</td>
<td valign="top" align="left">YR_Low</td>
<td valign="top" align="left">216</td>
<td valign="top" align="left">0-25</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">YR_Medium</td>
<td valign="top" align="left">288</td>
<td valign="top" align="left">25-50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">YR_High</td>
<td valign="top" align="left">501</td>
<td valign="top" align="left">&gt;50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" rowspan="3" align="left">Brown Rust</td>
<td valign="top" align="left">BR_Low</td>
<td valign="top" align="left">395</td>
<td valign="top" align="left">0-25</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">BR_Medium</td>
<td valign="top" align="left">595</td>
<td valign="top" align="left">25-50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">BR_High</td>
<td valign="top" align="left">216</td>
<td valign="top" align="left">&gt;50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" rowspan="3" align="left">Stem Rust</td>
<td valign="top" align="left">SR_Low</td>
<td valign="top" align="left">780</td>
<td valign="top" align="left">0-25</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">SR_Medium</td>
<td valign="top" align="left">706</td>
<td valign="top" align="left">25-50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" align="left">SR_High</td>
<td valign="top" align="left">489</td>
<td valign="top" align="left">&gt;50</td>
<td valign="top" align="left">1000</td>
</tr>
<tr>
<td valign="top" colspan="2" align="left">Total no. of images</td>
<td valign="top" align="left">5438</td>
<td valign="top" align="left"/>
<td valign="top" align="left">10252</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Data pre-processing and augmentation</title>
<p>Prior to model training, image pre-processing, and augmentation were performed to enhance model performance. Initially, duplicate, out-of-focus, noisy, or blurry photos were eliminated from the dataset to ensure data quality. Subsequently, the Augmentor Python package was utilized to augment the images by employing various techniques such as zooming, flipping, and rotating. This augmentation process aimed to diversify the dataset and enrich it with variations, thereby facilitating robust model training. Each class of severity stage images was augmented to contain 1000 images, ensuring balanced representation across classes (refer to <xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>). Additionally, the images were resized to 256 x 256 pixels to accommodate hardware constraints, optimize computational efficiency, and enhance the model&#x2019;s generalization and performance. This pre-processing and augmentation pipeline laid the groundwork for effective model training on the augmented dataset.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Framework overview and structure</title>
<p>The research methodology is visually depicted in <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2</bold>
</xref>. Initially, images were captured from real-world wheat crop fields using mobile devices. Domain experts labeled each image with the corresponding type of wheat rust and its severity level, organizing them into distinct folders. Subsequently, image processing techniques, including resizing, filtering, and noise reduction, were applied to refine the raw images. Augmentation techniques, such as random rotation, translation, flipping, and zooming, were employed to diversify the image dataset and validate the models before experimentation. Two datasets were created: the original dataset containing 5438 images and an augmented dataset comprising 10252 images, both segregated into train, test, and validation sets in an 80:10:10 ratio for experimentation purposes. Initially, the performance of the fine-tuned EfficientNet B0 model was evaluated on both datasets. Subsequently, to enhance the model&#x2019;s performance, the proposed model was developed by integrating the CBAM module (<xref ref-type="bibr" rid="B40">Tan and Le, 2020</xref>) into the fine-tuned EfficientNet B0 model. The attention mechanism&#x2019;s channel and spatial modules focus on key disease symptoms, aiding in determining the severity level of wheat rust. <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2</bold>
</xref> illustrates the flowchart of the wheat disease identification and severity stage estimation framework, with subsequent sections elaborating on each phase of the framework.</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>Overview of automated severity stage estimation framework.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g002.tif"/>
</fig>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Architectural overview of EfficientNet and attention mechanism integration</title>
<p>The EfficientNet B0 serves as the foundational model within the EfficientNet family, encompassing a total of eight variants (B0-B7) (<xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>). EfficientNet B0 architecture achieves high accuracy and computational efficiency through a compound scaling approach as described by <xref ref-type="bibr" rid="B2">Atila et&#xa0;al. (2021)</xref>. The EfficientNet architecture employs the Mobile Inverted Bottleneck Convolution (MBConv) as its primary building block, introduced by <xref ref-type="bibr" rid="B35">Sandler et&#xa0;al. (2018)</xref> and illustrated in <xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>. The MBConv block comprises three components: a 1 &#xd7; 1 convolution (1 &#xd7; 1 Conv), Depth-wise convolution (Depth-wise Conv), and a Squeeze-and-Excitation (SE) module. Initially, the output of the preceding layer is passed through the MBConv block, where the number of channels is expanded using a 1 &#xd7; 1 Conv. Subsequently, a 3 &#xd7; 3 Depth-wise Conv reduces the number of parameters, followed by channel pruning that compresses the channel count through another 1 &#xd7; 1 Conv. A residual connection is then introduced between the input and output of the projection layer to enhance feature representation. The SE module, as shown in <xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>, incorporates two key operations: squeeze and excitation. The squeeze operation is performed using global average pooling (AvgPooling), while the excitation operation involves two fully connected layers activated sequentially with a Swish activation and a Sigmoid activation function. This design facilitates efficient parameter utilization while maintaining high performance. However, the SE module focuses on channel-wise feature recalibration by emphasizing informative channel characteristics while suppressing less relevant ones. However, this approach primarily addresses channel-specific information and overlooks spatial context, which is critical for visual recognition tasks such as severity estimation. This limitation negatively affected the model&#x2019;s classification accuracy for severity estimation.</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Baseline architecture of EfficientNet B0.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g003.tif"/>
</fig>
<p>To address this, the (CBAM) was integrated into EfficientNet-B0 in place of the SE module to enhance feature extraction by simultaneously considering both channel and spatial information. The modified network, referred to as EfficientNet-CBAM, is illustrated in <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4</bold>
</xref>. Key modifications include the replacement of the SE module in each MBConv layer with a CBAM module, allowing the network to retain vital spatial information alongside channel-specific features, particularly for identifying disease severity symptoms. Additionally, a CBAM module was introduced after the final convolutional layer, refining the extracted features and improving the network&#x2019;s classification performance. The final convolutional layer of EfficientNet B0 produces feature maps, which serve as input for the Convolutional Block Attention Mechanism (CBAM) module (see <xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4</bold>
</xref>).</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Improved EfficientNet B0 architecture: SE module replaced with CBAM module.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g004.tif"/>
</fig>
<p>Attention mechanisms, extensively utilized in research, augment feature extraction and boost model performance in image classification tasks (<xref ref-type="bibr" rid="B46">Woo et&#xa0;al., 2018</xref>; <xref ref-type="bibr" rid="B45">Wang et&#xa0;al., 2017</xref>). Our architectural design incorporates a convolutional block attention module with two key components: the Channel Attention Module (CAM) and the Spatial Attention Module (SAM) (refer to <xref ref-type="fig" rid="f5">
<bold>Figures&#xa0;5</bold>
</xref>). These two modules work together to improve feature extraction and representation within the generated feature maps (<xref ref-type="bibr" rid="B46">Woo et&#xa0;al., 2018</xref>). The input feature map W, representing the wheat rust-infected leaf image, undergoes processing within the CAM, producing the channel attention feature map Mc. This map highlights essential image information, which is then used to generate the refined feature map W&#x2019;. Element-wise multiplication between M<sub>c</sub> and W yields the improved feature map W&#x2019;. Subsequently, W&#x2019; is subjected to processing within the Spatial Attention Module, generating the spatial feature map M<sub>s</sub>. This map selectively emphasizes significant image areas. The enhanced feature map W&#x2019; is subsequently combined with the spatial feature map M<sub>s</sub>. This multiplication yields the ultimate feature map W&#x2019;&#x2019;, which encapsulates the representation of the wheat rust image. The Convolutional Block Attention Mechanism operates through the following <xref ref-type="disp-formula" rid="eq1">Equations 1</xref>, <xref ref-type="disp-formula" rid="eq2">2</xref>:</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Overall architecture of CBAM module.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g005.tif"/>
</fig>
<disp-formula id="eq1">
<label>(1)</label>
<mml:math display="block" id="M1">
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>'</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mtext>M</mml:mtext>
<mml:mtext>c</mml:mtext>
</mml:msub>
<mml:mo stretchy="false">(</mml:mo>
<mml:mtext>W</mml:mtext>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo>&#x2297;</mml:mo>
<mml:mtext>W</mml:mtext>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq2">
<label>(2)</label>
<mml:math display="block" id="M2">
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>&#x201d;</mml:mo>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mtext>M</mml:mtext>
<mml:mtext>s</mml:mtext>
</mml:msub>
<mml:mo stretchy="false">(</mml:mo>
<mml:mtext>W</mml:mtext>
<mml:mo>'</mml:mo>
<mml:mo stretchy="false">)</mml:mo>
<mml:mo>&#x2297;</mml:mo>
<mml:mtext>W</mml:mtext>
<mml:mo>'</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The CBAM module&#x2019;s channel attention mechanism utilizes pooling operations to compress the feature map W, focusing solely on essential symptom regions within the image while disregarding extraneous information or features. Conversely, the spatial attention mechanism identifies significant feature locations post-CAM processing. This process involves spatial dimension compression of the feature maps W&#x2019; and the generation of the spatial attention feature M<sub>s</sub> utilizing the sigmoid activation function. It highlights critical features within specific image area, enhancing intermediate features.</p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>Proposed severity estimation framework</title>
<p>The proposed methodology employs transfer learning, where a novel model aimed at disease severity stage identification is trained utilizing a pre-trained model, EfficientNet B0, as the foundation for learning. While retaining the initial layers of the EfficientNet B0 model, the final layer is replaced with new layers. These newly introduced layers are subsequently fine-tuned to classify infected leaves into ten distinct classes using, &#x2018;WheatSev&#x2019; dataset developed by us, as per the methodology given by <xref ref-type="bibr" rid="B42">Too et&#xa0;al. (2019)</xref>. Thus, the WheatSevNet model is designed to accurately classify the severity stages of wheat rust infections by enhancing feature extraction and representation. It builds upon EfficientNet-B0, a widely used deep learning architecture known for its computational efficiency and high performance. However, EfficientNet-B0&#x2019;s Squeeze-and-Excitation (SE) module, while effective for channel-wise recalibration, lacks spatial feature extraction capabilities. To overcome this limitation, WheatSevNet integrates the Convolutional Block Attention Module (CBAM) in place of the SE module within each MBConv block of EfficientNet-B0 as depicted in <xref ref-type="fig" rid="f6">
<bold>Figure&#xa0;6</bold>
</xref>. The CBAM module consists of two components: the Channel Attention Module (CAM), which selectively enhances significant channels using global average pooling, max pooling, a multi-layer perceptron (MLP), and a sigmoid activation function, and the Spatial Attention Module (SAM), which refines feature extraction by applying average and max pooling across the channel axis, followed by a convolutional layer and a sigmoid activation function. In WheatSevNet, each MBConv block of EfficientNet-B0 is modified by replacing the SE module with CBAM, ensuring both spatial and channel-wise attention are effectively incorporated. Additionally, a final CBAM layer is introduced after the last convolutional layer to further refine feature maps before classification. The model leverages transfer learning and fine-tuning, utilizing pre-trained EfficientNet-B0 weights with the initial layers frozen while optimizing the later layers for severity classification. Further modification involves the addition of a normalization layers, fully connected (FC) layer, a dropout, and a convolutional layer, as depicted in <xref ref-type="fig" rid="f6">
<bold>Figure&#xa0;6</bold>
</xref> for detecting and classifying different severity stages of wheat rust. By integrating CBAM at multiple levels, WheatSevNet achieves enhanced feature extraction, capturing both disease-specific spatial structures and critical channel-wise characteristics, thereby improving classification performance over traditional EfficientNet-based approaches.</p>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Overall proposed model framework for Wheat disease severity estimation.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g006.tif"/>
</fig>
<p>Additionally, <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref> presents the hyperparameters employed for the disease severity estimation models. Owing to low stopping, the number of epochs ranged from 27 to 50, with a fixed learning rate of 0.001. To mitigate this risk, the authors have incorporated several measures to prevent overfitting. These include data augmentation, which helps in artificially increasing the dataset size and providing more diverse training examples. The augmentation process involved applying random transformations such as rotation, flipping, scaling, and color adjustments to create new variations of the existing images. This technique aimed to simulate real-world variations in the data, which helps prevent the model from overfitting to the specifics of the training set. Furthermore, class balancing was ensured by augmenting each class of severity stage images to contain a total of 1000 images per class, ensuring that all classes were equally represented in the dataset. This balanced representation prevents the model from being biased towards any particular class, enhancing its ability to generalize across different categories. Further, a dropout rate of 0.20 was implemented during the training process. Dropout works by randomly disabling 20% of the neurons in each layer during training, which helps prevent the model from becoming overly reliant on specific features and encourages it to learn more generalized patterns. Additionally, L2 regularization has been applied to control the complexity of the model and prevent it from overfitting to the training data. In addition to these measures, early stopping with patience as 3 was incorporated as an extra safeguard against overfitting. This technique monitors the validation loss during training, and if no improvement is seen for a specified number of epochs (the patience parameter), the training process is halted early. This prevents the model from continuing to learn noise and overfitting to the training data. The patience parameter was set to allow the model to train for several epochs without improvement before stopping, ensuring that it had enough time to converge but also preventing unnecessary overfitting. All these hyperparameters for the severity estimation models are reported in <xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>. Categorical cross-entropy served as the loss function during model training, while the batch size for experimentation was fixed at 32. The subsequent subsection will address the third objective of the research study, focusing on elucidating the validation of the developed models and their integration into mobile applications.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Hyperparameters set for wheat disease severity model.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">Epochs</th>
<th valign="top" align="left">Batch size</th>
<th valign="top" align="left">Optimizer</th>
<th valign="top" align="left">Learning rate</th>
<th valign="top" align="left">Momentum</th>
<th valign="top" align="left">Loss function</th>
<th valign="top" align="left">Dropout</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">27-50</td>
<td valign="top" align="left">32</td>
<td valign="top" align="left">Adamax</td>
<td valign="top" align="left">0.001</td>
<td valign="top" align="left">0.99</td>
<td valign="top" align="left">Categorical cross-entropy</td>
<td valign="top" align="left">0.20</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_6">
<label>2.6</label>
<title>Evaluating model performance and efficacy</title>
<p>During the experiment, various pre-trained classical deep-learning models were compared to the proposed model. These models included VGGNet (<xref ref-type="bibr" rid="B38">Simonyan and Zisserman, 2014</xref>), ResNet152 (<xref ref-type="bibr" rid="B14">He et&#xa0;al., 2016</xref>), InceptionV3 (<xref ref-type="bibr" rid="B39">Szegedy et&#xa0;al., 2016</xref>), MobileNetV2 (<xref ref-type="bibr" rid="B35">Sandler et&#xa0;al., 2018</xref>), and DenseNet121 (<xref ref-type="bibr" rid="B18">Huang et&#xa0;al., 2017</xref>). These models underwent parameter resetting before training, followed by modifications to the bottom layers of the pre-trained networks. The bottom layer was substituted with a new SoftMax and output layers containing ten severity stage classes from the datasets.</p>
</sec>
<sec id="s2_7">
<label>2.7</label>
<title>Experimental implementation</title>
<p>The experimentation was conducted on a robust DGX server featuring GPU capabilities, with computations executed using the Keras and TensorFlow frameworks. The system has Ubuntu as the operating system, supported by an Intel<sup>&#xae;</sup> Xeon<sup>&#xae;</sup> CPU. All computationally intensive tasks were handled by the NVIDIA Tesla V100-SXM2 GPU, boasting ample memory resources of 528 GB (refer to <xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>).</p>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Experimental setup.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">Operating&#xa0;system</th>
<th valign="top" align="left">GPU</th>
<th valign="top" align="left">Memory</th>
<th valign="top" align="left">Frameworks</th>
<th valign="top" align="left">Programming language</th>
<th valign="top" align="left">System</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Ubuntu</td>
<td valign="top" align="left">NVIDIA&#xa0;Tesla&#xa0;V100-SXM2</td>
<td valign="top" align="left">528 GB</td>
<td valign="top" align="left">Keras&#xa0;and&#xa0;TensorFlow</td>
<td valign="top" align="left">Python</td>
<td valign="top" align="left">Intel<sup>&#xae;</sup> Xeon<sup>&#xae;</sup>
</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_8">
<label>2.8</label>
<title>Evaluation metrics</title>
<p>The accuracy of classification predictions in machine learning experiments is assessed through confusion matrices. It is used to analyze the correspondence between the predicted and actual prediction scores for individual classes of a classification model. Other metrics such as precision, accuracy, F1 score, and recall are also used to assess the performance of our model. Precision describes the ratio of true positives to all positive predictions, while accuracy refers to the proportion of correctly identified predictions relative to the total number of the predictions. On the other hand, Recall measures the ratio of true positive cases to all positive predictions (<xref ref-type="bibr" rid="B15">Hossin and Sulaiman, 2015</xref>). The F1 score is further calculated using the harmonic mean of precision and recall, providing a balanced evaluation of the model performance (<xref ref-type="disp-formula" rid="eq3">Equations 3</xref>&#x2013;<xref ref-type="disp-formula" rid="eq6">6</xref>).</p>
<disp-formula id="eq3">
<label>(3)</label>
<mml:math display="block" id="M3">
<mml:mrow>
<mml:mtext mathvariant="bold-italic">Precision</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FP</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq4">
<label>(4)</label>
<mml:math display="block" id="M4">
<mml:mrow>
<mml:mtext mathvariant="bold-italic">Accuracy</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">TN</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">TN</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FN</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq5">
<label>(5)</label>
<mml:math display="block" id="M5">
<mml:mrow>
<mml:mtext mathvariant="bold-italic">Recall</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FN</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq6">
<label>(6)</label>
<mml:math display="block" id="M6">
<mml:mrow>
<mml:mtext mathvariant="bold-italic">F</mml:mtext>
<mml:mn mathvariant="bold">1</mml:mn>
<mml:mo>=</mml:mo>
<mml:mo>&#xa0;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn mathvariant="bold">2</mml:mn>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn mathvariant="bold">2</mml:mn>
<mml:mtext mathvariant="bold-italic">TP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FP</mml:mtext>
<mml:mo>+</mml:mo>
<mml:mtext mathvariant="bold-italic">FN</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>The &#x201c;true positive&#x201d; (TP) represents the count of images accurately detected within each severity stage class. Conversely, &#x201c;true negative&#x201d; (TN) represents the overall number of images correctly identified across all severity stages, excluding the specific severity stage to which they belong. &#x201c;False negative&#x201d; refers to the count of images wrongly classified within each relevant severity class, while &#x201c;false positive&#x201d; (FP) indicates the number of images incorrectly classified as belonging to different severity stage classifications. Finally, the predictive performance of the proposed model is summarized by the F1 score.</p>
</sec>
</sec>
<sec id="s3" sec-type="results">
<label>3</label>
<title>Results and discussion</title>
<p>The experiment aimed to estimate the severity stages for all three wheat rusts utilizing the image dataset. The wheat disease severity estimation model was crafted using the EfficientNet architecture as the foundational model, augmented with the Convolutional Block Attention Mechanism (CBAM) integrated at the network&#x2019;s base. Performance evaluation of the disease severity estimation model was conducted, posing it against state-of-the-art CNN models and a simple fine-tuned EfficientNet B0 model, as outlined in <xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>. Results demonstrate that the proposed severity model achieved the highest test accuracy, reaching 93.88% and 96.68% on both the non-augmented and augmented datasets, respectively. Upon comparing the experimental results of EfficientNet B0 and EfficientNet B0-CBAM, as presented in <xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>, a notable enhancement in disease severity identification was observed upon the integration of an attention mechanism into the model.</p>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Performance comparison of State-of-the-art CNN models with proposed severity estimation model.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" colspan="3" align="center">Model parameters</th>
<th valign="top" colspan="3" align="center">Accuracies on non- augmented dataset (%)</th>
<th valign="top" colspan="3" align="center">Accuracies on augmented dataset (%)</th>
</tr>
<tr>
<th valign="top" align="left">Model</th>
<th valign="top" align="left">Image size</th>
<th valign="top" align="left">Ep</th>
<th valign="top" align="left">Training</th>
<th valign="top" align="left">Validation</th>
<th valign="top" align="left">Testing</th>
<th valign="top" align="left">Training</th>
<th valign="top" align="left">Validation</th>
<th valign="top" align="left">Testing</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">VGGNet19</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">85.14</td>
<td valign="top" align="left">84.78</td>
<td valign="top" align="left">83.27</td>
<td valign="top" align="left">86.73</td>
<td valign="top" align="left">85.09</td>
<td valign="top" align="left">85.12</td>
</tr>
<tr>
<td valign="top" align="left">ResNet152</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">87.72</td>
<td valign="top" align="left">86.03</td>
<td valign="top" align="left">86.91</td>
<td valign="top" align="left">88.14</td>
<td valign="top" align="left">87.65</td>
<td valign="top" align="left">87.08</td>
</tr>
<tr>
<td valign="top" align="left">MobileNetV2</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">86.90</td>
<td valign="top" align="left">85.87</td>
<td valign="top" align="left">84.78</td>
<td valign="top" align="left">87.05</td>
<td valign="top" align="left">86.31</td>
<td valign="top" align="left">84.93</td>
</tr>
<tr>
<td valign="top" align="left">DenseNet169</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">90.13</td>
<td valign="top" align="left">90.15</td>
<td valign="top" align="left">89.97</td>
<td valign="top" align="left">91.44</td>
<td valign="top" align="left">90.7</td>
<td valign="top" align="left">90.56</td>
</tr>
<tr>
<td valign="top" align="left">InceptionV3</td>
<td valign="top" align="left">299*299</td>
<td valign="top" align="left">50</td>
<td valign="top" align="left">91.82</td>
<td valign="top" align="left">90.50</td>
<td valign="top" align="left">90.22</td>
<td valign="top" align="left">94.62</td>
<td valign="top" align="left">92.05</td>
<td valign="top" align="left">91.84</td>
</tr>
<tr>
<td valign="top" align="left">EfficientNet B0</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">25</td>
<td valign="top" align="left">96.33</td>
<td valign="top" align="left">92.68</td>
<td valign="top" align="left">91.56</td>
<td valign="top" align="left">98.54</td>
<td valign="top" align="left">93.43</td>
<td valign="top" align="left">92.19</td>
</tr>
<tr>
<td valign="top" align="left">Proposed model</td>
<td valign="top" align="left">224*224</td>
<td valign="top" align="left">27</td>
<td valign="top" align="left">98.67</td>
<td valign="top" align="left">94.83</td>
<td valign="top" align="left">93.88</td>
<td valign="top" align="left">99.51</td>
<td valign="top" align="left">95.97</td>
<td valign="top" align="left">96.68</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In the absence of an attention mechanism in EfficientNetB0, the overall testing accuracy on the WheatSev dataset was recorded at 92.19%. However, upon integrating the CBAM module into EfficientNetB0, the overall testing accuracy markedly increased to 96.68%, as illustrated in <xref ref-type="fig" rid="f7">
<bold>Figure&#xa0;7</bold>
</xref>. Analysis of the disease severity stage classification results revealed that the performance enhancement observed in EfficientNetB0, when augmented with the attention mechanism, could be attributed to the spatial attention module of the CBAM module, which adeptly locates key information with greater accuracy. Furthermore, the channel attention module of CBAM exhibits the ability to amplify important features while suppressing irrelevant ones, thereby yielding a more refined feature representation. Consequently, it can be inferred that the incorporation of the CBAM module into the base model effectively contributes to improving the model&#x2019;s performance in identifying the severity level of wheat rust diseases.</p>
<fig id="f7" position="float">
<label>Figure&#xa0;7</label>
<caption>
<p>Comparison of proposed model testing accuracies with various CNN models.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g007.tif"/>
</fig>
<sec id="s3_1">
<label>3.1</label>
<title>Confusion matrix and various performance metrics</title>
<p>
<xref ref-type="fig" rid="f8">
<bold>Figure&#xa0;8</bold>
</xref> illustrates the confusion matrix of the severity estimation model, depicting ten classes based on actual and predicted labels from augmented datasets. The numbers along the diagonal signify accurately identified images, while those outside the diagonal indicate instances of misclassification (<xref ref-type="bibr" rid="B41">Ting, 2017</xref>). Specifically, among the four low severity images of brown Rust, two were erroneously identified as medium-stage severity, and one as healthy (<xref ref-type="fig" rid="f8">
<bold>Figure&#xa0;8</bold>
</xref>). Similarly, misclassifications occur in other severity stages of brown Rust, with images of high and medium levels misclassified as medium and low, respectively. Furthermore, low-severity images of brown Rust were mistakenly classified as medium severity and healthy. Additionally, misclassifications were observed in stem rust and yellow rust severity stages.</p>
<fig id="f8" position="float">
<label>Figure&#xa0;8</label>
<caption>
<p>Confusion matrix for proposed disease severity estimation model.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g008.tif"/>
</fig>
<p>Upon analyzing the confusion matrix of EfficientNetB0 embedded with CBAM, it was noted that Rust spread on the upper surface of the leaf leads to significant confusion between the low and medium severity stages of the disease. Referencing <xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref>, the classification report derived from the confusion matrices includes F1, precision, and recall metrics for the proposed disease severity estimation model. A model is deemed appropriate if its F1 score approaches one. After evaluating these performance metrics, the following findings emerge: On the augmented dataset, the average precision, recall, and F1 score for identifying severity stages in brown rust and stem rust is 97%, whereas, for yellow Rust, the average score for precision, recall, and F1 measure is 95%.</p>
<table-wrap id="T5" position="float">
<label>Table&#xa0;5</label>
<caption>
<p>Performance metrics for severity estimation model on both datasets.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="left">Classes</th>
<th valign="middle" colspan="3" align="center">Augmented dataset</th>
<th valign="middle" colspan="3" align="center">Non-augmented dataset</th>
</tr>
<tr>
<th valign="middle" align="left">Precision</th>
<th valign="middle" align="left">Recall</th>
<th valign="middle" align="left">F1</th>
<th valign="middle" align="left">Precision</th>
<th valign="middle" align="left">Recall</th>
<th valign="middle" align="left">F1</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">Healthy</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fed280">0.96</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
</tr>
<tr>
<td valign="middle" align="left">BR_Low</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#fed280">0.96</td>
<td valign="middle" align="right" style="background-color:#fed280">0.96</td>
<td valign="middle" align="right" style="background-color:#d8df81">0.9</td>
<td valign="middle" align="right" style="background-color:#a9d27f">0.84</td>
<td valign="middle" align="right" style="background-color:#c0d980">0.87</td>
</tr>
<tr>
<td valign="middle" align="left">BR_Medium</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#8ac97d">0.8</td>
<td valign="middle" align="right" style="background-color:#63be7b">0.75</td>
<td valign="middle" align="right" style="background-color:#99cd7e">0.82</td>
</tr>
<tr>
<td valign="middle" align="left">BR_High</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#c8db80">0.88</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#dfe282">0.91</td>
</tr>
<tr>
<td valign="middle" align="left">SR_Low</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#f7e883">0.94</td>
<td valign="middle" align="right" style="background-color:#f7e883">0.94</td>
</tr>
<tr>
<td valign="middle" align="left">SR_Medium</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#fed280">0.96</td>
<td valign="middle" align="right" style="background-color:#d8df81">0.9</td>
<td valign="middle" align="right" style="background-color:#dfe282">0.91</td>
<td valign="middle" align="right" style="background-color:#dfe282">0.91</td>
</tr>
<tr>
<td valign="middle" align="left">SR_High</td>
<td valign="middle" align="right" style="background-color:#fa8370">0.99</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#efe683">0.93</td>
<td valign="middle" align="right" style="background-color:#f7e883">0.94</td>
<td valign="middle" align="right" style="background-color:#efe683">0.93</td>
</tr>
<tr>
<td valign="middle" align="left">YR_Low</td>
<td valign="middle" align="right" style="background-color:#dfe282">0.91</td>
<td valign="middle" align="right" style="background-color:#fed280">0.96</td>
<td valign="middle" align="right" style="background-color:#f7e883">0.94</td>
<td valign="middle" align="right" style="background-color:#8ac97d">0.8</td>
<td valign="middle" align="right" style="background-color:#91cb7d">0.81</td>
<td valign="middle" align="right" style="background-color:#8ac97d">0.8</td>
</tr>
<tr>
<td valign="middle" align="left">YR_Medium</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#dfe282">0.91</td>
<td valign="middle" align="right" style="background-color:#efe683">0.93</td>
<td valign="middle" align="right" style="background-color:#7ac47c">0.78</td>
<td valign="middle" align="right" style="background-color:#63be7b">0.75</td>
<td valign="middle" align="right" style="background-color:#72c27b">0.77</td>
</tr>
<tr>
<td valign="middle" align="left">YR_High</td>
<td valign="middle" align="right" style="background-color:#f8696b">1</td>
<td valign="middle" align="right" style="background-color:#fdb87b">0.97</td>
<td valign="middle" align="right" style="background-color:#fb9d75">0.98</td>
<td valign="middle" align="right" style="background-color:#ffeb84">0.95</td>
<td valign="middle" align="right" style="background-color:#e7e482">0.92</td>
<td valign="middle" align="right" style="background-color:#efe683">0.93</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Shades of red color for higher, yellow for moderate, and green for lower values of performance parameters.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>
<xref ref-type="table" rid="T5">
<bold>Table&#xa0;5</bold>
</xref> shows that augmented datasets yield superior results compared to non-augmented datasets across all diseases and their categories. However, for stem rust disease, all performance parameters in non-augmented datasets surpass 90% across all stages, possibly due to easily identifiable features of stem rusts. Conversely, for brown and yellow Rust, the performance of non-augmented datasets is notably inferior to augmented datasets. Another noteworthy finding pertains to the higher classification accuracy of healthy leaves, which can be attributed to (i) the more significant number of images (1252) and (ii) the absence of any classes for healthy leaves. Although the accuracy improves as the disease stage matures, even in low stages, precision exceeds 90% for all types of rusts. For the augmented dataset, precision reaches 97% for brown Rust and 91% for yellow Rust.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Model accuracy and loss curves</title>
<p>
<xref ref-type="fig" rid="f9">
<bold>Figure&#xa0;9</bold>
</xref> depicts the training and validation curves for the wheat severity estimation model, offering insights into the learning process. Notably, the accuracy curves indicate that our proposed severity model serves as a commendable and efficient fit model (<xref ref-type="bibr" rid="B15">Hossin and Sulaiman, 2015</xref>). The findings of the disease severity model underscore its efficacy in automatically identifying wheat severity stages based on images.</p>
<fig id="f9" position="float">
<label>Figure&#xa0;9</label>
<caption>
<p>Training and validation curves for severity estimation model on <bold>(a)</bold> non-augmented dataset <bold>(b)</bold> augmented dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g009.tif"/>
</fig>
<p>In conclusion, the proposed models prove useful for identifying diseases at a low severity level, as evidenced by the accurate identification of most images in the low stages of Rust. This early severity assessment holds promise for crop preservation and can significantly minimize crop loss.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Visualization of disease symptoms model interpretability</title>
<p>The interpretability of the proposed model is carried out using the GradCAM (<xref ref-type="bibr" rid="B36">Selvaraju et&#xa0;al., 2020</xref>). <xref ref-type="fig" rid="f10">
<bold>Figure&#xa0;10</bold>
</xref> illustrates that the proposed severity estimation model focuses explicitly on the features and the symptoms that play an important role in identifying the severity level of the type of the wheat rusts. The activation maps shown in the figure facilitate specific regions in the&#xa0;input test images necessary for estimating the severity of the disease.&#xa0;Thus, we aimed to illustrate how the model is directing its attention towards the areas where the symptoms are most noticeable in order to identify the disease at a low stage of severity. The attention mechanism has been found to improve the model&#x2019;s ability to identify the appropriate symptoms in the correct location accurately.</p>
<fig id="f10" position="float">
<label>Figure&#xa0;10</label>
<caption>
<p>GradCAM visualization of the <bold>(a)</bold> Yellow rust <bold>(b)</bold> Brown rust <bold>(c)</bold> Stem rust diseases at low severity level.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g010.tif"/>
</fig>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Development of an Android based mobile application for wheat disease severity estimation</title>
<p>In this study, we also developed an Android-based mobile application as a practical tool for the automatic identification of wheat diseases and their corresponding severity stages in agricultural fields. The proposed model was integrated into the application&#x2019;s backend to facilitate this functionality. The mobile application allows users to capture real-time images from agricultural fields or select images stored in their mobile device gallery. These images are subsequently uploaded to a server for analysis. The analytical process begins with the application determining whether the uploaded image depicts a healthy or diseased wheat plant. For images identified as healthy, the result is directly displayed as &#x201c;healthy&#x201d;. Conversely, if the image is diagnosed as diseased, the application proceeds to identify the specific type of disease. Following disease identification, the application further evaluates the image to estimate the severity stage of the detected disease. <xref ref-type="fig" rid="f11">
<bold>Figure&#xa0;11</bold>
</xref> provides a detailed illustration of the application&#x2019;s process flow, from image acquisition to disease identification and severity stage estimation.</p>
<fig id="f11" position="float">
<label>Figure&#xa0;11</label>
<caption>
<p>Overall flow of disease severity estimation through mobile application.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g011.tif"/>
</fig>
<p>The developed mobile application for wheat disease severity estimation follows a streamlined workflow <xref ref-type="fig" rid="f12">
<bold>Figure&#xa0;12</bold>
</xref>. Upon launching, a splash screen introduces the app, followed by an interface that allows users to capture or upload a wheat leaf image. Once an image is uploaded, the &#x201c;Identify&#x201d; button determines if the leaf is healthy or diseased. For healthy images, the app displays a message indicating no further action is required. If a disease is detected, the application identifies the disease type and provides an option to predict its severity stage. The final screen presents the identified disease along with its severity stage, providing a complete diagnostic result for the uploaded image.</p>
<fig id="f12" position="float">
<label>Figure&#xa0;12</label>
<caption>
<p>Screenshots of the developed mobile application for severity estimation.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1540642-g012.tif"/>
</fig>
</sec>
<sec id="s3_5">
<label>3.5</label>
<title>Comparison with existing studies in the literature</title>
<p>The <xref ref-type="table" rid="T6">
<bold>Table&#xa0;6</bold>
</xref> offers a comprehensive overview and comparative assessment of various models utilized for disease classification across different crops, alongside their respective training or testing accuracies. Our proposed model, specifically for diagnosing three major rusts in wheat crops, achieved a commendable testing accuracy of 96.68%. When compared with existing literature, our model emerges as a strong contender, demonstrating competitive performance. Notably, existing models developed for crops such as apple, tomato, coffee, cucumber, and grape achieved accuracies ranging from 90.4% to 97.75%, albeit focusing on single disease classes. Importantly, prior attempts at wheat severity estimation encompassing three diseases and their severity levels were scarce. Despite this, our model&#x2019;s accuracy not only matches but also exceeds the reported accuracies in the literature, underscoring its efficacy in wheat rust severity estimation.</p>
<table-wrap id="T6" position="float">
<label>Table&#xa0;6</label>
<caption>
<p>Summarization of the previous studies for image-based disease severity estimation.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">References</th>
<th valign="top" align="left">Crops/diseases/classes</th>
<th valign="top" align="left">No. of diseases</th>
<th valign="top" align="left">Algorithm/ Architecture</th>
<th valign="top" align="left">Training/Testing accuracy (%)</th>
<th valign="top" align="left">Dataset with number of images</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B45">Wang et&#xa0;al. (2017)</xref>
</td>
<td valign="top" align="left">Apple (Leaf black rot)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">VGG16 model</td>
<td valign="top" align="left">90.4</td>
<td valign="top" align="left">PlantVillage (1986)</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B24">Liang et&#xa0;al. (2019)</xref>
</td>
<td valign="top" align="left">Multiple fruit crops<break/>(healthy, general, serious)</td>
<td valign="top" align="left">&#x2013;</td>
<td valign="top" align="left">ResNet &amp; ShuffleNet</td>
<td valign="top" align="left">91</td>
<td valign="top" align="left">PlantVillage</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B34">Prabhakar et&#xa0;al. (2020)</xref>
</td>
<td valign="top" align="left">Tomato (early blight)<break/>(mild, moderate, severe)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">ResNet101</td>
<td valign="top" align="left">94.6</td>
<td valign="top" align="left">PlantVillage</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B43">Verma et&#xa0;al. (2020)</xref>
</td>
<td valign="top" align="left">Tomato (Late blight)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">AlexNet</td>
<td valign="top" align="left">93.4</td>
<td valign="top" align="left">PlantVillage</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B10">Esgario et&#xa0;al. (2020)</xref>
</td>
<td valign="top" align="left">Coffee (Leaf biotic stress)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">ResNet50</td>
<td valign="top" align="left">95.24</td>
<td valign="top" align="left">Own (2293)</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B44">Wang et&#xa0;al. (2021)</xref>
</td>
<td valign="top" align="left">Cucumber (downy and powdery mildew)</td>
<td valign="top" align="left">2</td>
<td valign="top" align="left">DeepLabV3</td>
<td valign="top" align="left">92.85</td>
<td valign="top" align="left">Own (1000)</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B48">Zhao et&#xa0;al. (2021)</xref>
</td>
<td valign="top" align="left">Tomato (fungal diseases)</td>
<td valign="top" align="left">&#x2013;</td>
<td valign="top" align="left">SENet &amp; CBAM</td>
<td valign="top" align="left">95.37</td>
<td valign="top" align="left">PlantVillage</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B26">Liu et&#xa0;al. (2022)</xref>
</td>
<td valign="top" align="left">Apple (Alternaria leaf blotch)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">DeepLabV3, UNet</td>
<td valign="top" align="left">96.41</td>
<td valign="top" align="left">Own (5382)</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B21">Ji and Wu et&#xa0;al. (2022)</xref>
</td>
<td valign="top" align="left">Grape (black measles)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">ResNet &amp; Fuzzy logic</td>
<td valign="top" align="left">97.75</td>
<td valign="top" align="left">PlantVillage</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B23">Li et&#xa0;al. (2023)</xref>
</td>
<td valign="top" align="left">Wheat (yellow Rust)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">GhostNetV2</td>
<td valign="top" align="left">95.44</td>
<td valign="top" align="left">Public</td>
</tr>
<tr>
<td valign="top" align="left">
<xref ref-type="bibr" rid="B16">Hu et&#xa0;al. (2023)</xref>
</td>
<td valign="top" align="left">Tea (leaf blight)</td>
<td valign="top" align="left">1</td>
<td valign="top" align="left">GBM</td>
<td valign="top" align="left">&#x2013;</td>
<td valign="top" align="left">Own (300)</td>
</tr>
<tr>
<td valign="top" align="left">WheatSevNet (Proposed)</td>
<td valign="top" align="left">Wheat (three major rusts)</td>
<td valign="top" align="left">3</td>
<td valign="top" align="left">EfficientNet &amp; CBAM</td>
<td valign="top" align="left">96.68</td>
<td valign="top" align="left">Own (10000)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>In this study, our primary contributions are twofold: Firstly, we curated a robust dataset for wheat disease classification, encompassing the estimation of severity categories. The non-augmented dataset comprises 5438 images, while the augmented dataset boasts a total of 10252 images. This dataset lays a strong foundation for future research in this domain. Secondly, we introduced WheatSevNet, an algorithm capable of identifying various wheat disease categories and assessing their severity levels. Despite the challenges posed by multi-disease classes and multi-severity levels, our enhanced model achieved an impressive accuracy rate of over 96%. This performance is comparable even to other algorithms designed for single disease identification. We could not compare with wheat severity estimation models for three diseases as no such published attempt is available to the best of our knowledge. The success of our approach not only addresses an immediate need in agricultural research and opens up promising avenues for future investigations in this field. We hope our study will inspire further exploration and innovation in automated plant disease diagnosis and severity estimation.</p>
</sec>
</sec>
<sec id="s4" sec-type="conclusions">
<label>4</label>
<title>Conclusion</title>
<p>The major fungal diseases in a wheat crop significantly impact crop quality and quantity, leading to substantial agricultural yield losses. In our study, we aim to diagnose major wheat fungal diseases and its corresponding severity level, utilizing a model based on EfficientNet architecture and enhanced with a Convolutional Block Attention Mechanism. The proposed model demonstrates exceptional effectiveness, boasting a training accuracy of 99.51% and a testing accuracy of 96.68%. In comparative analyses, our model surpasses state-of-the-art CNN models and a fine-tuned EfficientNet B0 model, highlighting its superior performance in severity estimation. To ensure the robustness of our approach across various disease categories, we conducted experiments using images from real-life field conditions, encompassing three major types of wheat rusts: yellow, brown, and black. Notably, our model&#x2019;s ability to classify severity stages into medium and high stages provides precise information, facilitating timely intervention. The integration of the CBAM module significantly enhances the model&#x2019;s performance, boosting the testing accuracy from 93.21% to an impressive 96.68% on the WheatSev dataset. This improvement is largely attributed to the attention module within CBAM, which adeptly identifies critical information and enhances the representation of features. Furthermore, the channel attention module demonstrates its effectiveness in amplifying features while suppressing ones, thereby contributing to a more precise and accurate identification of the severity level of wheat rust disease. The results validate that the inclusion of the CBAM module substantially improves the efficiency of the model in detecting and assessing the severity of wheat rust disease.</p>
</sec>
</body>
<back>
<sec id="s5" sec-type="data-availability">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec id="s6" sec-type="author-contributions">
<title>Author contributions</title>
<p>SN: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing &#x2013; original draft, Writing &#x2013; review &amp; editing. RJ: Conceptualization, Formal analysis, Investigation, Methodology, Supervision, Validation, Visualization, Writing &#x2013; review &amp; editing. VS: Data curation, Funding acquisition, Investigation, Resources, Supervision, Validation, Writing &#x2013; review &amp; editing. AS: Conceptualization, Formal analysis, Methodology, Project administration, Software, Visualization, Writing &#x2013; original draft. HK: Investigation, Writing &#x2013; review &amp; editing.</p>
</sec>
<sec id="s7" sec-type="funding-information">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. This work was supported by Bill and Melinda Gates Foundation under the project ICAR BMGF (Grant Number OPP1194767).</p>
</sec>
<sec id="s8" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s9" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec id="s10" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arnal Barbedo</surname> <given-names>J. G.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Plant disease identification from individual lesions and spots using deep learning</article-title>. <source>Biosyst. Eng.</source> <volume>180</volume>, <fpage>96</fpage>&#x2013;<lpage>107</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.biosystemseng.2019.02.002</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Atila</surname> <given-names>&#xdc;.</given-names>
</name>
<name>
<surname>U&#xe7;ar</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Akyol</surname> <given-names>K.</given-names>
</name>
<name>
<surname>U&#xe7;ar</surname> <given-names>E.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Plant leaf disease classification using EfficientNet deep learning model</article-title>. <source>Ecol. Inf.</source> <volume>61</volume>, <fpage>101182</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ecoinf.2020.101182</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bock</surname> <given-names>C. H.</given-names>
</name>
<name>
<surname>Pethybridge</surname> <given-names>S. J.</given-names>
</name>
<name>
<surname>Barbedo</surname> <given-names>J. G. A.</given-names>
</name>
<name>
<surname>Esker</surname> <given-names>P. D.</given-names>
</name>
<name>
<surname>Mahlein</surname> <given-names>A. K.</given-names>
</name>
<name>
<surname>Del Ponte</surname> <given-names>E. M.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>A phytopathometry glossary for the twenty-first century: Towards consistency and precision in intra- and inter-disciplinary dialogues</article-title>. <source>Trop. Plant Pathol.</source> <volume>47</volume>, <fpage>14</fpage>&#x2013;<lpage>24</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s40858-021-00454-0</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bock</surname> <given-names>C. H.</given-names>
</name>
<name>
<surname>Poole</surname> <given-names>G. H.</given-names>
</name>
<name>
<surname>Parker</surname> <given-names>P. E.</given-names>
</name>
<name>
<surname>Gottwald</surname> <given-names>T. R.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Plant disease severity estimated visually, by digital photography and image analysis, and by hyperspectral imaging</article-title>. <source>Crit. Rev. Plant Sci.</source> <volume>29</volume>, <fpage>59</fpage>&#x2013;<lpage>107</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1080/07352681003617285</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Ban</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>Y.</given-names>
</name>
<etal/>
</person-group>. (<year>2021</year>). <article-title>An approach for rice bacterial leaf streak disease segmentation and disease severity estimation</article-title>. <source>Agriculture</source> <volume>11</volume>, <fpage>420</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agriculture11050420</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Wellings</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Kang</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Wheat stripe (yellow)&#xa0;rust caused by <italic>P uccinia striiformis f. sp. tritici</italic>
</article-title>. <source>Molecular plant pathology</source> <volume>15</volume> (<issue>5</issue>), <page-range>433&#x2013;446</page-range>.</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chin</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Catal</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Kassahun</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Plant disease detection using drones in precision agriculture</article-title>. <source>Precis. Agric.</source> <volume>24</volume>, <fpage>1663</fpage>&#x2013;<lpage>1682</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-023-10014-y</pub-id>
</citation>
</ref>
<ref id="B8">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dheeraj</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Chand</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2022</year>). &#x201c;<article-title>Deep learning model for automated image based plant disease classification</article-title>,&#x201d; in <source>International Conference on Intelligent Vision and Computing</source> (<publisher-name>Springer Nature Switzerland</publisher-name>, <publisher-loc>Cham</publisher-loc>), <fpage>21</fpage>&#x2013;<lpage>32</lpage>.</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dheeraj</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Chand</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>LWDN: lightweight DenseNet model for plant&#xa0;disease diagnosis</article-title>. <source>J. Plant Dis. Prot.</source> <volume>131</volume> (<issue>3</issue>), <page-range>1043&#x2013;59</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s41348-024-00915-z</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Esgario</surname> <given-names>J. G.</given-names>
</name>
<name>
<surname>Krohling</surname> <given-names>R. A.</given-names>
</name>
<name>
<surname>Ventura</surname> <given-names>J. A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Deep learning for classification and severity estimation of coffee leaf biotic stress</article-title>. <source>Comput. Electron. Agric.</source> <volume>169</volume>, <fpage>105162</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2019.105162</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fuentes</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Yoon</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>M. H.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>D. S.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Improving accuracy of tomato plant disease diagnosis based on deep learning with explicit control of hidden classes</article-title>. <source>Front. Plant Sci.</source> <volume>12</volume>, <elocation-id>682230</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2021.682230</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Haque</surname> <given-names>M. A.</given-names>
</name>
<name>
<surname>Marwaha</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Arora</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Deb</surname> <given-names>C. K.</given-names>
</name>
<name>
<surname>Misra</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Nigam</surname> <given-names>S.</given-names>
</name>
<etal/>
</person-group>. (<year>2022</year>). <article-title>A lightweight convolutional neural network for recognition of severity stages of maydis leaf blight disease of maize</article-title>. <source>Front. Plant Sci.</source> <volume>13</volume>, <elocation-id>1077568</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2022.1077568</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hayit</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Erbay</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Var&#xe7;&#x131;n</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Hayit</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Akci</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Determination of the severity level of yellow rust disease in wheat by using convolutional neural networks</article-title>. <source>J.&#xa0;Plant Pathol.</source> <volume>103</volume>, <fpage>923</fpage>&#x2013;<lpage>934</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s42161-021-00886-2</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Identity mappings in deep residual networks</article-title>,&#x201d; in <source>European conference on computer vision</source> (<publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>), <fpage>630</fpage>&#x2013;<lpage>645</lpage>.</citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hossin</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Sulaiman</surname> <given-names>M. N.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>A review on evaluation metrics for data classification evaluations</article-title>. <source>International journal of data mining &amp; knowledge management process</source> <volume>5</volume> (<issue>2</issue>), <fpage>1</fpage>.</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Wan</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Wei</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Ye</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Computer vision based method for severity estimation of tea leaf blight in natural scene images</article-title>. <source>Eur. J. Agron.</source> <volume>144</volume>, <fpage>126756</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.eja.2023.126756</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Wei</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Bao</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Liang</surname> <given-names>D.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Estimation of tea leaf blight severity in natural scene images</article-title>. <source>Precis. Agric.</source> <volume>22</volume>, <fpage>1239</fpage>&#x2013;<lpage>1262</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-020-09782-8</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Huang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>van der Maaten</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Weinberger</surname> <given-names>K. Q.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Densely connected convolutional networks</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <fpage>4700</fpage>&#x2013;<lpage>4708</lpage>.</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jiang</surname> <given-names>Q.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>H. L.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>H. G.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Two new methods for severity assessment of wheat stripe rust caused by <italic>Puccinia striiformis</italic> f. sp. <italic>tritici</italic>
</article-title>. <source>Front. Plant Sci.</source> <volume>13</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2022.1002627</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jiang</surname> <given-names>Q.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Severity assessment of wheat stripe rust based on machine learning</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>, <elocation-id>1150855</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1150855</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ji</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Z.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Automatic detection and severity analysis of grape black measles disease based on deep learning and fuzzy logic</article-title>. <source>Computers and Electronics in Agriculture.</source> <volume>193</volume>, <elocation-id>106718</elocation-id>.</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>LeCun</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Bengio</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Hinton</surname> <given-names>G.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Deep learning</article-title>. <source>Nature</source> <volume>521</volume>, <fpage>436</fpage>&#x2013;<lpage>444</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/nature14539</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Fang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Zhen</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Zhu</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Detection of wheat yellow rust disease severity based on improved GhostNetV2</article-title>. <source>Appl. Sci.</source> <volume>13</volume> (<issue>17</issue>), <fpage>9987</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/app13179987</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liang</surname> <given-names>Q.</given-names>
</name>
<name>
<surname>Xiang</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Hu</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Coppola</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>W.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>PD2SE-Net: Computer-assisted plant disease diagnosis and severity estimation network</article-title>. <source>Comput. Electron. Agric.</source> <volume>157</volume>, <fpage>518</fpage>&#x2013;<lpage>529</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2019.01.034</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Ding</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Tian</surname> <given-names>L.</given-names>
</name>
<name>
<surname>He</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Grape leaf disease identification using improved deep convolutional neural networks</article-title>. <source>Front. Plant Sci.</source> <volume>11</volume>, <elocation-id>1082</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2020.01082</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>B. Y</given-names>
</name>
<name>
<surname>Fan</surname> <given-names>K. J.</given-names>
</name>
<name>
<surname>Su</surname> <given-names>W. H.</given-names>
</name>
<name>
<surname>Peng</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Two-stage convolutional neural networks for diagnosing the severity of alternaria leaf blotch disease of the apple tree</article-title>. <source>Remote Sensing.</source> <volume>14</volume> (<issue>11</issue>), <elocation-id>2519</elocation-id>.</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Hu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Mei</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>C.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>An in-field automatic wheat disease diagnosis system</article-title>. <source>Comput. Electron. Agric.</source> <volume>142</volume>, <fpage>369</fpage>&#x2013;<lpage>379</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2017.09.012</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mi</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Su</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Han</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Su</surname> <given-names>B.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Wheat stripe rust grading by deep learning with attention mechanism and images from mobile devices</article-title>. <source>Front. Plant Sci.</source> <volume>11</volume>, <elocation-id>558126</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2020.558126</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mohanty</surname> <given-names>S. P.</given-names>
</name>
<name>
<surname>Hughes</surname> <given-names>D. P.</given-names>
</name>
<name>
<surname>Salath&#xe9;</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Using deep learning for image-based plant disease detection</article-title>. <source>Front. Plant Sci.</source> <volume>7</volume>, <elocation-id>1419</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2016.01419</pub-id>
</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nigam</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jain</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Plant disease identification using Deep Learning: A review</article-title>. <source>Indian J. Agric. Sci.</source> <volume>90</volume>, <fpage>249</fpage>&#x2013;<lpage>257</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.56093/ijas.v90i2.98996</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nigam</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jain</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Marwaha</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Arora</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Haque</surname> <given-names>M. A.</given-names>
</name>
<name>
<surname>Dheeraj</surname> <given-names>A.</given-names>
</name>
<etal/>
</person-group>. (<year>2023</year>). <article-title>Deep transfer learning model for disease identification in wheat crop</article-title>. <source>Ecol. Inf.</source> <volume>75</volume>, <fpage>102068</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ecoinf.2023.102068</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Nigam</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Jain</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Prakash</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Marwaha</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Arora</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Singh</surname> <given-names>V. K.</given-names>
</name>
<etal/>
</person-group>. (<year>2021</year>). &#x201c;<article-title>Wheat disease severity estimation: A deep learning approach</article-title>,&#x201d; in <source>International Conference on Internet of Things and Connected Technologies</source> (<publisher-name>Springer</publisher-name>, <publisher-loc>Cham</publisher-loc>), <fpage>185</fpage>&#x2013;<lpage>193</lpage>.</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pavithra</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Kalpana</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Vigneswaran</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Deep learning-based automated disease detection and classification model for precision agriculture</article-title>. <source>Soft Computing</source>. <volume>28</volume> (<supplement>Suppl 2</supplement>), <fpage>463</fpage>&#x2013;<lpage>483</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s00500-023-07936-0</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Prabhakar</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Purushothaman</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Awasthi</surname> <given-names>D. P.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Deep learning based assessment of disease severity for early blight in tomato crop</article-title>. <source>Multimedia tools and applications</source>, <volume>79</volume>, <fpage>28773</fpage>&#x2013;<lpage>84</lpage>.</citation>
</ref>
<ref id="B35">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Sandler</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Howard</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Zhu</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Zhmoginov</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>L. C.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Mobilenetv2: Inverted residuals and linear bottlenecks</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <fpage>4510</fpage>&#x2013;<lpage>4520</lpage>.</citation>
</ref>
<ref id="B36">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Selvaraju</surname> <given-names>R. R.</given-names>
</name>
<name>
<surname>Cogswell</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Das</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Vedantam</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Parikh</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Batra</surname> <given-names>D.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Grad-cam: Visual explanations from deep networks via gradient-based localization</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE international conference on computer vision</conf-name>. <volume>128</volume>, <fpage>336</fpage>&#x2013;<lpage>359</lpage>.</citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sibiya</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Sumbwanyambe</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Automatic fuzzy logic-based maize common rust disease severity predictions with thresholding and deep learning</article-title>. <source>Pathogens</source> <volume>10</volume>, <fpage>131</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/pathogens10020131</pub-id>
</citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Zisserman</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2014</year>). <source>Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556</source>.</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Vanhoucke</surname> <given-names>V.</given-names>
</name>
<name>
<surname>Ioffe</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Shlens</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Wojna</surname> <given-names>Z.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Rethinking the inception architecture for computer vision</article-title>. <source>Proc. IEEE Conf. Comput. Vision Pattern Recognition</source>. <volume>2016</volume>, <fpage>2818</fpage>&#x2013;<lpage>2826</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2016.308</pub-id>
</citation>
</ref>
<ref id="B40">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Tan</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Le</surname> <given-names>Q.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>EfficientNet: Rethinking model scaling for convolutional neural networks</article-title>,&#x201d; in <source>International conference on machine learning</source> (<publisher-name>PMLR</publisher-name>), <fpage>6105</fpage>&#x2013;<lpage>6114</lpage>. Available online at: <uri xlink:href="https://proceedings.mlr.press/v97/tan19a.html">https://proceedings.mlr.press/v97/tan19a.html</uri>
</citation>
</ref>
<ref id="B41">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ting</surname> <given-names>K. M.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Confusion matrix</article-title>,&#x201d; in <source>Encyclopedia of Machine Learning and Data Mining</source>. Eds. <person-group person-group-type="editor">
<name>
<surname>Sammut</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Webb</surname> <given-names>G. I.</given-names>
</name>
</person-group> (<publisher-name>Springer</publisher-name>, <publisher-loc>Boston, MA</publisher-loc>). doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-1-4899-7687-1_50</pub-id>
</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Too</surname> <given-names>E. C.</given-names>
</name>
<name>
<surname>Yujian</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Njuki</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Yingchun</surname> <given-names>L.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A comparative study of fine-tuning deep learning models for plant disease identification</article-title>. <source>Comput. Electron. Agric.</source> <volume>161</volume>, <fpage>272</fpage>&#x2013;<lpage>279</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2018.03.032</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Verma</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Chug</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Singh</surname> <given-names>A. P.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Application of convolutional neural networks for evaluation of disease severity in tomato plant</article-title>. <source>Journal of Discrete Mathematical Sciences and Cryptography.</source> <volume>2020</volume> <issue>23</issue> (<issue>1</issue>), <page-range>273-282</page-range>.</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Du</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>J</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>C</given-names>
</name>
<name>
<surname>Zhu</surname> <given-names>H</given-names>
</name>
</person-group>. (<year>2021</year>). <article-title>A cucumber leaf disease severity classification method based on the fusion of DeepLabV3+ and U-Net</article-title>. <source>Computers and electronics in agriculture.</source> <volume>189</volume>, <fpage>106373</fpage>.</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Automatic image-based plant disease severity estimation using deep learning</article-title>. <source>Comput. Intell. Neurosci.</source> <volume>2017</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.1155/2017/2917536</pub-id>
</citation>
</ref>
<ref id="B46">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Woo</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>J. Y.</given-names>
</name>
<name>
<surname>Kweon</surname> <given-names>I. S.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Cbam: Convolutional block attention module</article-title>,&#x201d; in <conf-name>Proceedings of the European conference on computer vision (ECCV)</conf-name>. <fpage>3</fpage>&#x2013;<lpage>19</lpage>.</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wspanialy</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Moussa</surname> <given-names>M.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A detection and severity estimation system for generic diseases of tomato greenhouse plants</article-title>. <source>Comput. Electron. Agric.</source> <volume>178</volume>, <fpage>105701</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2020.105701</pub-id>
</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Lei</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Zhou</surname> <given-names>W.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>SEV-Net: residual network embedded with attention mechanism for plant disease severity detection</article-title>. <source>Concurrency Computation: Pract. Exp.</source> <volume>33</volume>, <fpage>e6161</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/cpe.6161</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>