<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2026.1737464</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>LMP-PM: a lightweight multi-path pruning method for plant leaf disease recognition</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name><surname>Hua</surname><given-names>Jing</given-names></name>
<xref ref-type="corresp" rid="c001"><sup>*</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3281325/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project-administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Zou</surname><given-names>Fendong</given-names></name>
<uri xlink:href="https://loop.frontiersin.org/people/3201538/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhu</surname><given-names>Yuanhao</given-names></name>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Deng</surname><given-names>Jize</given-names></name>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>He</surname><given-names>Ruimin</given-names></name>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><institution>School of Software, Jiangxi Agricultural University</institution>, <city>Nanchang</city>,&#xa0;<country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>*</label>Correspondence: Jing Hua, <email xlink:href="mailto:15179195967@163.com">15179195967@163.com</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-26">
<day>26</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>17</volume>
<elocation-id>1737464</elocation-id>
<history>
<date date-type="received">
<day>01</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>11</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="rev-recd">
<day>04</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Hua, Zou, Zhu, Deng and He.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Hua, Zou, Zhu, Deng and He</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-26">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Plant leaf diseases pose a significant threat to plant growth and productivity, necessitating accurate and timely identification. While high-performance deep learning models exist, their complexity often hinders deployment in real-world, resource-constrained agricultural settings. To address the need for efficient and accurate plant disease identification, we developed a novel lightweight approach named the Lightweight Multi-Path Pruning Method (LMP-PM). LMP-PM offers flexible lightweight optimization, configurable via pruning parameters and path expansion ratios, enabling users to balance significant reductions in model parameters and FLOPs against potential inference time increases, thereby tailoring model size, performance, and real-time needs to specific application scenarios. Specifically, we first constructed an original, high-performance, and complex model (OMNet) incorporating various structures and a three-branch parallel module (TBP block). We then applied LMP-PM to OMNet to perform lightweight processing, resulting in several lightweight models. Through extensive experimentation, we identified the optimal model that balances performance and complexity, which we named LMNet (Lightweight Multi-Path Network). LMNet demonstrates remarkable efficiency, utilizing only 5.69% of the parameters and 3.80% of the FLOPs of OMNet. Despite this substantial reduction in complexity, LMNet achieved superior accuracy: 99.23% on the Plant Village dataset, representing an improvement of 0.58% over OMNet, and 87.27% on the AI 2018 Challenger dataset, surpassing OMNet by 1.91%. These results highlight that LMP-PM successfully creates highly efficient models like LMNet, which not only drastically reduce computational resources but also improve classification accuracy. This flexibility and enhanced performance make LMNet particularly suitable for real-time plant disease identification in resource-constrained environments, offering a practical and effective solution for agricultural applications.</p>
</abstract>
<kwd-group>
<kwd>convolutional neural network</kwd>
<kwd>deep learning</kwd>
<kwd>lightweight</kwd>
<kwd>LMP-PM</kwd>
<kwd>plant disease identification</kwd>
</kwd-group>
<funding-group>
<award-group id="gs1">
<funding-source id="sp1">
<institution-wrap>
<institution>Natural Science Foundation of Jiangxi Province</institution>
<institution-id institution-id-type="doi" vocab="open-funder-registry" vocab-identifier="10.13039/open_funder_registry">10.13039/501100004479</institution-id>
</institution-wrap>
</funding-source>
</award-group>
<award-group id="gs2">
<funding-source id="sp2">
<institution-wrap>
<institution>National Natural Science Foundation of China</institution>
<institution-id institution-id-type="doi" vocab="open-funder-registry" vocab-identifier="10.13039/open_funder_registry">10.13039/501100001809</institution-id>
</institution-wrap>
</funding-source>
</award-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This research was funded by the Natural Science Foundation of Jiangxi Province (Grant No. 20224BAB202038) and the National Natural Science Foundation of China (Grant No.61861021). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.</funding-statement>
</funding-group>
<counts>
<fig-count count="13"/>
<table-count count="9"/>
<equation-count count="24"/>
<ref-count count="59"/>
<page-count count="23"/>
<word-count count="10758"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Sustainable and Intelligent Phytoprotection</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Plant leaf diseases are multifactorial and pose a serious threat to plant productivity (<xref ref-type="bibr" rid="B51">Xiao et&#xa0;al., 2024</xref>). These diseases can lead to reduced yields, plant death, and significant losses, thereby hindering overall food production. Accurate identification of leaf diseases is crucial for mitigating their impact on plant production (<xref ref-type="bibr" rid="B20">Khanna et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B39">Swaminathan and Vairavasundaram, 2024</xref>; <xref ref-type="bibr" rid="B43">Tang L. et&#xa0;al., 2024</xref>). However, the identification of leaf diseases presents challenges due to the wide variety of disease types, the difficulty in recognition, and the short disease cycles. Traditionally, the identification of plant diseases has relied on field sampling conducted by specialists, who then visually observe or analyze samples in a laboratory to determine the specific types of diseases present. However, this method often faces limitations due to a lack of available experts, leading to inefficient time management and the potential to miss critical treatment windows (<xref ref-type="bibr" rid="B32">Pal and Kumar, 2023</xref>; <xref ref-type="bibr" rid="B35">Sahu and Pandey, 2023</xref>; <xref ref-type="bibr" rid="B38">Shi et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B7">Deng et&#xa0;al., 2024</xref>).</p>
<p>In recent years, with the advancement of artificial intelligence technology, deep learning and computer vision techniques have been widely used for the identification of plant diseases (<xref ref-type="bibr" rid="B33">Parthiban et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B36">Salamai et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B49">Wu et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B8">Dong et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B22">Li H. et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B46">Wang H. et&#xa0;al., 2024</xref>). Techniques such as image classification (<xref ref-type="bibr" rid="B4">Cheng et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B25">Lin et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B42">Tang W. et&#xa0;al., 2024</xref>), object detection (<xref ref-type="bibr" rid="B6">Denarda et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B12">Hern&#xe1;ndez et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B56">Yu X. et&#xa0;al., 2024</xref>), and semantic segmentation (<xref ref-type="bibr" rid="B11">Heidari et&#xa0;al., [[NoYear]]</xref>; <xref ref-type="bibr" rid="B19">Jain et&#xa0;al., [[NoYear]]</xref>; <xref ref-type="bibr" rid="B5">Dai et&#xa0;al., 2024</xref>) have gained prominence in this field. Currently, deep learning applications in the identification of plant leaf diseases primarily focus on two key directions. The first direction involves the development of large models (<xref ref-type="bibr" rid="B57">Zhang T. et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B21">Kuska et&#xa0;al., 2024</xref>) and multimodal fusion (<xref ref-type="bibr" rid="B55">Yu H. et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B58">Zhang et&#xa0;al., 2024</xref>). However, large models and multimodal tasks often rely on high-performance computing equipment, which is typically expensive. Consequently, agricultural practitioners in real-world settings frequently lack the purchasing power for such equipment, limiting the applicability of these methods. The second trending direction is the use of lightweight models (<xref ref-type="bibr" rid="B16">Hua et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B17">Huang et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B48">Wang B. et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B52">Xie et&#xa0;al., 2024</xref>). These models do not require expensive hardware and can be effectively applied in practical agricultural production. This advantage has facilitated their wider application and promotion within the agricultural sector. Lightweight models can run on resource-constrained devices such as smartphones, edge computing devices, and small embedded systems, providing quick responses that meet the demands of real-time agricultural production monitoring (<xref ref-type="bibr" rid="B54">Yongda et&#xa0;al., 2024</xref>).</p>
<p>Existing methods for lightweighting plant leaf disease models often achieve their goals by introducing new modules or functions to replace existing, more complex components. For instance, <xref ref-type="bibr" rid="B23">Li R. et&#xa0;al. (2024)</xref> introduced the ECIoU_Loss (EfficiCLoss Loss) function, replacing the original CIoU_Loss. This modification resulted in a 50.2% reduction in model size and a significant 43.1% decrease in FLOPs, achieving a best accuracy of 87.5% in corn leaf disease identification tasks. Similarly, Sandesh et&#xa0;al. (<xref ref-type="bibr" rid="B1">Bhagat et&#xa0;al., 2024</xref>) replaced the standard convolution in Vgg16 with MDsConv, which reduced model parameters by 60% while achieving an accuracy of 94.14%, surpassing the performance of the standard Vgg16 model. Additionally, Siyu et&#xa0;al. (<xref ref-type="bibr" rid="B34">Quan et&#xa0;al., 2024</xref>) utilized partial convolution and point-wise convolution techniques to replace traditional deep convolutions, thereby reducing computational complexity and attaining a best accuracy of 99.04% on the PlantVillage dataset. <xref ref-type="bibr" rid="B53">Xing et&#xa0;al. (2024)</xref> adopted the lightweight network structure ShuffleNetV2 to replace the traditional backbone network Xception in DeepLabv3+. The improved DeepLabv3+ achieved an average pixel accuracy and mean intersection over union of 95.84% and 96.87%, respectively, with a detection rate and weight file size superior to other algorithms. Lastly, <xref ref-type="bibr" rid="B59">Zhang P. et&#xa0;al. (2023)</xref> employed mixed lightweight convolution and spatial pyramid dilated convolution to replace standard convolutions. Their proposed LW-Segnet and LW-Unet models achieved higher F1 scores and intersection-over-union values in seedling detection and cross-variety row segmentation while reducing model parameters.</p>
<p>The lightweight methods described above have successfully enhanced model performance while reducing the complexity of the original models, demonstrating the feasibility of lightweight models in the identification of plant leaf diseases. However, these approaches primarily involve structural replacements of the original model without considering the effects of model pruning and path expansion on both model complexity and performance, often resulting in limited degrees of lightweighting. To address this research gap, we developed a new lightweighting approach named the Lightweight Multi-Path Pruning Method (LMP-PM). LMP-PM significantly reduces the complexity of the original model while allowing for selective control of the degree of lightweighting through the adjustment of pruning parameters and path expansion ratios, thereby catering to varying task requirements. Specifically, we initially constructed an original model (OMNet), which boasts a high level of complexity and incorporates various structures, including a three-branch parallel module (TBP block) designed to optimize performance. Subsequently, we applied LMP-PM to perform lightweight processing on OMNet, resulting in several lightweight models. We selected the best one and named it LMNet. The contributions of this paper are as follows:</p>
<p>LMP-PM: We proposed a new lightweight method, the Lightweight Multi-Path Pruning Method (LMP-PM), which selectively controls the degree of lightweight processing.</p>
<p>OMNet: We introduced a novel base model, OMNet, which incorporates various structures and exhibits high performance and complexity.</p>
<p>LMNet: LMNet is derived from OMNet through the lightweight processing of LMP-PM. It features fewer parameters while achieving superior performance in the identification of leaf diseases.</p>
<p>TBP Block: We developed a new three-branch parallel module designed for multi-scale extraction of plant leaf disease characteristics, effectively enhancing recognition performance.</p>
</sec>
<sec id="s2" sec-type="materials|methods">
<label>2</label>
<title>Materials and methods</title>
<sec id="s2_1">
<label>2.1</label>
<title>PlantVillage dataset</title>
<p>In this study, we utilized the highly regarded PlantVillage dataset (Davi et&#xa0;al., 2015) (<xref ref-type="bibr" rid="B18">Hughes and Marcel, 2015</xref>), which includes 38 categories of plant leaf diseases and a total of 54,305 images. This dataset consists of 36,219 images in the training set, 9,036 images in the validation set, and 9,050 images in the test set, with example images shown in <xref ref-type="fig" rid="f1"><bold>Figure&#xa0;1</bold></xref>.A. This dataset was chosen for its widespread use as a benchmark in plant disease recognition, providing a robust foundation for initial model validation and comparison with existing literature. Each image in this dataset has been rigorously validated and annotated by experienced plant pathologists, ensuring a high level of accuracy and reliability.</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Sample images of diseased leaves from the PlantVillage dataset and the AI Challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g001.tif">
<alt-text content-type="machine-generated">Composite graphic showing two sets of leaf disease images. The Plant Village dataset section presents six images of apple, corn, potato, and tomato leaves with different disease symptoms, each labeled by disease type. The AI Challenger 2018 dataset section displays six leaf images of similar crops, each labeled with specific disease symptoms and severity. Each dataset is grouped and labeled for comparison purposes.</alt-text>
</graphic></fig>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>AI challenger 2018 dataset</title>
<p>To further validate the advancements and generalizability of our method, we also utilized the AI Challenger 2018 Dataset (Wu et&#xa0;al., 2019) (<xref ref-type="bibr" rid="B50">Wu et&#xa0;al., [[NoYear]]</xref>), which comprises 61 categories of plant leaf diseases at varying degrees of severity, totaling 36,075 images. This dataset includes 25,252 images in the training set, 6,289 images in the validation set, and 4,534 images in the test set, with example images displayed in <xref ref-type="fig" rid="f1"><bold>Figure&#xa0;1</bold></xref>.B. Its highly imbalanced category labels make recognition inherently more challenging. This dataset serves as a critical test for model robustness under difficult conditions. In this study, we treated it as an ablation dataset and did not apply any augmentation techniques to specifically assess the model&#x2019;s inherent performance on imbalanced data without external data manipulation.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Lightweight multi-path pruning method</title>
<p>We propose a novel Lightweight Multi-Path Pruning Method (LMP-PM), as illustrated in <xref ref-type="fig" rid="f2"><bold>Figure&#xa0;2</bold></xref>. First, we design an original model, assuming the input feature map is represented by <inline-formula>
<mml:math display="inline" id="im1"><mml:mrow><mml:mi>X</mml:mi><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mi>C</mml:mi><mml:mo>&#xd7;</mml:mo><mml:mi>A</mml:mi><mml:mo>&#xd7;</mml:mo><mml:mi>B</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, with C indicating the number of input channels, and <inline-formula>
<mml:math display="inline" id="im2"><mml:mi>A</mml:mi></mml:math></inline-formula> and <inline-formula>
<mml:math display="inline" id="im3"><mml:mi>B</mml:mi></mml:math></inline-formula> representing the height and width of the input feature map, respectively. We define <inline-formula>
<mml:math display="inline" id="im4"><mml:mi>W</mml:mi></mml:math></inline-formula> as the weight matrix of the original convolutional layer; therefore, the output of the original model can be expressed by formula (1):</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>The overall architecture of lightweight multi-path pruning method.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g002.tif">
<alt-text content-type="machine-generated">Flowchart diagram illustrating a neural network optimization process. The top section shows an original fully connected model. Subsequent sections detail channel pruning, standard to depthwise convolution structural replacement, and generation of multi-path structures, with arrows connecting each transformation stage and decision diamonds guiding logical choices.</alt-text>
</graphic></fig>
<disp-formula id="eq1"><label>(1)</label>
<mml:math display="block" id="M1"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>i</mml:mi><mml:mi>g</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>C</mml:mi></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>f</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>+</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im5"><mml:mi>f</mml:mi></mml:math></inline-formula> represents the index of the output channels, and <inline-formula>
<mml:math display="inline" id="im6"><mml:mrow><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>+</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> is the input feature map. <inline-formula>
<mml:math display="inline" id="im7"><mml:mi>K</mml:mi></mml:math></inline-formula> denote the size of the convolutional kernel, with its center position located at (<inline-formula>
<mml:math display="inline" id="im8"><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:math></inline-formula>, <inline-formula>
<mml:math display="inline" id="im9"><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:math></inline-formula>). Considering the symmetry of the convolutional kernel, indices from <inline-formula>
<mml:math display="inline" id="im10"><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:math></inline-formula> to <inline-formula>
<mml:math display="inline" id="im11"><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:math></inline-formula> can cover the entire kernel. Next, we proceed to lightweight the original model by reducing the number of input feature map channels to <inline-formula>
<mml:math display="inline" id="im12"><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:math></inline-formula>. <inline-formula>
<mml:math display="inline" id="im13"><mml:mi>N</mml:mi></mml:math></inline-formula> is the pruning parameter, and its size is determined based on the complexity of the original model. The&#xa0;specific steps for pruning are as follows: First, we iterate through all layers, sequentially extracting each standard convolutional layer from the original model. For the current convolutional layer, we retrieve its output channel count, denoted as C. Next, a custom pruning parameter, N, is introduced. The value of this parameter N is typically determined based on the original model&#x2019;s complexity and the desired degree of lightweighting. Subsequently, we check if the current layer&#x2019;s output channel count satisfies the pruning condition, specifically C/N &gt; 1. If this condition is not met, the layer is not pruned, thereby preventing excessive compression that could lead to a drastic performance drop. If the pruning condition is met, the layer&#x2019;s output channel count C is updated to C = C/N. This action, in turn, reduces the computational burden on this layer and subsequent layers. Finally, these steps are repeated until all convolutional layers have been traversed and their channel counts adjusted according to the pruning conditions. We define the new convolutional weights as W1, therefore, the output feature map can be expressed as formula&#xa0;(2):</p>
<disp-formula id="eq2"><label>(2)</label>
<mml:math display="block" id="M2"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>+</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where the new convolutional weights <inline-formula>
<mml:math display="inline" id="im14"><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#xd7;</mml:mo><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac><mml:mo>&#xd7;</mml:mo><mml:mi>K</mml:mi><mml:mo>&#xd7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, <inline-formula>
<mml:math display="inline" id="im15"><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> serve as indices for the new output channels. Following this, we perform the next lightweight operation by replacing the standard convolution with depthwise (DW) convolution, as outlined in (<xref ref-type="bibr" rid="B13">Howard et&#xa0;al., 2017</xref>). Subsequently, the process entails re-traversing all layers within the channel-pruned model to determine whether the current layer is a standard convolutional layer. If so, its critical parameters&#x2014;namely, in_channels, out_channels, kernel_size, stride, and padding&#x2014;are extracted. These extracted parameters are then utilized to construct a depthwise convolutional layer. At this point, the input feature map, denoted as <inline-formula>
<mml:math display="inline" id="im16"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>, processed to obtain formula (3):</p>
<disp-formula id="eq3"><label>(3)</label>
<mml:math display="block" id="M3"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>+</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where the weights of the depthwise convolution are <inline-formula>
<mml:math display="inline" id="im17"><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mi>d</mml:mi></mml:msub><mml:mo>&#x2208;</mml:mo><mml:msup><mml:mi>R</mml:mi><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac><mml:mo>&#xd7;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mi>K</mml:mi><mml:mo>&#xd7;</mml:mo><mml:mi>K</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:math></inline-formula>, where <inline-formula>
<mml:math display="inline" id="im18"><mml:mi>j</mml:mi></mml:math></inline-formula> and <inline-formula>
<mml:math display="inline" id="im19"><mml:mi>k</mml:mi></mml:math></inline-formula> serve as the indices for the convolution operation. Immediately following this, we apply pointwise convolution to <inline-formula>
<mml:math display="inline" id="im20"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula>, resulting in formula (4):</p>
<disp-formula id="eq4"><label>(4)</label>
<mml:math display="block" id="M4"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msub><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>Then, to balance the performance loss associated with the two lightweight operations, we convert the single-path model into a multi-path model by controlling the path expansion factor E. The specific steps are as follows: First, we examine the value of the custom path expansion factor, E. If E &#x2264; 1, multi-path expansion is not performed; instead, the lightweight single-path model is utilized directly. However, if E &gt; 1, an empty list is initialized to store the multiple model branches that will be subsequently created. A loop is then initiated, with the loop variable &#x2018;i&#x2019; iterating from 1 to E. In each iteration, a deep copy of the structurally-transformed lightweight single-path model is performed. This implies copying not only the model&#x2019;s architecture but also all of its current weights. Afterwards, the weights of all branches except the original one are randomly initialized. The duplicated model instance is then appended to the pre-initialized list of branches. Upon the loop&#x2019;s completion, the number of paths is determined based on the specific task, resulting in formula (5):</p>
<disp-formula id="eq5"><label>(5)</label>
<mml:math display="block" id="M5"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:msub><mml:mi>f</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:msub><mml:mi>F</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>p</mml:mi><mml:mo>,</mml:mo><mml:msub><mml:mi>f</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:mo>(</mml:mo><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mfrac><mml:mi>C</mml:mi><mml:mi>N</mml:mi></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>j</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:mstyle displaystyle="true"><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow><mml:mrow><mml:mfrac><mml:mi>K</mml:mi><mml:mn>2</mml:mn></mml:mfrac></mml:mrow></mml:munderover><mml:mrow><mml:msub><mml:mi>W</mml:mi><mml:mrow><mml:mi>d</mml:mi><mml:mo>,</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle></mml:mrow></mml:mstyle><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>X</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>a</mml:mi><mml:mo>+</mml:mo><mml:mi>j</mml:mi><mml:mo>,</mml:mo><mml:mi>b</mml:mi><mml:mo>+</mml:mo><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>Finally, we concatenate all the paths, after lightweight optimization, the final output results of the model are shown in formula (6):</p>
<disp-formula id="eq6"><label>(6)</label>
<mml:math display="block" id="M6"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>f</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mi>E</mml:mi></mml:msub><mml:msub><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:msub></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>&#xb7;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mi>n</mml:mi><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mi>E</mml:mi><mml:mo>=</mml:mo><mml:mi>n</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:mi>dim</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>OMNet</title>
<p>We propose an original model (OMNet), whose structure is illustrated in <xref ref-type="fig" rid="f3"><bold>Figure&#xa0;3A</bold></xref>. OMNet was designed to deliver high performance and high complexity, serving as the foundation for subsequent lightweight processing in LMP-PM. OMNet comprises fundamental components such as convolutional layers, pooling layers, activation functions, and normalization functions, while also integrating residual modules (<xref ref-type="bibr" rid="B10">He et&#xa0;al., 2016</xref>), SE modules (<xref ref-type="bibr" rid="B15">Hu et&#xa0;al., 2018</xref>), and TBP blocks. The overall design approach of OMNet involves first downsampling the input image twice, reducing the feature map size by four times while maintaining the number of channels to facilitate subsequent feature extraction. Next, these feature maps enter the main feature extraction structure, which is also the focus of the subsequent validation of LMP-PM. For ease of representation, we divide this portion of the structure into four stages. Let the input image for the first stage be represented as X. Here, we omit the calculations related to batch normalization (BN) and ReLU, and the output of the first stage can be expressed by formula (7) as:</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Model structure diagrams of ONET, LMNet, and TBP block.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g003.tif">
<alt-text content-type="machine-generated">Diagram showing deep learning network architectures and their blocks. OMNet(a) and LMNet(b) visualize multi-stage image classification pipelines using convolution, pooling, TBP, residual, and SE blocks, with data flow and tensor dimensions. Panel (c) illustrates a TBP block with three branches performing various convolutions and aggregations. Panel (d) explains the SE block with global average pooling and dense layer. Panel (e) depicts a residual block applying convolution and summation.</alt-text>
</graphic></fig>
<disp-formula id="eq7"><label>(7)</label>
<mml:math display="block" id="M7"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>X</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where, <inline-formula>
<mml:math display="inline" id="im21"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mi>X</mml:mi></mml:mrow></mml:math></inline-formula> represents the TBP block operation in the first stage, while <inline-formula>
<mml:math display="inline" id="im22"><mml:mi>&#x3b7;</mml:mi></mml:math></inline-formula> denotes the residual module operation. The output of the second stage is similar to that of the first stage, as shown in equation (8):</p>
<disp-formula id="eq8"><label>(8)</label>
<mml:math display="block" id="M8"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>Compared to the first two stages, the third stage includes an additional SE block. The output of the third stage can be expressed by formula (9):</p>
<disp-formula id="eq9"><label>(9)</label>
<mml:math display="block" id="M9"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x3b4;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>b</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>A</mml:mi><mml:mi>v</mml:mi><mml:mi>g</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>Y</mml:mi><mml:mn>3</mml:mn><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>Y</mml:mi><mml:mn>3</mml:mn><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow></mml:msubsup><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where, <inline-formula>
<mml:math display="inline" id="im23"><mml:mi>&#x3b4;</mml:mi></mml:math></inline-formula> represents the Sigmoid activation function in the SE block, while <inline-formula>
<mml:math display="inline" id="im24"><mml:mrow><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi></mml:mrow></mml:math></inline-formula> refers to the fully connected layer within the SE block. In the fourth stage, an additional TBP block is added compared to the third stage, and the other components remain similar. Output as shown in formula (10):</p>
<disp-formula id="eq10"><label>(10)</label>
<mml:math display="block" id="M10"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi>&#x3b4;</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>D</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>s</mml:mi><mml:mi>e</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>G</mml:mi><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>b</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>A</mml:mi><mml:mi>v</mml:mi><mml:mi>g</mml:mi><mml:mi>p</mml:mi><mml:mi>o</mml:mi><mml:mi>o</mml:mi><mml:mi>l</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>Y</mml:mi><mml:mn>4</mml:mn><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:msubsup><mml:mo>'</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>&#x3b7;</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msubsup><mml:mi>Y</mml:mi><mml:mn>4</mml:mn><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow></mml:msubsup><mml:mo>'</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where, <inline-formula>
<mml:math display="inline" id="im25"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>'</mml:mo></mml:mrow></mml:mrow></mml:math></inline-formula> denotes the operations of the TBP block performed twice. Subsequently, the feature maps outputted from the fourth stage are passed through a 1&#xd7;1 convolutional layer followed by an Adaptive Pooling layer. Finally, these are flattened and mapped to the output classes by a fully connected layer.</p>
</sec>
<sec id="s2_5">
<label>2.5</label>
<title>LMNet</title>
<p>LMNet is a lightweight model derived by applying our proposed Lightweight Multi-Path Pruning Method (LMP-PM) to OMNet, with its structure illustrated in <xref ref-type="fig" rid="f3"><bold>Figure&#xa0;3B</bold></xref>. LMNet significantly reduces the model&#x2019;s parameter count and computational complexity while maintaining performance superior to OMNet.</p>
<p>The lightweighting of LMNet is primarily achieved through two core strategies: global channel pruning and path expansion. Compared to OMNet, LMNet reduces the global channel count by one-fourth across the entire network. This implies that the channel dimensions of both convolutional and fully connected layers within the network are proportionally scaled down, leading to a substantial reduction in model parameters and FLOPs.</p>
<p>To compensate for potential performance degradation caused by channel pruning and to enhance the model&#x2019;s feature extraction capabilities, OMNet&#x2019;s main feature extraction component is expanded into two independent, parallel paths. Each path, in itself, constitutes a pruned and lightweighted OMNet main feature extraction structure. These two paths share the same internal structure, but their parallel processing enables the capture of richer and more diverse feature representations.</p>
<p>Specifically, LMNet&#x2019;s feature extraction process unfolds&#xa0;as&#xa0;follows: The input image first passes through the same&#xa0;initial&#xa0;downsampling layer as OMNet. Subsequently, the resulting&#xa0;feature maps simultaneously feed into two independent, pruned&#xa0;OMNet main feature extraction structures. These two computational processes are analogous to OMNet&#x2019;s main feature extraction process detailed in Section 2.4, but with proportionally pruned channel dimensions. The output feature maps from the two parallel paths are then concatenated dimensionally to form a broader feature representation. The final output can be expressed by formula (11):</p>
<disp-formula id="eq11"><label>(11)</label>
<mml:math display="block" id="M11"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mrow><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>B</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>B</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo>'</mml:mo><mml:mo>;</mml:mo><mml:mi>dim</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where, the computation processes for <inline-formula>
<mml:math display="inline" id="im26"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>4</mml:mn></mml:msub></mml:mrow></mml:math></inline-formula> and <inline-formula>
<mml:math display="inline" id="im27"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mn>4</mml:mn></mml:msub><mml:mo>'</mml:mo></mml:mrow></mml:math></inline-formula> are similar to those described in Section 2.4. Subsequently, <inline-formula>
<mml:math display="inline" id="im28"><mml:mrow><mml:msub><mml:mi>B</mml:mi><mml:mrow><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> passes through a 1&#xd7;1 convolutional layer and then an Adaptive Pooling layer before being flattened. Finally, it is mapped to the output classes by a fully connected layer.</p>
</sec>
<sec id="s2_6">
<label>2.6</label>
<title>TBP block</title>
<p>We designed a TBP block, and its specific structure is illustrated in <xref ref-type="fig" rid="f3"><bold>Figure&#xa0;3C</bold></xref>. The input image simultaneously enters branches 1, 2, and 3, where corresponding feature extraction tasks are performed before the features are fused. This design facilitates the extraction of multi-scale complex feature information, thereby enhancing the performance of OMNet. Let X denote the input with dimensions (N,C,H,W), where N is the batch size, C is the number of input channels, and H and W represent the height and width of the input feature map, respectively. The output is denoted as Y. The feature map input to branch 1 first passes through a 1&#xd7;1 convolutional layer, the output is as shown in formula (12):</p>
<disp-formula id="eq12"><label>(12)</label>
<mml:math display="block" id="M12"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>X</mml:mi><mml:mo>;</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where, <inline-formula>
<mml:math display="inline" id="im29"><mml:mi>S</mml:mi></mml:math></inline-formula> represents the stride, which is set to 1 by default. The feature map then passes through a second convolutional layer, as shown in formula (13):</p>
<disp-formula id="eq13"><label>(13)</label>
<mml:math display="block" id="M13"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mi>S</mml:mi><mml:mo>,</mml:mo><mml:mi>P</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mi>d</mml:mi><mml:mi>i</mml:mi><mml:mi>n</mml:mi><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>After passing through a third convolutional layer, we obtain formula (14):</p>
<disp-formula id="eq14"><label>(14)</label>
<mml:math display="block" id="M14"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>v</mml:mi><mml:mn>2</mml:mn><mml:msub><mml:mi>d</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>;</mml:mo><mml:mi>C</mml:mi><mml:mo>,</mml:mo><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im30"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:math></inline-formula> undergoes batch normalization, resulting in formula (15):</p>
<disp-formula id="eq15"><label>(15)</label>
<mml:math display="block" id="M15"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>4</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>&#x3bc;</mml:mi></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msup><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:mi>&#x3b3;</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x3b2;</mml:mi><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where <inline-formula>
<mml:math display="inline" id="im31"><mml:mi>&#x3bc;</mml:mi></mml:math></inline-formula> and <inline-formula>
<mml:math display="inline" id="im32"><mml:mrow><mml:msup><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> represent the mean and variance of the input, <inline-formula>
<mml:math display="inline" id="im33"><mml:mi>&#x3b3;</mml:mi></mml:math></inline-formula> and <inline-formula>
<mml:math display="inline" id="im34"><mml:mi>&#x3b2;</mml:mi></mml:math></inline-formula> are the learnable parameters, and <inline-formula>
<mml:math display="inline" id="im35"><mml:mi>&#x3f5;</mml:mi></mml:math></inline-formula> is a small constant to prevent division by zero. Ultimately, the activation mapping produces the result as shown in formula (16):</p>
<disp-formula id="eq16"><label>(16)</label>
<mml:math display="block" id="M16"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>&#x3bc;</mml:mi></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msup><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:mi>&#x3b3;</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x3b2;</mml:mi><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>Similarly, the output of the second branch can be expressed as formula (17):</p>
<disp-formula id="eq17"><label>(17)</label>
<mml:math display="block" id="M17"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>&#x3bc;</mml:mi></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msup><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:mi>&#x3b3;</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x3b2;</mml:mi><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>the output of the third branch is shown in Formula (18):</p>
<disp-formula id="eq18"><label>(18)</label>
<mml:math display="block" id="M18"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>3</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:mi>&#x3bc;</mml:mi></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msup><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msup><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:mi>&#x3b3;</mml:mi><mml:mo>+</mml:mo><mml:mi>&#x3b2;</mml:mi><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>.</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>Finally, the outputs of the three branches are merged through a concatenation operation as shown in Formula (19):</p>
<disp-formula id="eq19"><label>(19)</label>
<mml:math display="block" id="M19"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mi>Y</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>;</mml:mo><mml:mi>dim</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>After substituting the numerical values, we obtain the comprehensive output formula of the TBP block as shown in Formula (20):</p>
<disp-formula id="eq20"><label>(20)</label>
<mml:math display="block" id="M20"><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mi>o</mml:mi><mml:mi>u</mml:mi><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mi>C</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>t</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x3bc;</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mi>&#x3c3;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mn>2</mml:mn><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>&#x3b3;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x3b2;</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mi>Y</mml:mi><mml:mrow><mml:mn>2</mml:mn><mml:mo>,</mml:mo><mml:mn>3</mml:mn></mml:mrow></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x3bc;</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mi>&#x3c3;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mn>2</mml:mn><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>&#x3b3;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x3b2;</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>,</mml:mo><mml:mi>M</mml:mi><mml:mi>a</mml:mi><mml:mi>x</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mfrac><mml:mrow><mml:mi>X</mml:mi><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mi>&#x3bc;</mml:mi><mml:mn>3</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msqrt><mml:mrow><mml:msub><mml:mi>&#x3c3;</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mn>2</mml:mn><mml:mo>+</mml:mo><mml:mi>&#x3f5;</mml:mi></mml:mrow></mml:msqrt></mml:mrow></mml:mfrac><mml:mo>&#xb7;</mml:mo><mml:msub><mml:mi>&#x3b3;</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>&#x3b2;</mml:mi><mml:mn>3</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mn>0</mml:mn><mml:mo stretchy="false">)</mml:mo><mml:mo>;</mml:mo><mml:mi>dim</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math>
</disp-formula>
</sec>
</sec>
<sec id="s3" sec-type="results">
<label>3</label>
<title>Results</title>
<sec id="s3_1">
<label>3.1</label>
<title>Experimental design</title>
<p>Our experimental conditions are presented in <xref ref-type="table" rid="T1"><bold>Table&#xa0;1</bold></xref>. All ablation experiments were conducted under the following conditions: an initial learning rate of 1&#xd7;10<sup>-4</sup>, 100 iterations, the AdamW optimizer (<xref ref-type="bibr" rid="B28">Loshchilov and Hutter, 2017</xref>), a weight decay of 5&#xd7;10<sup>-2</sup>, and a cross-entropy loss function. To ensure the comparison reflects the model&#x2019;s original performance, no pre-training was applied in any of the experiments. In order to validate the effectiveness of LMP-PM and to demonstrate that LMNet possesses strong performance and generalization capabilities, we conducted four relevant experiments. These included verifying the effectiveness of LMP-PM using two excellent open-source agricultural image datasets, testing the efficacy of the TBP block, assessing the lightweight high-performance attributes of LMNet, and finally, performing a visualization validation. To enhance model robustness and prevent overfitting, standard data augmentation techniques were applied to the training data for the PlantVillage dataset. These included Random resized crop (224x224), Random rotation ([-10&#xb0;, 10&#xb0;]), and Random horizontal flip (0.5 probability). Images were then normalized to a [0.0, 1.0] range. For the AI Challenger 2018 dataset, no augmentation techniques were applied, as stated in Section 2.2, to specifically evaluate the model&#x2019;s performance on its inherent data distribution and imbalance.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Experimental conditions.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Device</th>
<th valign="middle" align="center">Version</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">GPU</td>
<td valign="middle" align="center">ASUS RTX 4070 Ti</td>
</tr>
<tr>
<td valign="middle" align="center">CPU</td>
<td valign="middle" align="center">Intel i5-12490F</td>
</tr>
<tr>
<td valign="middle" align="center">Memory</td>
<td valign="middle" align="center">G.Skill 32 GB RAM</td>
</tr>
<tr>
<td valign="middle" align="center">Frame</td>
<td valign="middle" align="center">PyTorch 2.0.1</td>
</tr>
<tr>
<td valign="middle" align="center">Tool</td>
<td valign="middle" align="center">Python 3.9.13</td>
</tr>
<tr>
<td valign="middle" align="center">GPU accelerator 1</td>
<td valign="middle" align="center">CUDA 11.8</td>
</tr>
<tr>
<td valign="middle" align="center">GPU accelerator 2</td>
<td valign="middle" align="center">CUDNN 8.9.4.25</td>
</tr>
<tr>
<td valign="middle" align="center">Learning rate</td>
<td valign="middle" align="center">1&#xd7;10<sup>-4</sup></td>
</tr>
<tr>
<td valign="middle" align="center">weight decay</td>
<td valign="middle" align="center">5&#xd7;10<sup>-2</sup></td>
</tr>
<tr>
<td valign="middle" align="center">Optimizer</td>
<td valign="middle" align="center">AdamW</td>
</tr>
<tr>
<td valign="middle" align="center">Epoch</td>
<td valign="middle" align="center">100</td>
</tr>
<tr>
<td valign="middle" align="center">Random resized crop</td>
<td valign="middle" align="center">224&#xd7;224</td>
</tr>
<tr>
<td valign="middle" align="center">Random rotation</td>
<td valign="middle" align="center">[-10&#xb0;, 10&#xb0;]</td>
</tr>
<tr>
<td valign="middle" align="center">Random horizontal flip</td>
<td valign="middle" align="center">0.5</td>
</tr>
<tr>
<td valign="middle" align="center">Normalization</td>
<td valign="middle" align="center">[0.0, 1.0]</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Evaluation indicators</title>
<p>We employed several commonly used evaluation metrics to assess the performance of LMNet, including accuracy, precision, recall, and the F1 score, as defined by formulas (21-24). We used the number of parameters and FLOPs as indicators of the model&#x2019;s lightweight characteristics. Additionally, we utilized precision-recall (PR) curves, ROC curves, and confusion matrices to further evaluate LMNet&#x2019;s classification performance.</p>
<disp-formula id="eq21"><label>(21)</label>
<mml:math display="block" id="M21"><mml:mrow><mml:mi>A</mml:mi><mml:mi>c</mml:mi><mml:mi>c</mml:mi><mml:mi>u</mml:mi><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>c</mml:mi><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>T</mml:mi><mml:mi>N</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<disp-formula id="eq22"><label>(22)</label>
<mml:math display="block" id="M22"><mml:mrow><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>N</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<disp-formula id="eq23"><label>(23)</label>
<mml:math display="block" id="M23"><mml:mrow><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>T</mml:mi><mml:mi>P</mml:mi></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>T</mml:mi><mml:mi>P</mml:mi><mml:mo>+</mml:mo><mml:mi>F</mml:mi><mml:mi>P</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<disp-formula id="eq24"><label>(24)</label>
<mml:math display="block" id="M24"><mml:mrow><mml:mi>F</mml:mi><mml:mn>1</mml:mn><mml:mo>-</mml:mo><mml:mi>s</mml:mi><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mo>&#xd7;</mml:mo><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>&#xd7;</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>P</mml:mi><mml:mi>r</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>i</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>o</mml:mi><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mi>R</mml:mi><mml:mi>e</mml:mi><mml:mi>c</mml:mi><mml:mi>a</mml:mi><mml:mi>l</mml:mi><mml:mi>l</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mfrac><mml:mo>,</mml:mo></mml:mrow></mml:math>
</disp-formula>
<p>where TP denotes the number of true-positive samples, FP denotes the number of false-positive samples, FN denotes the number of false-negative samples, and TN denotes the number of true-negative samples.</p>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Verification of the effectiveness of the LMP-PM</title>
<p>To validate the effectiveness of LMP-PM, we designed two sets of experiments, including verification of the lightweight capability of model pruning and assessment of the accuracy enhancement capability through path expansion, thereby comprehensively evaluating the performance of LMP-PM.</p>
<sec id="s3_3_1">
<label>3.3.1</label>
<title>Verification of the effectiveness of model pruning</title>
<p>To verify the impact of model pruning on performance, we compared various performance metrics between the original model (N = 1, E = 1) and the pruned model, with experimental results presented in <xref ref-type="table" rid="T2"><bold>Table&#xa0;2</bold></xref>. As the pruning parameters increased, the parameters and FLOPs of OMNet significantly decreased. When N = 16, further pruning of OMNet resulted in the first convolution having fewer than three output channels, rendering it unable to extract features from color images. Therefore, N = 16 represents the minimum pruned model, with OMNet&#x2019;s parameters reduced from 99,799,144 to 279,884, and FLOPs decreased from 10,371,351,552 to 14,956,000, representing reductions of approximately 358 times and 693 times, respectively. It is noteworthy that while the size of OMNet was greatly reduced at this stage, excessive pruning led to a significant decline in model accuracy. On the Plant Village dataset, the model (N = 16, E = 1) exhibited a 6.27% decrease in Test accuracy compared to OMNet, alongside reductions of 7.76% in Precision, 7.59% in F1-score, and 7.41% in Recall. A similar phenomenon was observed on the AI Challenger 2018 dataset, where the model (N = 16, E = 1) showed a 5.47% decrease in Test accuracy compared to OMNet, with Precision reduced by 10.45%, F1-score by 12.41%, and Recall by 12.91%.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Comparison of model performance after pruning (due to the limited lightweight effect of OMNet when the pruning parameter N = 2, we have excluded the models from this group).</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
<th valign="middle" rowspan="2" align="center">FLOPs</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">N=1, E = 1</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center">98.20%</td>
<td valign="middle" align="center">97.94%</td>
<td valign="middle" align="center">97.79%</td>
<td valign="middle" align="center">85.36%</td>
<td valign="middle" align="center">82.16%</td>
<td valign="middle" align="center">81.04%</td>
<td valign="middle" align="center">80.67%</td>
<td valign="middle" align="center">99799144</td>
<td valign="middle" align="center">10371351552</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 1</td>
<td valign="middle" align="center">98.21%</td>
<td valign="middle" align="center">97.69%</td>
<td valign="middle" align="center">97.32%</td>
<td valign="middle" align="center">97.20%</td>
<td valign="middle" align="center">84.98%</td>
<td valign="middle" align="center">79.43%</td>
<td valign="middle" align="center">78.47%</td>
<td valign="middle" align="center">78.20%</td>
<td valign="middle" align="center">3353720</td>
<td valign="middle" align="center">200291712</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 1</td>
<td valign="middle" align="center">95.80%</td>
<td valign="middle" align="center">94.16%</td>
<td valign="middle" align="center">94.32%</td>
<td valign="middle" align="center">94.75%</td>
<td valign="middle" align="center">83.79%</td>
<td valign="middle" align="center">78.58%</td>
<td valign="middle" align="center">76.43%</td>
<td valign="middle" align="center">76.20%</td>
<td valign="middle" align="center">1102640</td>
<td valign="middle" align="center">53515968</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 1</td>
<td valign="middle" align="center">92.38%</td>
<td valign="middle" align="center">90.44%</td>
<td valign="middle" align="center">90.35%</td>
<td valign="middle" align="center">90.38%</td>
<td valign="middle" align="center">79.89%</td>
<td valign="middle" align="center">71.71%</td>
<td valign="middle" align="center">68.63%</td>
<td valign="middle" align="center">67.76%</td>
<td valign="middle" align="center">279884</td>
<td valign="middle" align="center">14956000</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="fig" rid="f4"><bold>Figure&#xa0;4</bold></xref> details the weight distribution of the first convolutional layer and the final fully connected layer under different pruning parameters. The frequency of weight values is represented on the left vertical axis, while the weight values themselves are displayed on the horizontal axis. From the figure, it is evident that the frequency in the final fully connected layer is significantly higher than that in the first convolutional layer, indicating that the fully connected layer has a greater impact on the model&#x2019;s decisions. In contrast, the first convolutional layer primarily extracts shallow features, which have a relatively smaller influence on decision-making. The model with N = 1 and E = 1 exhibits the highest frequency in both the first convolutional layer and the last fully connected layer. For the other three models, the frequency gradually decreases as the pruning parameters increase, highlighting an improvement in parameter utilization. The first three models display similar weight distribution patterns in the final fully connected layer, suggesting that during the initial pruning stages, the model successfully eliminated many unimportant connections or neurons. Although the remaining weights are fewer in number, they are more concentrated, allowing the pruned models to maintain good decision-making capabilities. However, the third pruning stage resulted in a significant decline in the model&#x2019;s decision-making ability, as the remaining weights could not be effectively concentrated. Notably, the first convolutional layer demonstrated strong sparsity at E = 16, with considerable differences in height among the values, indicating substantial variation among the remaining weights. This suggests that certain important features or connections are critical to the model, while others are deemed extraneous. While this enhances parameter utilization, it may also lead to a certain degree of information loss and variations in model performance.</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Weight distribution of the model before and after pruning.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g004.tif">
<alt-text content-type="machine-generated">Four rows of paired histograms illustrate weight distributions in neural networks at increasing model sizes. Left column shows blue first convolutional layer weights with decreasing frequency as model size increases. Right column shows orange last layer weights, all normally distributed but wider as model size increases. Each row is labeled Model(N=1, E=1), Model(N=4, E=1), Model(N=8, E=1), and Model(N=16, E=1), indicating larger model widths.</alt-text>
</graphic></fig>
</sec>
<sec id="s3_3_2">
<label>3.3.2</label>
<title>Verification of path expansion effectiveness</title>
<p>To address the significant performance degradation caused by model pruning, we conducted path expansion operations. Due to hardware constraints, we only explored five models with different path expansion ratios under varying pruning parameters. The results are presented in <xref ref-type="table" rid="T3"><bold>Tables&#xa0;3</bold></xref>&#x2013;<xref ref-type="table" rid="T5"><bold>5</bold></xref>.</p>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Comparison of models with different path expansion ratios when pruning parameter N = 4 (Bold font indicates the optimal model data).</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">parameters</th>
<th valign="middle" rowspan="2" align="center">FLOPs</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">N=1, E = 1</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center">98.20%</td>
<td valign="middle" align="center">97.94%</td>
<td valign="middle" align="center">97.79%</td>
<td valign="middle" align="center">85.36%</td>
<td valign="middle" align="center">82.16%</td>
<td valign="middle" align="center">81.04%</td>
<td valign="middle" align="center">80.67%</td>
<td valign="middle" align="center">99799144</td>
<td valign="middle" align="center">10371351552</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 1</td>
<td valign="middle" align="center">98.21%</td>
<td valign="middle" align="center">97.69%</td>
<td valign="middle" align="center">97.32%</td>
<td valign="middle" align="center">97.20%</td>
<td valign="middle" align="center">84.98%</td>
<td valign="middle" align="center">79.43%</td>
<td valign="middle" align="center">78.47%</td>
<td valign="middle" align="center">78.20%</td>
<td valign="middle" align="center">3353720</td>
<td valign="middle" align="center">200291712</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 2</td>
<td valign="middle" align="center"><bold>99.23%</bold></td>
<td valign="middle" align="center"><bold>98.97%</bold></td>
<td valign="middle" align="center"><bold>98.94%</bold></td>
<td valign="middle" align="center"><bold>98.91%</bold></td>
<td valign="middle" align="center"><bold>87.27%</bold></td>
<td valign="middle" align="center"><bold>83.72%</bold></td>
<td valign="middle" align="center"><bold>82.91%</bold></td>
<td valign="middle" align="center"><bold>83.06%</bold></td>
<td valign="middle" align="center">5679944</td>
<td valign="middle" align="center">394008320</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 3</td>
<td valign="middle" align="center">98.86%</td>
<td valign="middle" align="center">98.71%</td>
<td valign="middle" align="center">98.51%</td>
<td valign="middle" align="center">98.35%</td>
<td valign="middle" align="center">85.75%</td>
<td valign="middle" align="center">79.47%</td>
<td valign="middle" align="center">78.47%</td>
<td valign="middle" align="center">78.16%</td>
<td valign="middle" align="center">8519192</td>
<td valign="middle" align="center">588302976</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 4</td>
<td valign="middle" align="center">98.96%</td>
<td valign="middle" align="center">98.67%</td>
<td valign="middle" align="center">98.56%</td>
<td valign="middle" align="center">98.49%</td>
<td valign="middle" align="center">86.52%</td>
<td valign="middle" align="center">81.87%</td>
<td valign="middle" align="center">81.80%</td>
<td valign="middle" align="center">82.25%</td>
<td valign="middle" align="center">11358440</td>
<td valign="middle" align="center">782597632</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 5</td>
<td valign="middle" align="center">98.95%</td>
<td valign="middle" align="center">98.59%</td>
<td valign="middle" align="center">98.61%</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center">85.80%</td>
<td valign="middle" align="center">80.79%</td>
<td valign="middle" align="center">79.46%</td>
<td valign="middle" align="center">78.82%</td>
<td valign="middle" align="center">14197688</td>
<td valign="middle" align="center">976892288</td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Comparison of models with different path expansion ratios when pruning parameter N = 8.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
<th valign="middle" rowspan="2" align="center">FLOPs</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">N=1, E = 1</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center">98.20%</td>
<td valign="middle" align="center">97.94%</td>
<td valign="middle" align="center">97.79%</td>
<td valign="middle" align="center">85.36%</td>
<td valign="middle" align="center">82.16%</td>
<td valign="middle" align="center">81.04%</td>
<td valign="middle" align="center">80.67%</td>
<td valign="middle" align="center">99799144</td>
<td valign="middle" align="center">10371351552</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 1</td>
<td valign="middle" align="center">95.80%</td>
<td valign="middle" align="center">94.16%</td>
<td valign="middle" align="center">94.32%</td>
<td valign="middle" align="center">94.75%</td>
<td valign="middle" align="center">83.79%</td>
<td valign="middle" align="center">78.58%</td>
<td valign="middle" align="center">76.43%</td>
<td valign="middle" align="center">76.20%</td>
<td valign="middle" align="center">1102640</td>
<td valign="middle" align="center">53515968</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 2</td>
<td valign="middle" align="center">98.48%</td>
<td valign="middle" align="center">97.76%</td>
<td valign="middle" align="center">97.93%</td>
<td valign="middle" align="center">98.15%</td>
<td valign="middle" align="center"><bold>86.66%</bold></td>
<td valign="middle" align="center"><bold>82.95%</bold></td>
<td valign="middle" align="center"><bold>81.74%</bold></td>
<td valign="middle" align="center"><bold>81.54%</bold></td>
<td valign="middle" align="center">1691032</td>
<td valign="middle" align="center">103744384</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 3</td>
<td valign="middle" align="center">97.16%</td>
<td valign="middle" align="center">96.09%</td>
<td valign="middle" align="center">96.12%</td>
<td valign="middle" align="center">96.24%</td>
<td valign="middle" align="center">84.85%</td>
<td valign="middle" align="center">79.44%</td>
<td valign="middle" align="center">78.18%</td>
<td valign="middle" align="center">77.82%</td>
<td valign="middle" align="center">2535936</td>
<td valign="middle" align="center">154261824</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 4</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center"><bold>98.35%</bold></td>
<td valign="middle" align="center">98.27%</td>
<td valign="middle" align="center">98.22%</td>
<td valign="middle" align="center">86.35%</td>
<td valign="middle" align="center">82.36%</td>
<td valign="middle" align="center">81.53%</td>
<td valign="middle" align="center">81.39%</td>
<td valign="middle" align="center">3380840</td>
<td valign="middle" align="center">204779264</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 5</td>
<td valign="middle" align="center"><bold>98.74%</bold></td>
<td valign="middle" align="center">98.16%</td>
<td valign="middle" align="center"><bold>98.31%</bold></td>
<td valign="middle" align="center"><bold>98.48%</bold></td>
<td valign="middle" align="center">85.75%</td>
<td valign="middle" align="center">82.10%</td>
<td valign="middle" align="center">81.25%</td>
<td valign="middle" align="center">80.94%</td>
<td valign="middle" align="center">4225744</td>
<td valign="middle" align="center">255296704</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
<table-wrap id="T5" position="float">
<label>Table&#xa0;5</label>
<caption>
<p>Comparison of models with different path expansion ratios when pruning parameter N = 16.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
<th valign="middle" rowspan="2" align="center">FLOPs</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">N=1, E = 1</td>
<td valign="middle" align="center">98.65%</td>
<td valign="middle" align="center">98.20%</td>
<td valign="middle" align="center">97.94%</td>
<td valign="middle" align="center">97.79%</td>
<td valign="middle" align="center">85.36%</td>
<td valign="middle" align="center">82.16%</td>
<td valign="middle" align="center">81.04%</td>
<td valign="middle" align="center">80.67%</td>
<td valign="middle" align="center">99799144</td>
<td valign="middle" align="center">10371351552</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 1</td>
<td valign="middle" align="center">92.38%</td>
<td valign="middle" align="center">90.44%</td>
<td valign="middle" align="center">90.35%</td>
<td valign="middle" align="center">90.38%</td>
<td valign="middle" align="center">79.89%</td>
<td valign="middle" align="center">71.71%</td>
<td valign="middle" align="center">68.63%</td>
<td valign="middle" align="center">67.76%</td>
<td valign="middle" align="center">279884</td>
<td valign="middle" align="center">14956000</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 2</td>
<td valign="middle" align="center">97.30%</td>
<td valign="middle" align="center">96.18%</td>
<td valign="middle" align="center">96.30%</td>
<td valign="middle" align="center">96.53%</td>
<td valign="middle" align="center">83.35%</td>
<td valign="middle" align="center">77.85%</td>
<td valign="middle" align="center">75.40%</td>
<td valign="middle" align="center">75.02%</td>
<td valign="middle" align="center">558656</td>
<td valign="middle" align="center">28557248</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 3</td>
<td valign="middle" align="center">96.51%</td>
<td valign="middle" align="center">95.12%</td>
<td valign="middle" align="center">95.41%</td>
<td valign="middle" align="center">95.89%</td>
<td valign="middle" align="center">82.49%</td>
<td valign="middle" align="center">76.39%</td>
<td valign="middle" align="center">73.26%</td>
<td valign="middle" align="center">72.76%</td>
<td valign="middle" align="center">837428</td>
<td valign="middle" align="center">42158496</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 4</td>
<td valign="middle" align="center"><bold>97.48%</bold></td>
<td valign="middle" align="center"><bold>96.84%</bold></td>
<td valign="middle" align="center"><bold>96.89%</bold></td>
<td valign="middle" align="center"><bold>96.99%</bold></td>
<td valign="middle" align="center"><bold>84.80%</bold></td>
<td valign="middle" align="center"><bold>79.79%</bold></td>
<td valign="middle" align="center"><bold>78.64%</bold></td>
<td valign="middle" align="center"><bold>78.48%</bold></td>
<td valign="middle" align="center">1116200</td>
<td valign="middle" align="center">55759744</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 5</td>
<td valign="middle" align="center">97.31%</td>
<td valign="middle" align="center">96.48%</td>
<td valign="middle" align="center">96.41%</td>
<td valign="middle" align="center">96.44%</td>
<td valign="middle" align="center">82.31%</td>
<td valign="middle" align="center">76.15%</td>
<td valign="middle" align="center">73.39%</td>
<td valign="middle" align="center">72.79%</td>
<td valign="middle" align="center">1394972</td>
<td valign="middle" align="center">69360992</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
<p><xref ref-type="table" rid="T3"><bold>Table&#xa0;3</bold></xref> provides a detailed record of various model performance metrics when N = 4, under five different path expansion ratios. Compared to OMNet (E = 1, N = 1), all models, except for the model without path expansion (E = 4, N = 1), exhibited performance that surpassed that of OMNet, with a significant reduction in both parameters and FLOPs. Among the five models under the condition of N = 4, the model with N = 4 and E = 2 demonstrated the best performance. On the Plant Village dataset, this model achieved a Test accuracy improvement of 0.58% over OMNet, with Precision increasing by 0.77%, F1-score by 1.00%, and Recall by 1.12%. On the AI Challenger 2018 dataset, the model (N = 4, E = 2) outperformed OMNet with a Test accuracy improvement of 1.91%, Precision rising by 1.56%, F1-score by 1.87%, and Recall by 2.39%. Notably, the parameters and FLOPs for the model (N = 4, E = 2) were only 5.69% and 3.80% of those for OMNet, respectively.</p>
<p><xref ref-type="table" rid="T4"><bold>Table&#xa0;4</bold></xref> presents the model performance metrics under five different path expansion ratios when N = 8. Compared to N = 4, the model sizes continued to decrease, and some models after path expansion still slightly outperformed OMNet, such as the models with N = 4 and N = 5. Notably, the model (E = 8, N = 5) achieved the best performance on the Plant Village dataset, with a Test accuracy improvement of 0.09% over OMNet, an increase in F1-score of 0.37%, and a Recall enhancement of 0.69%. The parameters and FLOPs for this model were only 4.24% and 2.46% of those for OMNet, respectively. Meanwhile, the model (E = 8, N = 2) exhibited the best performance on the AI Challenger 2018 dataset, achieving a Test accuracy improvement of 1.30% over OMNet, with Precision rising by 0.79%, F1-score by 0.70%, and Recall by 0.87%. The&#xa0;parameters and FLOPs for this model were merely 1.69% and 0.998% of those for OMNet.</p>
<p><xref ref-type="table" rid="T5"><bold>Table&#xa0;5</bold></xref> presents the performance metrics of five models obtained through path expansion on the minimal model when N = 16. At this stage, the model complexity has reached its lowest, resulting in a significant decrease in model accuracy. However, applying path expansion to the minimal model still yielded substantial performance improvements, with the model (N = 16, E = 4) achieving the best results. Compared to the model (N = 16, E = 1), the Test accuracy on the Plant Village dataset improved by 5.10%, Precision increased by 6.10%, F1-score rose by 6.54%, and Recall enhanced by 6.61%. On the AI Challenger 2018 dataset, the Test accuracy increased by 4.91%, Precision improved by 8.08%, F1-score rose by 10.01%, and Recall increased by 10.72%. Although the accuracy of the model (N = 16, E = 4) does not surpass that of OMNet, its Test accuracy on the Plant Village dataset is approximately 98.81% of OMNet&#x2019;s, and on the AI Challenger 2018 dataset, it is about 99.34% of OMNet&#x2019;s. Notably, the parameters of the model (N = 16, E = 4) constitute only 1.12% of those for OMNet, while its FLOPs are merely 0.537% of OMNet&#x2019;s. The weight distribution of the model (N = 16, E = 4) is illustrated in <xref ref-type="fig" rid="f5"><bold>Figure&#xa0;5</bold></xref>, which shows a more concentrated weight distribution in the final fully connected layer compared to the model (N = 16, E = 1) shown in <xref ref-type="fig" rid="f4"><bold>Figure&#xa0;4</bold></xref>. This phenomenon indicates that path expansion is effective in mitigating the performance degradation caused by excessive pruning.</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Weight distribution of the model (N = 16, E = 4).</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g005.tif">
<alt-text content-type="machine-generated">Side-by-side bar charts visualizing weight distributions for a neural network. Left: histogram of first convolution layer weights, blue bars, ranging roughly from minus 0.15 to 0.15. Right: histogram of last layer weights, orange bars, centered near zero, with weights ranging from approximately minus 0.6 to 0.6. Both charts show frequency on the y-axis and weight values on the x-axis with corresponding legends.</alt-text>
</graphic></fig>
<p>We collected inference time data for OMNet and path expansion models on the same device and during the same time period, recording ten data points for each model, as shown in <xref ref-type="fig" rid="f6"><bold>Figure&#xa0;6</bold></xref>. The results indicate that the path expansion ratio significantly impacts the model&#x2019;s inference time. When the path expansion ratio is the same, the inference time does not decrease with the increased degree of pruning. This may be due to the additional reconstruction or adjustments required by model pruning, which could increase inference time. Furthermore, when varying the path expansion ratios, the inference time noticeably increases with higher path expansion ratios. At E = 3 and E = 4, the inference time of the models has already surpassed that of OMNet. This observation suggests that when utilizing LMP-PM, it is essential to consider the increase in inference time associated with higher path expansion ratios.</p>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Inference time of the model on the plant village test dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g006.tif">
<alt-text content-type="machine-generated">Four line charts show inference time in seconds over eleven epochs for machine learning models with varying parameters N and E. Each subplot compares different model configurations, indicated by distinct shapes and colors in the legends.</alt-text>
</graphic></fig>
<p>The experimental results indicate that pruning operations on OMNet can effectively reduce the model&#x2019;s complexity, though this often comes at the cost of model performance. However, the path expansion operation within LMP-PM successfully alleviates this performance decline and can even enhance the performance of OMNet, demonstrating the effectiveness of LMP-PM. In practical applications of LMP-PM, users can select the most appropriate values of N and E based on specific task requirements to achieve optimal lightweight benefits. The accuracy metrics of all models derived from pruning OMNet with LMP-PM are illustrated in <xref ref-type="fig" rid="f7"><bold>Figure&#xa0;7</bold></xref>, where the model with N = 4 and E = 2 exhibits the best performance while also maintaining a shorter inference time than OMNet. We have designated this model as LMNet. Further experiments will be conducted to validate the high-performance characteristics of LMNet.</p>
<fig id="f7" position="float">
<label>Figure&#xa0;7</label>
<caption>
<p>Accuracy evaluation metrics of the pruned model. <bold>(a)</bold> Results on the plant village dataset. <bold>(b)</bold> Results on the AI challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g007.tif">
<alt-text content-type="machine-generated">Radar chart comparing multiple models based on six metrics: Test accuracy, F1-score, Precision, and Recall for cases labeled a and b. Each model is represented by a different colored and styled line, with a legend mapping each color and style to specific model configurations varying in N and E values. The metrics are arranged radially, with higher values closer to the outer edge, and most models cluster toward high values across metrics.</alt-text>
</graphic></fig>
</sec>
<sec id="s3_3_3">
<label>3.3.3</label>
<title>LMP-PM generalization validation</title>
<p>To further validate the generalizability of LMP-PM, we used ResNet101 as the baseline model and pruned it using the LMP-PM method, obtaining multiple models. The experimental results on the Plant Village Dataset and AI Challenger 2018 Dataset are presented in <xref ref-type="table" rid="T6"><bold>Table&#xa0;6</bold></xref>. When N = 4 and E = 4, the pruned model achieved the best performance, with test accuracy on both datasets higher than that of the original Resnet101, and a substantial reduction in the number of parameters. This further demonstrates the effectiveness of LMP-PM.</p>
<table-wrap id="T6" position="float">
<label>Table&#xa0;6</label>
<caption>
<p>Results of ResNet101 pruning using LMP-PM.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" align="center">Plant village dataset</th>
<th valign="middle" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Test-acc</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">ResNet101</td>
<td valign="middle" align="center">98.54%</td>
<td valign="middle" align="center">86.30%</td>
<td valign="middle" align="center">42.63 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=1, E = 1</td>
<td valign="middle" align="center">98.32%</td>
<td valign="middle" align="center">86.08%</td>
<td valign="middle" align="center">23.72 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=2, E = 1</td>
<td valign="middle" align="center">98.19%</td>
<td valign="middle" align="center">85.16%</td>
<td valign="middle" align="center">6.01 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=2, E = 4</td>
<td valign="middle" align="center">98.63%</td>
<td valign="middle" align="center">86.76%</td>
<td valign="middle" align="center">24.02 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 1</td>
<td valign="middle" align="center">97.77%</td>
<td valign="middle" align="center">83.13%</td>
<td valign="middle" align="center">1.54 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=4, E = 4</td>
<td valign="middle" align="center"><bold>99.12%</bold></td>
<td valign="middle" align="center"><bold>86.83%</bold></td>
<td valign="middle" align="center">6.16 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 1</td>
<td valign="middle" align="center">96.82%</td>
<td valign="middle" align="center">82.75%</td>
<td valign="middle" align="center">0.40 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=8, E = 2</td>
<td valign="middle" align="center">98.04%</td>
<td valign="middle" align="center">85.81%</td>
<td valign="middle" align="center">0.81 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 1</td>
<td valign="middle" align="center">93.97%</td>
<td valign="middle" align="center">81.36%</td>
<td valign="middle" align="center">0.11 M</td>
</tr>
<tr>
<td valign="middle" align="center">N=16, E = 4</td>
<td valign="middle" align="center">97.62%</td>
<td valign="middle" align="center">85.01%</td>
<td valign="middle" align="center">0.44 M</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
</sec>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Verification of the effectiveness of the TBP block</title>
<p>To validate the effectiveness of the TBP block, we used LMNet as the baseline model to compare the performance of models with and without the TBP block. The results are presented in <xref ref-type="table" rid="T7"><bold>Table&#xa0;7</bold></xref>. The model with the TBP block outperformed the model without it on both test sets, achieving a significant increase of 3.88% on the Plant Village dataset and 3.48% on the AI 2018 Challenger dataset. Additionally, all other performance evaluation metrics for the model with the TBP block also surpassed those of the model without it. The accuracy and loss trends during training on the training and validation sets for both models are illustrated in <xref ref-type="fig" rid="f8"><bold>Figure&#xa0;8</bold></xref>. From the figure, it can be observed that the model with the TBP block demonstrates superior convergence speed and performance on both datasets. Notably, in the validation set of the Plant Village dataset, the stability of convergence with the TBP block is significantly better than that without it, effectively proving the validity of the TBP block.</p>
<table-wrap id="T7" position="float">
<label>Table&#xa0;7</label>
<caption>
<p>Comparison of the performance between models with and without the TBP block.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">Without TBP block module</td>
<td valign="middle" align="center">95.35%</td>
<td valign="middle" align="center">93.88%</td>
<td valign="middle" align="center">93.91%</td>
<td valign="middle" align="center">94.26%</td>
<td valign="middle" align="center">83.79%</td>
<td valign="middle" align="center">77.95%</td>
<td valign="middle" align="center">76.09%</td>
<td valign="middle" align="center">75.30%</td>
</tr>
<tr>
<td valign="middle" align="center"><bold>With TBP block module</bold></td>
<td valign="middle" align="center"><bold>99.23%</bold></td>
<td valign="middle" align="center"><bold>98.97%</bold></td>
<td valign="middle" align="center"><bold>98.94%</bold></td>
<td valign="middle" align="center"><bold>98.91%</bold></td>
<td valign="middle" align="center"><bold>87.27%</bold></td>
<td valign="middle" align="center"><bold>83.72%</bold></td>
<td valign="middle" align="center"><bold>82.91%</bold></td>
<td valign="middle" align="center"><bold>83.06%</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
<fig id="f8" position="float">
<label>Figure&#xa0;8</label>
<caption>
<p>Comparison of the training processes for models with and without the TBP block. Up: Results on the Plant Village dataset. Down: Results on the AI Challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g008.tif">
<alt-text content-type="machine-generated">Four pairs of line graphs compare model performance over 100 training epochs with and without the TBPblock module. Left graphs show train and validation loss decreasing, and right graphs show train and validation accuracy increasing. Red lines (with TBPblock) consistently outperform black lines (without TBPblock) in all metrics.</alt-text>
</graphic></fig>
</sec>
<sec id="s3_5">
<label>3.5</label>
<title>Comparison of LMNet with other outstanding models</title>
<p>To validate the performance of LMNet, we compared it with six other high-performing models. Among these, ConvNext served as the benchmark model due to its normal complexity and high performance, while the other five models are lightweight models with parameter counts similar to LMNet. The results are presented in <xref ref-type="table" rid="T8"><bold>Table&#xa0;8</bold></xref>. The experimental results indicate that LMNet achieved a Test accuracy on the Plant Village dataset that is 0.62% higher than the second-ranked Shufflenet_v2_x2_0. Additionally, its Precision is 0.97% higher, F1-score is 0.91% higher, and Recall is 0.82% higher. Notably, LMNet&#x2019;s Test accuracy surpasses that of the benchmark model Convnext_tiny by 1.12%, while having only one-fifth of the parameters. On the AI 2018 Challenger dataset, LMNet continued to deliver the best performance, exceeding the second-ranked Repvit_m0_9 by 0.92% in Test accuracy, achieving 0.75% higher Precision, 1.38% higher F1-score, and 1.64% higher Recall. Furthermore, LMNet&#x2019;s Test accuracy also outperformed the benchmark model Convnext_tiny by 5.31%. The changes in Accuracy and loss during training for these seven models on the training set and validation set are shown in <xref ref-type="fig" rid="f9"><bold>Figure&#xa0;9</bold></xref>. From the figure, it is evident that LMNet exhibits superior convergence speed, stability, and overall performance compared to the other models, providing strong evidence for its outstanding capabilities.</p>
<table-wrap id="T8" position="float">
<label>Table&#xa0;8</label>
<caption>
<p>Performance comparison of LMNet with other models (no pre-training).</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">ConvNeXt_tiny (<xref ref-type="bibr" rid="B27">Liu et&#xa0;al., 2022</xref>)</td>
<td valign="middle" align="center">98.11%</td>
<td valign="middle" align="center">97.75%</td>
<td valign="middle" align="center">97.42%</td>
<td valign="middle" align="center">97.15%</td>
<td valign="middle" align="center">81.96%</td>
<td valign="middle" align="center">74.74%</td>
<td valign="middle" align="center">73.62%</td>
<td valign="middle" align="center">73.35%</td>
<td valign="middle" align="center">28589128</td>
</tr>
<tr>
<td valign="middle" align="center">RegNetX_400MF (<xref ref-type="bibr" rid="B40">Szegedy et&#xa0;al., 2015</xref>)</td>
<td valign="middle" align="center">97.06%</td>
<td valign="middle" align="center">95.76%</td>
<td valign="middle" align="center">96.00%</td>
<td valign="middle" align="center">96.32%</td>
<td valign="middle" align="center">85.29%</td>
<td valign="middle" align="center">79.40%</td>
<td valign="middle" align="center">78.31%</td>
<td valign="middle" align="center">77.91%</td>
<td valign="middle" align="center">5157512</td>
</tr>
<tr>
<td valign="middle" align="center">Shufflenet_v2_x2_0 (<xref ref-type="bibr" rid="B30">Ma et&#xa0;al., 2018</xref>)</td>
<td valign="middle" align="center">98.61%</td>
<td valign="middle" align="center">98.00%</td>
<td valign="middle" align="center">98.03%</td>
<td valign="middle" align="center">98.09%</td>
<td valign="middle" align="center">84.14%</td>
<td valign="middle" align="center">78.30%</td>
<td valign="middle" align="center">77.50%</td>
<td valign="middle" align="center">77.40%</td>
<td valign="middle" align="center">7393996</td>
</tr>
<tr>
<td valign="middle" align="center">Mobilenet_v3_large (<xref ref-type="bibr" rid="B14">Howard et&#xa0;al., 2019</xref>)</td>
<td valign="middle" align="center">97.45%</td>
<td valign="middle" align="center">96.83%</td>
<td valign="middle" align="center">96.68%</td>
<td valign="middle" align="center">96.65%</td>
<td valign="middle" align="center">86.13%</td>
<td valign="middle" align="center">81.72%</td>
<td valign="middle" align="center">80.65%</td>
<td valign="middle" align="center">80.27%</td>
<td valign="middle" align="center">5483032</td>
</tr>
<tr>
<td valign="middle" align="center">DemoNet_d12_w256_sum (<xref ref-type="bibr" rid="B29">Ma et&#xa0;al., 2024</xref>)</td>
<td valign="middle" align="center">94.69%</td>
<td valign="middle" align="center">92.85%</td>
<td valign="middle" align="center">93.19%</td>
<td valign="middle" align="center">93.68%</td>
<td valign="middle" align="center">82.40%</td>
<td valign="middle" align="center">76.14%</td>
<td valign="middle" align="center">74.80%</td>
<td valign="middle" align="center">74.37%</td>
<td valign="middle" align="center">7716584</td>
</tr>
<tr>
<td valign="middle" align="center">Repvit_m0_9 (<xref ref-type="bibr" rid="B44">Wang A. et&#xa0;al., 2024</xref>)</td>
<td valign="middle" align="center">98.39%</td>
<td valign="middle" align="center">97.66%</td>
<td valign="middle" align="center">97.81%</td>
<td valign="middle" align="center">98.00%</td>
<td valign="middle" align="center">86.35%</td>
<td valign="middle" align="center">82.97%</td>
<td valign="middle" align="center">81.53%</td>
<td valign="middle" align="center">81.42%</td>
<td valign="middle" align="center">5103560</td>
</tr>
<tr>
<td valign="middle" align="center">EfficientNetB0 (<xref ref-type="bibr" rid="B41">Tan and Le, 2019</xref>)</td>
<td valign="middle" align="center">98.57%</td>
<td valign="middle" align="center">98.24%</td>
<td valign="middle" align="center">98.09%</td>
<td valign="middle" align="center">98.11%</td>
<td valign="middle" align="center">84.45%</td>
<td valign="middle" align="center">78.96%</td>
<td valign="middle" align="center">78.20%</td>
<td valign="middle" align="center">78.07%</td>
<td valign="middle" align="center">8423848</td>
</tr>
<tr>
<td valign="middle" align="center">GhostNet_3.0x (<xref ref-type="bibr" rid="B9">Han et&#xa0;al., 2020</xref>)</td>
<td valign="middle" align="center">97.81%</td>
<td valign="middle" align="center">97.26%</td>
<td valign="middle" align="center">97.34%</td>
<td valign="middle" align="center">97.28%</td>
<td valign="middle" align="center">84.60%</td>
<td valign="middle" align="center">80.07%</td>
<td valign="middle" align="center">79.01%</td>
<td valign="middle" align="center">78.99%</td>
<td valign="middle" align="center">8469280</td>
</tr>
<tr>
<td valign="middle" align="center">PVT_Tiny (<xref ref-type="bibr" rid="B47">Wang et&#xa0;al., 2021</xref>)</td>
<td valign="middle" align="center">96.89%</td>
<td valign="middle" align="center">96.25%</td>
<td valign="middle" align="center">96.59%</td>
<td valign="middle" align="center">96.41%</td>
<td valign="middle" align="center">85.53%</td>
<td valign="middle" align="center">80.16%</td>
<td valign="middle" align="center">79.10%</td>
<td valign="middle" align="center">79.27%</td>
<td valign="middle" align="center">12900072</td>
</tr>
<tr>
<td valign="middle" align="center">EdgeNeXt_Small (<xref ref-type="bibr" rid="B31">Maaz et&#xa0;al., 2022</xref>)</td>
<td valign="middle" align="center">98.68%</td>
<td valign="middle" align="center">98.00%</td>
<td valign="middle" align="center">98.35%</td>
<td valign="middle" align="center">98.16%</td>
<td valign="middle" align="center">81.78%</td>
<td valign="middle" align="center">78.16%</td>
<td valign="middle" align="center">74.28%</td>
<td valign="middle" align="center">74.85%</td>
<td valign="middle" align="center">7463368</td>
</tr>
<tr>
<td valign="middle" align="center"><bold>LMNet</bold></td>
<td valign="middle" align="center"><bold>99.23%</bold></td>
<td valign="middle" align="center"><bold>98.97%</bold></td>
<td valign="middle" align="center"><bold>98.94%</bold></td>
<td valign="middle" align="center"><bold>98.91%</bold></td>
<td valign="middle" align="center"><bold>87.27%</bold></td>
<td valign="middle" align="center"><bold>83.72%</bold></td>
<td valign="middle" align="center"><bold>82.91%</bold></td>
<td valign="middle" align="center"><bold>83.06%</bold></td>
<td valign="middle" align="center">5679944</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
<fig id="f9" position="float">
<label>Figure&#xa0;9</label>
<caption>
<p>ROC curve diagram. Up: results on the plant village dataset. Down: Results on the AI challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g009.tif">
<alt-text content-type="machine-generated">Grid of eight ROC curve plots displaying the true positive rate versus false positive rate for multiple classes, each labeled with specific categories and corresponding AUC values, indicating high model performance in multi-class classification tasks.</alt-text>
</graphic></fig>
<p>All comparative models in <xref ref-type="table" rid="T9"><bold>Table&#xa0;9</bold></xref> utilized pre-trained weights. On the Plant Village dataset, LMNet, with only 5.67M parameters, achieved a test accuracy of 99.23%, a score that even surpassed the pre-trained ConvNeXt_tiny (99.11%), despite the latter having a significantly larger parameter count of 28.6M. This comparison strongly demonstrates LMNet&#x2019;s remarkable advantages in parameter efficiency and performance. Furthermore, on the AI Challenger 2018 dataset, LMNet achieved a test accuracy of 87.27%, ranking first among all models that used pre-trained weights. Moreover, LMNet also exhibited the best performance across key metrics such as Test-acc, Precision, F1-score, and Recall on this dataset. These results conclusively prove that LMNet can demonstrate exceptional generalization ability and leading performance, even without relying on pre-training.</p>
<table-wrap id="T9" position="float">
<label>Table&#xa0;9</label>
<caption>
<p>Performance comparison of LMNet with other outstanding models (pre-training).</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" rowspan="2" align="center">Method</th>
<th valign="middle" colspan="4" align="center">Plant village dataset</th>
<th valign="middle" colspan="4" align="center">AI challenger 2018 dataset</th>
<th valign="middle" rowspan="2" align="center">Parameters</th>
</tr>
<tr>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Test-acc</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">ConvNeXt_tiny (<xref ref-type="bibr" rid="B27">Liu et&#xa0;al., 2022</xref>)</td>
<td valign="middle" align="center">99.11%</td>
<td valign="middle" align="center">98.76%</td>
<td valign="middle" align="center">98.82%</td>
<td valign="middle" align="center">98.69%</td>
<td valign="middle" align="center">86.24%</td>
<td valign="middle" align="center">81.60%</td>
<td valign="middle" align="center">80.26%</td>
<td valign="middle" align="center">79.83%</td>
<td valign="middle" align="center">28589128</td>
</tr>
<tr>
<td valign="middle" align="center">RegNetX_400MF (<xref ref-type="bibr" rid="B40">Szegedy et&#xa0;al., 2015</xref>)</td>
<td valign="middle" align="center">99.03%</td>
<td valign="middle" align="center">98.78%</td>
<td valign="middle" align="center">98.90%</td>
<td valign="middle" align="center">98.79%</td>
<td valign="middle" align="center">86.70%</td>
<td valign="middle" align="center">82.00%</td>
<td valign="middle" align="center">79.60%</td>
<td valign="middle" align="center">79.32%</td>
<td valign="middle" align="center">5157512</td>
</tr>
<tr>
<td valign="middle" align="center">Mobilenet_v3_large (<xref ref-type="bibr" rid="B14">Howard et&#xa0;al., 2019</xref>)</td>
<td valign="middle" align="center">98.12%</td>
<td valign="middle" align="center">97.55%</td>
<td valign="middle" align="center">97.22%</td>
<td valign="middle" align="center">96.97%</td>
<td valign="middle" align="center">85.05%</td>
<td valign="middle" align="center">80.31%</td>
<td valign="middle" align="center">78.74%</td>
<td valign="middle" align="center">78.59%</td>
<td valign="middle" align="center">5483032</td>
</tr>
<tr>
<td valign="middle" align="center">EfficientNetB0 (<xref ref-type="bibr" rid="B41">Tan and Le, 2019</xref>)</td>
<td valign="middle" align="center">98.10%</td>
<td valign="middle" align="center">96.99%</td>
<td valign="middle" align="center">97.11%</td>
<td valign="middle" align="center">97.32%</td>
<td valign="middle" align="center">84.27%</td>
<td valign="middle" align="center">79.05%</td>
<td valign="middle" align="center">77.32%</td>
<td valign="middle" align="center">77.82%</td>
<td valign="middle" align="center">8423848</td>
</tr>
<tr>
<td valign="middle" align="center">PVT_Tiny (<xref ref-type="bibr" rid="B47">Wang et&#xa0;al., 2021</xref>)</td>
<td valign="middle" align="center">99.04%</td>
<td valign="middle" align="center">98.77%</td>
<td valign="middle" align="center">98.63%</td>
<td valign="middle" align="center">98.52%</td>
<td valign="middle" align="center">86.04%</td>
<td valign="middle" align="center">80.94%</td>
<td valign="middle" align="center">79.45%</td>
<td valign="middle" align="center">79.12%</td>
<td valign="middle" align="center">12900072</td>
</tr>
<tr>
<td valign="middle" align="center">EdgeNeXt_Small (<xref ref-type="bibr" rid="B31">Maaz et&#xa0;al., 2022</xref>)</td>
<td valign="middle" align="center">98.98%</td>
<td valign="middle" align="center">98.73%</td>
<td valign="middle" align="center">98.67%</td>
<td valign="middle" align="center">98.64%</td>
<td valign="middle" align="center">85.95%</td>
<td valign="middle" align="center">77.15%</td>
<td valign="middle" align="center">76.03%</td>
<td valign="middle" align="center">76.09%</td>
<td valign="middle" align="center">7463368</td>
</tr>
<tr>
<td valign="middle" align="center"><bold>LMNet</bold></td>
<td valign="middle" align="center"><bold>99.23%</bold></td>
<td valign="middle" align="center"><bold>98.97%</bold></td>
<td valign="middle" align="center"><bold>98.94%</bold></td>
<td valign="middle" align="center"><bold>98.91%</bold></td>
<td valign="middle" align="center"><bold>87.27%</bold></td>
<td valign="middle" align="center"><bold>83.72%</bold></td>
<td valign="middle" align="center"><bold>82.91%</bold></td>
<td valign="middle" align="center"><bold>83.06%</bold></td>
<td valign="middle" align="center">5679944</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Bold font indicates the optimal model data.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>Other results in <xref ref-type="table" rid="T9"><bold>Table&#xa0;9</bold></xref> indicate that models like ConvNeXt_tiny, RegNetX_400MF, and PVT_Tiny showed significant performance improvements after loading pre-trained weights. For instance, ConvNeXt_tiny&#x2019;s test accuracy on the Plant Village dataset increased from 98.11% (as shown in <xref ref-type="table" rid="T8"><bold>Table&#xa0;8</bold></xref>) to 99.11%, and on the AI Challenger 2018 dataset, it saw a substantial increase from 81.96% to 86.24%. RegNetX_400MF&#x2019;s test accuracy on the Plant Village dataset also rose from 97.06% to 99.03%. Similarly, PVT_Tiny&#x2019;s test accuracy on the Plant Village dataset improved from 96.89% to 99.04%. This highlights the critical role of pre-training in enhancing the performance of these general-purpose models on target tasks.</p>
<p>However, not all models exhibited improvement after loading pre-trained weights. For example, Mobilenet_v3_large and EfficientNetB0 even showed a slight performance decrease on the AI Challenger 2018 dataset, which can be attributed to the domain discrepancy between the pre-training task and the downstream task. LMNet&#x2019;s ability to achieve such excellent results without pre-training, especially surpassing SOTA models that rely on ImageNet pre-training, strongly suggests its architectural advantages in learning specific agricultural features. LMNet&#x2019;s design is more focused on extracting unique, fine-grained visual patterns specific to plant diseases and agricultural scenes, such as subtle variations in leaf texture, the shape and color characteristics of lesions, and nuanced differences in crop growth stages.</p>
<p>Unlike general SOTA models (e.g., ConvNeXt_tiny), which are typically pre-trained on large-scale, diverse general datasets like ImageNet, the features they learn may be more geared towards general object recognition. When these general features are transferred to highly specialized agricultural domains, a &#x201c;domain mismatch&#x201d; issue can arise, leading to reduced effectiveness of some general features for agricultural tasks, or even introducing noise. LMNet, through its specially designed modules and connection mechanisms, likely avoids the interference of general features, directly and efficiently learning and encoding these domain-specific, highly discriminative features from agricultural data. This &#x201c;from-scratch,&#x201d; targeted learning approach enables LMNet to better adapt to the characteristics of agricultural images, thereby demonstrating excellent generalization ability and performance even without the need for external large-scale pre-training data.</p>
</sec>
<sec id="s3_6">
<label>3.6</label>
<title>Visualization verification</title>
<p>Although LMNet achieved the best performance on both the Plant Village dataset and the AI Challenger 2018 dataset, the performance gap between these two datasets reached as much as 11.96%. To investigate the reasons behind this discrepancy, we utilized Matplotlib to plot the ROC curves, PR curves, and confusion matrices for LMNet on these two datasets, visualizing its performance and classification capabilities across different categories. The results are presented in <xref ref-type="fig" rid="f9"><bold>Figures&#xa0;9</bold></xref>&#x2013;<xref ref-type="fig" rid="f11"><bold>11</bold></xref>. From <xref ref-type="fig" rid="f9"><bold>Figure&#xa0;9</bold></xref>, it is evident that the ROC curves for LMNet on both datasets are very close to the top left corner, indicating a high true positive rate and a low false positive rate. Furthermore, the AUC values approach or equal 1, demonstrating LMNet&#x2019;s robust ability to identify positive samples. In <xref ref-type="fig" rid="f10"><bold>Figure&#xa0;10</bold></xref>, the PR curve for LMNet on the Plant Village dataset also approaches the top right corner, showing high precision and recall. However, the multiple PR curves for LMNet on the AI Challenger 2018 dataset do not converge towards the top right corner. Notably, labels 15, 39, 48, and 49 exhibit substantial deviations, suggesting that certain categories in the AI Challenger 2018 dataset have very few positive samples. This scarcity likely contributes to the poor performance of most classification models on this dataset. The confusion matrix depicted in <xref ref-type="fig" rid="f11"><bold>Figure&#xa0;11</bold></xref> clearly illustrates the number of correctly classified samples for all categories by LMNet on both datasets. LMNet successfully identified the majority of disease images in the Plant Village dataset; however, it made classification errors for several categories in the AI Challenger 2018 dataset, including labels 18, 19, 54, and 55. This indicates that specific categories require attention in future model optimization efforts.</p>
<fig id="f10" position="float">
<label>Figure&#xa0;10</label>
<caption>
<p>PR curve diagram. Up: Results on the plant village dataset. Down: Results on the AI challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g010.tif">
<alt-text content-type="machine-generated">Grouped image of eight precision-recall curve graphs, each comparing precision and recall metrics for multiple classes in a classification problem. Each subplot includes a legend specifying the class name and average precision (AP) value, with curves represented by distinct colors and line styles. Axes are labeled &#x201c;Precision&#x201d; and &#x201c;Recall,&#x201d; and class names relate to plant diseases or numeric classes. These charts visually assess model performance across classes or categories.</alt-text>
</graphic></fig>
<fig id="f11" position="float">
<label>Figure&#xa0;11</label>
<caption>
<p>Confusion matrix diagram. Up: Results on the plant village dataset. Down: Results on the AI challenger 2018 dataset.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g011.tif">
<alt-text content-type="machine-generated">Top panel shows a confusion matrix with true labels on the y-axis and predicted labels on the x-axis, numbers concentrated along the diagonal, and color intensity increasing with higher values; a vertical color bar ranges from zero to nine hundred. Bottom panel displays a similar confusion matrix layout, with diagonal dominance and a color bar ranging from zero to about three hundred, reflecting model classification performance for datasets with different class counts.</alt-text>
</graphic></fig>
<p>Finally, we used Grad-CAM (<xref ref-type="bibr" rid="B37">Selvaraju et&#xa0;al., [[NoYear]]</xref>) to visualize the feature extraction process of LMNet and the class activation maps of some exemplary images, as shown in <xref ref-type="fig" rid="f12"><bold>Figures&#xa0;12</bold></xref>, <xref ref-type="fig" rid="f13"><bold>13</bold></xref>. In <xref ref-type="fig" rid="f12"><bold>Figure&#xa0;12</bold></xref>, using the first 15 channels as an example, we observe that different models have similar extraction effects on the shallow features (such as shape and color) of diseased images during the feature extraction process. However, in the extraction of deeper features for plant leaf diseases, the model without the TBP block module fails to extract many abstract features due to its weaker multi-scale extraction capability. Both LMNet and OMNet can extract more in-depth abstract features. However, while OMNet extracts a large number of abstract features, some of its channel feature maps show overly smooth edges, such as channels 4, 5, 9, and 10. This could be due to a slight overfitting caused by OMNet&#x2019;s higher complexity. When outputting the disease category, OMNet has the highest score for the correct category but fails to sufficiently suppress incorrect categories. In contrast, LMNet shows the most concentrated correct category scores and the most suppressed incorrect category scores, indicating its superior feature extraction capabilities. In <xref ref-type="fig" rid="f13"><bold>Figure&#xa0;13</bold></xref>, we can see that LMNet identifies a larger area and with more precise localization of leaf diseases compared to ConvNeXt-Tiny and Repvit_m0_9. This suggests that LMNet performs better in identifying plant leaf diseases.</p>
<fig id="f12" position="float">
<label>Figure&#xa0;12</label>
<caption>
<p>Comparison of feature extraction process visualizations.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g012.tif">
<alt-text content-type="machine-generated">Diagram comparing feature extraction across three models for a leaf image. Each model displays shallow and deep feature maps for multiple channels and a classifier output bar graph. Models are labeled Without TBPblock module, OMNet, and LMNet.</alt-text>
</graphic></fig>
<fig id="f13" position="float">
<label>Figure&#xa0;13</label>
<caption>
<p>Comparison of Grad-CAM class activation mapping visualizations.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1737464-g013.tif">
<alt-text content-type="machine-generated">Comparison chart showing plant leaf images from Plant Village and AI Challenger 2018 datasets alongside heatmaps generated by ConvNeXt-Tiny, Repvit_m0_9, and LMNet models, highlighting regions of model focus for plant disease detection.</alt-text>
</graphic></fig>
</sec>
</sec>
<sec id="s4" sec-type="discussion">
<label>4</label>
<title>Discussion</title>
<p>We developed a new lightweight method for the task of plant leaf disease recognition, named the Lightweight Multi-Path Pruning Method (LMP-PM). LMP-PM flexibly performs selective lightweighting of the original model (OMNet) by adjusting the pruning parameter (N) and the path expansion factor (E) to meet the needs of different tasks. The OMNet has a complex structure that includes various modules, with the TBP block designed for multi-scale feature extraction to enhance recognition accuracy. Using LMP-PM, we lightweighted the original model to generate multiple variants and selected the model with a pruning parameter N of 4 and a path expansion factor E of 2, naming it LMNet. Subsequently, we conducted several ablation experiments on two large open-source plant leaf datasets: the Plant Village dataset and the AI Challenger 2018 dataset. The experimental results showed that, with fewer parameters, LMNet outperformed the original model and exceeded other similarly parametered lightweight models in multiple classification performance metrics. This further validates the effectiveness of our method.</p>
<p>The effectiveness of LMP-PM stems from its two-pronged approach, combining a pruning technique with a path expansion mechanism. This dual strategy enables the model to strike a precise balance between model size and accuracy. The pruning mechanism systematically reduces model complexity by decreasing the channel count of the input feature maps. As illustrated in <xref ref-type="table" rid="T2"><bold>Table&#xa0;2</bold></xref>, increasing the parameter N leads to a significant reduction in both parameters and FLOPs. However, excessive pruning alone can lead to a significant degradation in performance. For instance, a noticeable drop in accuracy is observed when N = 16 and E = 1. This suggests that while pruning effectively removes redundant information, it can also inadvertently discard critical features.</p>
<p>The path expansion mechanism counteracts this potential performance degradation by reintroducing model capacity and enhancing multi-scale feature learning. By transforming a single-path model into a multi-path architecture, it enables the model to learn more diverse and robust representations without significantly increasing the channel count for each individual path. As demonstrated in <xref ref-type="table" rid="T3"><bold>Tables&#xa0;3</bold></xref>-<xref ref-type="table" rid="T5"><bold>5</bold></xref>, path expansion effectively mitigates the performance loss incurred by pruning and can even elevate model accuracy to surpass that of the original OMNet. For instance, the model configured with N = 4 and E = 2 achieves higher accuracy than OMNet, despite a significant reduction in both parameters and FLOPs. This suggests that the parallel paths empower the model to capture a richer set of features, thereby compensating for the information loss introduced by pruning.</p>
<p>Despite LMP-PM demonstrating significant advantages in lightweighting models for plant disease recognition, several notable limitations warrant future research. For instance, the optimal selection of pruning parameter (N) and path expansion factor (E) currently relies heavily on empirical experimentation. There is no automatic mechanism to determine the best N and E based on varying datasets or tasks. Consequently, identifying the optimal N and E for new tasks or datasets still necessitates a time-consuming search process. Furthermore, this research primarily focuses on global channel reduction. The lack of exploration into finer-grained pruning strategies, as well as an investigation into the impact of specific layers or blocks within the pruned network on model lightweighting and application effectiveness, represents an area for potential further optimization. Moreover, as depicted in <xref ref-type="fig" rid="f6"><bold>Figure&#xa0;6</bold></xref>, increasing the path expansion factor (E) can lead to extended inference times. In latency-sensitive applications, this could potentially offset some of the lightweighting benefits. This suggests that while path expansion enhances accuracy, it may introduce computational overhead that requires further optimization for seamless real-time edge deployment. Finally, this study lacks experimental validation on actual mobile or embedded hardware. The method&#x2019;s robustness to environmental noise, mixed symptoms, and domain shift was also not assessed. Addressing these issues in future work will significantly enhance the proposed method&#x2019;s practical relevance and reliability.</p>
<p>We must explicitly state that OMNet was designed as a high-complexity, high-performance baseline model, integrating various structures such as residual modules, SE blocks, and our proposed TBP block. This design makes OMNet substantially large in terms of parameters and FLOPs. This is not accidental, but rather intended to serve as a rigorous stress test for the LMP-PM method. By applying lightweight processing to such a complex baseline, we aim to demonstrate that LMP-PM can not only simply compress models, but also effectively and significantly simplify a resource-intensive architecture while maintaining or even enhancing performance.</p>
<p>Transfer learning, particularly using ImageNet pre-trained weights, is a common and often effective method for accelerating model convergence and improving generalization capabilities on general tasks. However, for highly specialized domains like plant disease recognition, where subtle, fine-grained visual cues are critical, features learned from general datasets may not always be optimal. This can lead to domain shift issues, where general features may not transfer perfectly, or might even obscure domain-specific discriminative patterns.</p>
<p>It is in this context that LMNet achieved a test accuracy of 99.23% on the Plant Village dataset (outperforming the pre-trained ConvNeXt_tiny&#x2019;s 99.11%, despite its larger parameter count), and ranked first among all compared models on the AI Challenger 2018 dataset with an accuracy of 87.27%. Crucially, all of this was achieved without any pre-training. This demonstrates that a carefully designed, lightweight architecture, optimized for the target domain from scratch, can effectively learn domain-specific features. These features exhibit comparable or even stronger discriminative power compared to those obtained through fine-tuning after general large-scale pre-training.</p>
<p>Therefore, LMP-PM&#x2019;s unique value proposition extends beyond mere model compression. It offers a paradigm of architectural innovation and optimization, enabling the creation of highly efficient, domain-specific models. By systematically refining complex baseline models like OMNet, LMP-PM allows developed models to inherently adapt to the nuances of agricultural images, rather than relying on potentially mismatched general features. This provides an alternative pathway to achieving high performance in resource-constrained, domain-specific applications, where building optimized architectures from scratch may prove more effective than simply adapting pre-trained general models.</p>
<p>In the future, we aim to effectively apply LMP-PM in practical agricultural production to alleviate the application limitations posed by the large size of deep learning models. Additionally, we plan to further optimize LMP-PM based on its performance in real-world applications. This includes enhancing the adaptability of the lightweighted models for small hardware devices and researching the impact of pruning locations on the degree of model lightweighting and application efficacy. Building on this foundation, we will consider using LMP-PM for the lightweighting of more complex deep learning models, such as large-scale natural language processing models and vision models. Simultaneously, we will explore the effective integration of LMP-PM with deep learning optimization techniques. Knowledge distillation serves as an efficient lightweighting strategy by transferring features extracted from a high-performing, complex teacher model to a lower-complexity student model, thereby improving the performance of the student model (<xref ref-type="bibr" rid="B2">Cai et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B45">Wang J. et&#xa0;al., 2024</xref>). LMP-PM can provide multiple student models for knowledge distillation, further advancing the development of model lightweighting. Additionally, multimodal fusion techniques can combine data from different modes or sources (such as text, images, and audio) to acquire more comprehensive information, thereby enhancing the performance and effectiveness of tasks (<xref ref-type="bibr" rid="B26">Liu et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B3">Chen et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B24">Li J. et&#xa0;al., 2024</xref>). LMP-PM can facilitate the lightweighting of multimodal backbone networks while providing multiple path models for various input modalities, achieving an organic integration with multimodal tasks.</p>
</sec>
</body>
<back>
<sec id="s5" sec-type="data-availability">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author.</p></sec>
<sec id="s6" sec-type="author-contributions">
<title>Author contributions</title>
<p>JH: Funding acquisition, Project administration, Resources, Supervision, Validation, Writing &#x2013; review &amp; editing. FZ: Conceptualization, Data curation, Formal Analysis, Methodology, Software, Writing &#x2013; original draft, Writing &#x2013; review &amp; editing. YZ: Data curation, Investigation, Software, Validation, Writing &#x2013; review &amp; editing. JD: Formal Analysis, Investigation, Supervision, Validation, Writing &#x2013; review &amp; editing. RH: Investigation, Software, Supervision, Validation, Writing &#x2013; review &amp; editing.</p></sec>
<sec id="s8" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
<sec id="s9" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec id="s10" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Bhagat</surname> <given-names>S.</given-names></name>
<name><surname>Kokare</surname> <given-names>M.</given-names></name>
<name><surname>Haswani</surname> <given-names>V.</given-names></name>
<name><surname>Hambarde</surname> <given-names>P.</given-names></name>
<name><surname>Taori</surname> <given-names>T.</given-names></name>
<name><surname>Ghante</surname> <given-names>P. H.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Advancing real-time plant disease detection: A lightweight deep learning approach and novel dataset for pigeon pea crop</article-title>. <source>Smart. Agric. Technol.</source> <volume>7</volume>, <fpage>100408</fpage>. ISSN 2772-3755. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.atech.2024.100408</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Cai</surname> <given-names>X.</given-names></name>
<name><surname>Zhu</surname> <given-names>Y.</given-names></name>
<name><surname>Liu</surname> <given-names>S.</given-names></name>
<name><surname>Yu</surname> <given-names>Z.</given-names></name>
<name><surname>Xu</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>FastSegFormer: A knowledge distillation-based method for real-time semantic segmentation of surface defects in navel oranges</article-title>. <source>Comput. Electron. Agric.</source> <volume>217</volume>, <fpage>108604</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2023.108604</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Chen</surname> <given-names>W.</given-names></name>
<name><surname>Rao</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>F.</given-names></name>
<name><surname>Zhang</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>T.</given-names></name>
<name><surname>Jin</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>MLP-based multimodal tomato detection in complex scenarios: Insights from task-specific analysis of feature fusion architectures</article-title>. <source>Comput. Electron. Agric.</source> <volume>221</volume>, <fpage>108951</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.108951</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Cheng</surname> <given-names>G.</given-names></name>
<name><surname>Lai</surname> <given-names>P.</given-names></name>
<name><surname>Gao</surname> <given-names>D.</given-names></name>
<name><surname>Han</surname> <given-names>J.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Class attention network for image recognition</article-title>. <source>Sci. China Inf. Sci.</source> <volume>66</volume>, <fpage>132105</fpage>. 
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Dai</surname> <given-names>W.</given-names></name>
<name><surname>Zhu</surname> <given-names>W.</given-names></name>
<name><surname>Zhou</surname> <given-names>G.</given-names></name>
<name><surname>Liu</surname> <given-names>G.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<name><surname>Zhou</surname> <given-names>H.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>AISOA-SSformer: an effective image segmentation method for rice leaf disease based on the transformer architecture</article-title>. <source>Plant Phenom.</source> <volume>6</volume>, <fpage>0218</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0218</pub-id>, PMID: <pub-id pub-id-type="pmid">39105185</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Denarda</surname> <given-names>A. R.</given-names></name>
<name><surname>Crocetti</surname> <given-names>F.</given-names></name>
<name><surname>Costante</surname> <given-names>G.</given-names></name>
<name><surname>Valigi</surname> <given-names>P.</given-names></name>
<name><surname>Fravolini</surname> <given-names>M. L</given-names></name>
</person-group>. (<year>2024</year>). 
<article-title>MangoDetNet: a novel label-efficient weakly supervised fruit detection framework</article-title>. <source>Precis. Agric</source>. <volume>25</volume>(<issue>6</issue>), <fpage>3167</fpage>&#x2013;<lpage>3188</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-024-10187-0</pub-id>, PMID: <pub-id pub-id-type="pmid">41728210</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Deng</surname> <given-names>J.</given-names></name>
<name><surname>Huang</surname> <given-names>W.</given-names></name>
<name><surname>Zhou</surname> <given-names>G.</given-names></name>
<name><surname>Hu</surname> <given-names>Y.</given-names></name>
<name><surname>Li</surname> <given-names>L.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Identification of banana leaf disease based on KVA and GR-ARNet</article-title>. <source>J. Integr. Agric.</source> <volume>23</volume>, <fpage>3554</fpage>&#x2013;<lpage>3575</lpage>. ISSN 2095-3119. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jia.2023.11.037</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Dong</surname> <given-names>X.</given-names></name>
<name><surname>Zhao</surname> <given-names>K.</given-names></name>
<name><surname>Wang</surname> <given-names>Q.</given-names></name>
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<name><surname>Huang</surname> <given-names>Y.</given-names></name>
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>PlantPAD: A platform for large-scale image phenomics analysis of disease in plant science</article-title>. <source>Nucleic Acids Res.</source> <volume>52</volume>, <fpage>D1556</fpage>&#x2013;<lpage>D1568</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/nar/gkad917</pub-id>, PMID: <pub-id pub-id-type="pmid">37897364</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Han</surname> <given-names>K.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Tian</surname> <given-names>Q.</given-names></name>
<name><surname>Guo</surname> <given-names>J.</given-names></name>
<name><surname>Xu</surname> <given-names>C.</given-names></name>
<name><surname>Xu</surname> <given-names>C.</given-names></name>
</person-group> (<year>2020</year>). &#x201c;
<article-title>Ghostnet: More features from cheap operations</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF conference on computer vision and pattern recognition</conf-name>. <fpage>1580</fpage>&#x2013;<lpage>1589</lpage>.
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>He</surname> <given-names>K.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Ren</surname> <given-names>S.</given-names></name>
<name><surname>Sun</surname> <given-names>J</given-names></name>
</person-group>. (<year>2016</year>). &#x201c;
<article-title>Deep residual learning for image recognition</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>. <fpage>770</fpage>&#x2013;<lpage>778</lpage>.
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Heidari</surname> <given-names>M.</given-names></name>
<name><surname>Kazerouni</surname> <given-names>A.</given-names></name>
<name><surname>Soltany</surname> <given-names>M.</given-names></name>
<name><surname>Azad</surname> <given-names>R.</given-names></name>
<name><surname>Aghdam</surname> <given-names>E. K.</given-names></name>
<name><surname>Cohen-Adad</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). &#x201c;
<article-title>Hiformer: Hierarchical multiscale representations using trans-formers for medical image segmentation</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision</conf-name>, <conf-loc>Waikoloa, HI, USA</conf-loc>, <conf-date>2023 Jan 2&#x2013;7</conf-date>.
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hern&#xe1;ndez</surname> <given-names>I.</given-names></name>
<name><surname>Guti&#xe9;rrez</surname> <given-names>S.</given-names></name>
<name><surname>Barrio</surname> <given-names>I.</given-names></name>
<name><surname>&#xcd;&#xf1;iguez</surname> <given-names>R.</given-names></name>
<name><surname>Tardaguila</surname> <given-names>J.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>In-field disease symptom detection and localisation using explainable deep learning: Use case for downy mildew in grapevine</article-title>. <source>Comput. Electron. Agric.</source> <volume>226</volume>, <fpage>109478</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109478</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Howard</surname> <given-names>A. G.</given-names></name>
<name><surname>Zhu</surname> <given-names>M.</given-names></name>
<name><surname>Chen</surname> <given-names>B.</given-names></name>
<name><surname>Kalenichenko</surname> <given-names>D.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>Weyand</surname> <given-names>T.</given-names></name>
<etal/>
</person-group>. (<year>2017</year>). <source>&#x201c;MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.&#x201d; ArXiv abs/1704.04861</source>, n. pag.
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Howard</surname> <given-names>A.</given-names></name>
<name><surname>Sandler</surname> <given-names>M.</given-names></name>
<name><surname>Chu</surname> <given-names>G.</given-names></name>
<name><surname>Chen</surname> <given-names>L. C.</given-names></name>
<name><surname>Chen</surname> <given-names>B.</given-names></name>
<name><surname>Tan</surname> <given-names>M.</given-names></name>
<etal/>
</person-group>. (<year>2019</year>). &#x201c;
<article-title>Searching for mobileNetV3</article-title>,&#x201d; in <conf-name>2019 IEEE/CVF International Conference on Computer Vision (ICCV)</conf-name>, <conf-loc>Seoul, Korea (South</conf-loc>. <fpage>1314</fpage>&#x2013;<lpage>1324</lpage>.
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Hu</surname> <given-names>J.</given-names></name>
<name><surname>Shen</surname> <given-names>L.</given-names></name>
<name><surname>Sun</surname> <given-names>G.</given-names></name>
</person-group> (<year>2018</year>). &#x201c;
<article-title>Squeeze-and-excitation networks</article-title>,&#x201d; in <conf-name>2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>, <conf-loc>Salt Lake City, UT, USA</conf-loc>. <fpage>7132</fpage>&#x2013;<lpage>7141</lpage>.
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hua</surname> <given-names>J.</given-names></name>
<name><surname>Zhu</surname> <given-names>T.</given-names></name>
<name><surname>Zou</surname> <given-names>F.</given-names></name>
<name><surname>Zou</surname> <given-names>J.</given-names></name>
<name><surname>Tang</surname> <given-names>J.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>CSLSNet: A compressed domain classification model for pest and disease images</article-title>. <source>Agronomy</source> <volume>13</volume>, <fpage>2663</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy13102663</pub-id>, PMID: <pub-id pub-id-type="pmid">41725453</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Huang</surname> <given-names>Q.</given-names></name>
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<name><surname>Wang</surname> <given-names>Q.</given-names></name>
<name><surname>Dong</surname> <given-names>X.</given-names></name>
<name><surname>Qin</surname> <given-names>Y.</given-names></name>
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>Knowledge distillation facilitates the lightweight and efficient plant diseases detection model</article-title>. <source>Plant Phenom.</source> <volume>5</volume>, <fpage>0062</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0062</pub-id>, PMID: <pub-id pub-id-type="pmid">37396495</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hughes</surname> <given-names>D. P.</given-names></name>
<name><surname>Marcel</surname> <given-names>S.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>An open access repository of images on plant health to enable the development of mobile disease diagnostics</article-title>. <source>arXiv preprint arXiv:1511.08060</source>.
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Jain</surname> <given-names>J.</given-names></name>
<name><surname>Li</surname> <given-names>J.</given-names></name>
<name><surname>Chiu</surname> <given-names>M. T.</given-names></name>
<name><surname>Hassani</surname> <given-names>A.</given-names></name>
<name><surname>Orlov</surname> <given-names>N.</given-names></name>
<name><surname>Shi</surname> <given-names>H</given-names></name>
</person-group>. &#x201c;
<article-title>Oneformer: One transformer to rule universal image segmentation</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</conf-name>, <conf-loc>Vancouver, BC, Canada</conf-loc>, <conf-date>2023 Jun 17&#x2013;24</conf-date>.
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Khanna</surname> <given-names>M.</given-names></name>
<name><surname>Singh</surname> <given-names>L. K.</given-names></name>
<name><surname>Thawkar</surname> <given-names>S.</given-names></name>
<name><surname>Goyal</surname> <given-names>M.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Planet: A robust deep convolutional neural network model for plant leaves disease recognition</article-title>. <source>Multimed. Tools Appl.</source> <volume>83</volume>, <fpage>4465</fpage>&#x2013;<lpage>4517</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11042-023-15809-9</pub-id>, PMID: <pub-id pub-id-type="pmid">41728210</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Kuska</surname> <given-names>M. T.</given-names></name>
<name><surname>Wahabzada</surname> <given-names>M.</given-names></name>
<name><surname>Paulus</surname> <given-names>S.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>AI for crop production &#x2013; Where can large language models (LLMs) provide substantial value</article-title>? <source>Comput. Electron. Agric.</source> <volume>221</volume>, <fpage>108924</fpage>. ISSN 0168-1699.
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>H.</given-names></name>
<name><surname>Huang</surname> <given-names>L.</given-names></name>
<name><surname>Ruan</surname> <given-names>C.</given-names></name>
<name><surname>Huang</surname> <given-names>W.</given-names></name>
<name><surname>Wang</surname> <given-names>C.</given-names></name>
<name><surname>Zhao</surname> <given-names>J.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>A dual-branch neural network for crop disease recognition by integrating frequency domain and spatial domain information</article-title>. <source>Comput. Electron. Agric.</source> <volume>219</volume>, <fpage>108843</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.108843</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>R.</given-names></name>
<name><surname>Li</surname> <given-names>Y.</given-names></name>
<name><surname>Qin</surname> <given-names>W.</given-names></name>
<name><surname>Abbas</surname> <given-names>A.</given-names></name>
<name><surname>Li</surname> <given-names>S.</given-names></name>
<name><surname>Ji</surname> <given-names>R.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Lightweight network for corn leaf disease identification based on improved YOLO v8s</article-title>. <source>Agriculture</source> <volume>14</volume>, <fpage>220</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agriculture14020220</pub-id>, PMID: <pub-id pub-id-type="pmid">41725453</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>J.</given-names></name>
<name><surname>Zhao</surname> <given-names>F.</given-names></name>
<name><surname>Zhao</surname> <given-names>H.</given-names></name>
<name><surname>Zhou</surname> <given-names>G.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<name><surname>Gao</surname> <given-names>M.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>A multi-modal open object detection model for tomato leaf diseases with strong generalization performance using PDC-VLD</article-title>. <source>Plant Phenom.</source> <volume>6</volume>, <fpage>0220</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0220</pub-id>, PMID: <pub-id pub-id-type="pmid">39139386</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lin</surname> <given-names>J.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Qin</surname> <given-names>Y.</given-names></name>
<name><surname>Yang</surname> <given-names>S.</given-names></name>
<name><surname>Wen</surname> <given-names>X.</given-names></name>
<name><surname>Cernava</surname> <given-names>T.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Local and global feature-aware dual-branch networks for plant disease recognition</article-title>. <source>Plant Phenom.</source> <volume>6</volume>, <fpage>0208</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0208</pub-id>, PMID: <pub-id pub-id-type="pmid">39130161</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Liu</surname> <given-names>C.</given-names></name>
<name><surname>Feng</surname> <given-names>Q.</given-names></name>
<name><surname>Sun</surname> <given-names>Y.</given-names></name>
<name><surname>Li</surname> <given-names>Y.</given-names></name>
<name><surname>Ru</surname> <given-names>M.</given-names></name>
<name><surname>Xu</surname> <given-names>L.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>YOLACTFusion: An instance segmentation method for RGB-NIR multimodal image fusion based on an attention mechanism</article-title>. <source>Comput. Electron. Agric.</source> <volume>213</volume>, <fpage>108186</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2023.108186</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Mao</surname> <given-names>H.</given-names></name>
<name><surname>Wu</surname> <given-names>C.-Y.</given-names></name>
<name><surname>Feichtenhofer</surname> <given-names>C.</given-names></name>
<name><surname>Darrell</surname> <given-names>T.</given-names></name>
<name><surname>Xie</surname> <given-names>S.</given-names></name>
</person-group> (<year>2022</year>). &#x201c;
<article-title>A convNet for the 2020s</article-title>,&#x201d; in <conf-name>2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>New Orleans, LA, USA</conf-loc>. <fpage>11966</fpage>&#x2013;<lpage>11976</lpage>.
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Loshchilov</surname> <given-names>I.</given-names></name>
<name><surname>Hutter</surname> <given-names>F.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>Decoupled weight decay regularization</article-title>,&#x201d; in <conf-name>International Conference on Learning Representations</conf-name>.
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Ma</surname> <given-names>X.</given-names></name>
<name><surname>Dai</surname> <given-names>X.</given-names></name>
<name><surname>Bai</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Fu</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2024</year>). &#x201c;
<article-title>Rewrite the stars</article-title>,&#x201d; in <conf-name>2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Seattle, WA, USA</conf-loc>. <fpage>5694</fpage>&#x2013;<lpage>5703</lpage>.
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Ma</surname> <given-names>N.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Zheng</surname> <given-names>H. T.</given-names></name>
<name><surname>Sun</surname> <given-names>J</given-names></name>
</person-group>. (<year>2018</year>). &#x201c;
<article-title>Shufflenet v2: Practical guidelines for efficient cnn architecture design</article-title>,&#x201d; in <conf-name>Proceedings of the European conference on computer vision (ECCV)</conf-name>.
</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Maaz</surname> <given-names>M.</given-names></name>
<name><surname>Shaker</surname> <given-names>A.</given-names></name>
<name><surname>Cholakkal</surname> <given-names>H.</given-names></name>
<name><surname>Khan</surname> <given-names>S.</given-names></name>
<name><surname>Zamir</surname> <given-names>S. W.</given-names></name>
<name><surname>Anwer</surname> <given-names>R. M.</given-names></name>
<etal/>
</person-group>. (<year>2022</year>). &#x201c;
<article-title>Edgenext: efficiently amalgamated cnn-transformer architecture for mobile vision applications</article-title>,&#x201d; in <conf-name>European conference on computer vision</conf-name>, <conf-loc>Cham</conf-loc>. <fpage>3</fpage>&#x2013;<lpage>20</lpage> (<publisher-loc>Cham</publisher-loc>: 
<publisher-name>Springer Nature Switzerland</publisher-name>).
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Pal</surname> <given-names>A.</given-names></name>
<name><surname>Kumar</surname> <given-names>V.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Agridet: Plant leaf disease severity classification using agriculture detection framework</article-title>. <source>Eng. Appl. Artif. Intell.</source> <volume>119</volume>, <fpage>105754</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.engappai.2022.105754</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Parthiban</surname> <given-names>S.</given-names></name>
<name><surname>Moorthy</surname> <given-names>S.</given-names></name>
<name><surname>Sabanayagam</surname> <given-names>S.</given-names></name>
<name><surname>Shanmugasundaram</surname> <given-names>S.</given-names></name>
<name><surname>Naganathan</surname> <given-names>A.</given-names></name>
<name><surname>Annamalai</surname> <given-names>M.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). &#x201c;
<article-title>Deep learning based recognition of plant diseases</article-title>,&#x201d; in <conf-name>Computer Vision and Machine Intelligence Paradigms for SDGs: Select Proceedings of ICRTAC-CVMIP 2021</conf-name>. <fpage>83</fpage>&#x2013;<lpage>93</lpage> (<publisher-loc>Singapore</publisher-loc>: 
<publisher-name>Springer</publisher-name>).
</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Quan</surname> <given-names>S.</given-names></name>
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<name><surname>Jia</surname> <given-names>Z.</given-names></name>
<name><surname>Xu</surname> <given-names>Q.</given-names></name>
<name><surname>Yang</surname> <given-names>M.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Real-time field disease identification based on a lightweight model</article-title>. <source>Comput. Electron. Agric.</source> <volume>226</volume>, <fpage>109467</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109467</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sahu</surname> <given-names>S. K.</given-names></name>
<name><surname>Pandey</surname> <given-names>M.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>An optimal hybrid multiclass svm for plant leaf disease detection using spatial fuzzy c-means model</article-title>. <source>Expert Syst. Appl.</source> <volume>214</volume>, <fpage>118989</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.eswa.2022.118989</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Salamai</surname> <given-names>A. A.</given-names></name>
<name><surname>Ajabnoor</surname> <given-names>N.</given-names></name>
<name><surname>Khalid</surname> <given-names>W. E.</given-names></name>
<name><surname>Ali</surname> <given-names>M. M.</given-names></name>
<name><surname>Murayr</surname> <given-names>A. A.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Lesion-aware visual transformer network for paddy diseases detection in precision agriculture</article-title>. <source>Eur. J. Agron.</source> <volume>148</volume>, <fpage>126884</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.eja.2023.126884</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Selvaraju</surname> <given-names>R. R.</given-names></name>
<name><surname>Cogswell</surname> <given-names>M.</given-names></name>
<name><surname>Das</surname> <given-names>A.</given-names></name>
<name><surname>Vedantam</surname> <given-names>R.</given-names></name>
<name><surname>Parikh</surname> <given-names>D.</given-names></name>
<name><surname>Batra</surname> <given-names>D</given-names></name>
</person-group>. &#x201c;
<article-title>Grad-cam: Visual explanations from deep networks via gradient-based localization</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE International Conference on Computer Vision</conf-name>, <conf-loc>Venice, Italy</conf-loc>, <conf-date>2017 Oct 22&#x2013;29</conf-date>.
</mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Shi</surname> <given-names>T.</given-names></name>
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Zheng</surname> <given-names>X.</given-names></name>
<name><surname>Hu</surname> <given-names>K.</given-names></name>
<name><surname>Huang</surname> <given-names>H.</given-names></name>
<name><surname>Liu</surname> <given-names>H.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>Recent advances in plant disease severity assessment using convolutional neural networks</article-title>. <source>Sci. Rep.</source> <volume>13</volume>, <fpage>2336</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-023-29230-7</pub-id>, PMID: <pub-id pub-id-type="pmid">36759626</pub-id>
</mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Swaminathan</surname> <given-names>B.</given-names></name>
<name><surname>Vairavasundaram</surname> <given-names>S.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>D2CNN: Double-staged deep CNN for stress identification and classification in cropping system</article-title>. <source>Agric. Syst.</source> <volume>216</volume>, <fpage>103886</fpage>. ISSN 0308-521X.
</mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Szegedy</surname> <given-names>C.</given-names></name>
<name><surname>Liu</surname> <given-names>W.</given-names></name>
<name><surname>Jia</surname> <given-names>Y.</given-names></name>
<name><surname>Sermanet</surname> <given-names>P.</given-names></name>
<name><surname>Reed</surname> <given-names>S.</given-names></name>
<name><surname>Anguelov</surname> <given-names>D.</given-names></name>
<etal/>
</person-group>. (<year>2015</year>). &#x201c;
<article-title>Going deeper with convolutions</article-title>,&#x201d; in <conf-name>2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Boston, MA, USA</conf-loc>. <fpage>1</fpage>&#x2013;<lpage>9</lpage>.
</mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Tan</surname> <given-names>M.</given-names></name>
<name><surname>Le</surname> <given-names>Q.</given-names></name>
</person-group> (<year>2019</year>). &#x201c;
<article-title>Efficientnet: Rethinking model scaling for convolutional neural networks</article-title>,&#x201d; in <conf-name>International conference on machine learning</conf-name>. <fpage>6105</fpage>&#x2013;<lpage>6114</lpage> (
<publisher-name>PMLR</publisher-name>).
</mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Tang</surname> <given-names>W.</given-names></name>
<name><surname>Wen</surname> <given-names>X.</given-names></name>
<name><surname>Li</surname> <given-names>M.</given-names></name>
<name><surname>Chen</surname> <given-names>Y.</given-names></name>
<name><surname>Hu</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>ResiAdvNet: A named entity recognition model for potato diseases and pests based on progressive residual structures and adversarial training</article-title>. <source>Comput. Electron. Agric.</source> <volume>227</volume>, <fpage>109543</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109543</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Tang</surname> <given-names>L.</given-names></name>
<name><surname>Yi</surname> <given-names>J.</given-names></name>
<name><surname>Li</surname> <given-names>X.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Improved multi-scale inverse bottleneck residual network based on triplet parallel attention for apple leaf disease identification</article-title>. <source>J. Integr. Agric.</source> <volume>23</volume>, <fpage>901</fpage>&#x2013;<lpage>922</lpage>. ISSN 2095-3119. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jia.2023.06.023</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>A.</given-names></name>
<name><surname>Chen</surname> <given-names>H.</given-names></name>
<name><surname>Lin</surname> <given-names>Z.</given-names></name>
<name><surname>Han</surname> <given-names>J.</given-names></name>
<name><surname>Ding</surname> <given-names>G.</given-names></name>
</person-group> (<year>2024</year>). &#x201c;
<article-title>Rep viT: revisiting mobile CNN from viT perspective</article-title>,&#x201d; in <conf-name>2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Seattle, WA, USA</conf-loc>. <fpage>15909</fpage>&#x2013;<lpage>15920</lpage>.
</mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<name><surname>Lin</surname> <given-names>X.</given-names></name>
<name><surname>Luo</surname> <given-names>L.</given-names></name>
<name><surname>Chen</surname> <given-names>M.</given-names></name>
<name><surname>Wei</surname> <given-names>H.</given-names></name>
<name><surname>Xu</surname> <given-names>L.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Cognition of grape cluster picking point based on visual knowledge distillation in complex vineyard environment</article-title>. <source>Comput. Electron. Agric.</source> <volume>225</volume>, <fpage>109216</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109216</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>H.</given-names></name>
<name><surname>Pan</surname> <given-names>X.</given-names></name>
<name><surname>Zhu</surname> <given-names>Y.</given-names></name>
<name><surname>Li</surname> <given-names>S.</given-names></name>
<name><surname>Zhu</surname> <given-names>R.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Maize leaf disease recognition based on TC-MRSN model in sustainable agriculture</article-title>. <source>Comput. Electron. Agric.</source> <volume>221</volume>, <fpage>108915</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.108915</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>Xie</surname> <given-names>E.</given-names></name>
<name><surname>Li</surname> <given-names>X.</given-names></name>
<name><surname>Fan</surname> <given-names>D. P.</given-names></name>
<name><surname>Song</surname> <given-names>K.</given-names></name>
<name><surname>Liang</surname> <given-names>D.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). &#x201c;
<article-title>Pyramid vision transformer: A versatile backbone for dense prediction without convolutions</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE/CVF international conference on computer vision</conf-name>. <fpage>568</fpage>&#x2013;<lpage>578</lpage>.
</mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>B.</given-names></name>
<name><surname>Yang</surname> <given-names>H.</given-names></name>
<name><surname>Zhang</surname> <given-names>S.</given-names></name>
<name><surname>Li</surname> <given-names>L.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Identification of multiple diseases in apple leaf based on optimized lightweight convolutional neural network</article-title>. <source>Plants</source> <volume>13</volume>, <fpage>1535</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/plants13111535</pub-id>, PMID: <pub-id pub-id-type="pmid">38891344</pub-id>
</mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<name><surname>Deng</surname> <given-names>H.</given-names></name>
<name><surname>Wang</surname> <given-names>Q.</given-names></name>
<name><surname>Lei</surname> <given-names>L.</given-names></name>
<name><surname>Gao</surname> <given-names>Y.</given-names></name>
<name><surname>Hao</surname> <given-names>G.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Meta-learning shows great potential in plant disease recognition under few available samples</article-title>. <source>Plant J.</source> <volume>114</volume>, <fpage>767</fpage>&#x2013;<lpage>782</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1111/tpj.16176</pub-id>, PMID: <pub-id pub-id-type="pmid">36883481</pub-id>
</mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Wu</surname> <given-names>J.</given-names></name>
<name><surname>Zheng</surname> <given-names>H.</given-names></name>
<name><surname>Zhao</surname> <given-names>B.</given-names></name>
<name><surname>Li</surname> <given-names>Y.</given-names></name>
<name><surname>Yan</surname> <given-names>B.</given-names></name>
<name><surname>Liang</surname> <given-names>R.</given-names></name>
<etal/>
</person-group>. &#x201c;
<article-title>Large-scale datasets for going deeper in image understanding</article-title>,&#x201d; in <conf-name>2019 IEEE International Conference on Multimedia and Expo (ICME)</conf-name>.
</mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Xiao</surname> <given-names>S.</given-names></name>
<name><surname>Chen</surname> <given-names>H.</given-names></name>
<name><surname>Hou</surname> <given-names>Y.</given-names></name>
<name><surname>Shao</surname> <given-names>K.</given-names></name>
<name><surname>Bi</surname> <given-names>K.</given-names></name>
<name><surname>Wang</surname> <given-names>R.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Investigating the 3D distribution of Cercospora leaf spot disease in sugar beet through fusion methods</article-title>. <source>Comput. Electron. Agric.</source> <volume>224</volume>, <fpage>109107</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109107</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Xie</surname> <given-names>Z.</given-names></name>
<name><surname>Li</surname> <given-names>C.</given-names></name>
<name><surname>Yang</surname> <given-names>Z.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<name><surname>Jiang</surname> <given-names>J.</given-names></name>
<name><surname>Guo</surname> <given-names>H.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>YOLOv5s-biPCNeXt, a lightweight model for detecting disease in eggplant leaves</article-title>. <source>Plants</source> <volume>13</volume>, <fpage>2303</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/plants13162303</pub-id>, PMID: <pub-id pub-id-type="pmid">39204740</pub-id>
</mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Xing</surname> <given-names>Z.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Xu</surname> <given-names>P.</given-names></name>
<name><surname>Guo</surname> <given-names>Q.</given-names></name>
<name><surname>Zeng</surname> <given-names>C.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>SDC-deepLabv3+: lightweight and precise localization algorithm for safflower-harvesting robots</article-title>. <source>Plant Phenom.</source> <volume>6</volume>, <fpage>0194</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0194</pub-id>, PMID: <pub-id pub-id-type="pmid">38974378</pub-id>
</mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yongda</surname> <given-names>L.</given-names></name>
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Chen</surname> <given-names>T.</given-names></name>
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>L.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Monitoring system for peanut leaf disease based on a lightweight deep learning model</article-title>. <source>Comput. Electron. Agric.</source> <volume>222</volume>, <fpage>109055</fpage>. ISSN 0168-1699.
</mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yu</surname> <given-names>H.</given-names></name>
<name><surname>Weng</surname> <given-names>L.</given-names></name>
<name><surname>Wu</surname> <given-names>S.</given-names></name>
<name><surname>He</surname> <given-names>J.</given-names></name>
<name><surname>Yuan</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Time-series field phenotyping of soybean growth analysis by combining multimodal deep learning and dynamic modeling</article-title>. <source>Plant Phenom.</source> <volume>6</volume>, <fpage>0158</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0158</pub-id>, PMID: <pub-id pub-id-type="pmid">38524738</pub-id>
</mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yu</surname> <given-names>X.</given-names></name>
<name><surname>Yin</surname> <given-names>D.</given-names></name>
<name><surname>Xu</surname> <given-names>H.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Maize tassel number and tasseling stage monitoring based on near-ground and UAV RGB images by improved YoloV8</article-title>. <source>Precis. Agric.</source> <volume>25</volume>, <fpage>1800</fpage>&#x2013;<lpage>1838</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-024-10135-y</pub-id>, PMID: <pub-id pub-id-type="pmid">41728210</pub-id>
</mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>T.</given-names></name>
<name><surname>Hu</surname> <given-names>D.</given-names></name>
<name><surname>Wu</surname> <given-names>C.</given-names></name>
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Yang</surname> <given-names>J.</given-names></name>
<name><surname>Tang</surname> <given-names>K.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Large-scale apple orchard mapping from multi-source data using the semantic segmentation model with image- to- image translation and transfer learning</article-title>. <source>Comput. Electron. Agric.</source> <volume>213</volume>, <fpage>108204</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2023.108204</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>J.</given-names></name>
<name><surname>Shen</surname> <given-names>D.</given-names></name>
<name><surname>Chen</surname> <given-names>D.</given-names></name>
<name><surname>Ming</surname> <given-names>D.</given-names></name>
<name><surname>Ren</surname> <given-names>D.</given-names></name>
<name><surname>Diao</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>ISMSFuse: Multi-modal fusing recognition algorithm for rice bacterial blight disease adaptable in edge computing scenarios</article-title>. <source>Comput. Electron. Agric.</source> <volume>223</volume>, <fpage>109089</fpage>. ISSN 0168-1699. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109089</pub-id>, PMID: <pub-id pub-id-type="pmid">41727822</pub-id>
</mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>P.</given-names></name>
<name><surname>Sun</surname> <given-names>X.</given-names></name>
<name><surname>Zhang</surname> <given-names>D.</given-names></name>
<name><surname>Yang</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Lightweight deep learning models for high-precision rice seedling segmentation from UAV-based multispectral images</article-title>. <source>Plant Phenom.</source> <volume>5</volume>, <fpage>0123</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0123</pub-id>, PMID: <pub-id pub-id-type="pmid">38047001</pub-id>
</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3152000">Sathishkumar Samiappan</ext-link>, The University of Tennessee, United States</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2727168">Cevher &#xd6;zden</ext-link>, &#xc7;ukurova University, T&#xfc;rkiye</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3176208">Mohtar Yunianto</ext-link>, Universitas Sebelas Maret, Indonesia</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3176502">Elham Mohammed Thabit A. Alsaadi</ext-link>, University of Karbala, Iraq</p></fn>
</fn-group>
</back>
</article>