<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Sustain. Food Syst.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Sustainable Food Systems</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Sustain. Food Syst.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2571-581X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fsufs.2025.1734639</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>GET-YOLO: a lightweight deep learning framework for real-time pest detection in marginal tea plantations to support green and sustainable tea production</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Yu</surname><given-names>Chao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Bao</surname><given-names>Qiang</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn0001"><sup>&#x2020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Wang</surname><given-names>Yuanjiang</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Ni</surname><given-names>Xiao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Xiao</surname><given-names>Lei</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Zhou</surname><given-names>Chao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3258757"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &#x0026; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x0026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Hunan Academy of Agricultural Sciences</institution>, <city>Changsha</city>, <country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Yuelushan Laboratory</institution>, <city>Changsha</city>, <country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Chao Zhou, <email xlink:href="mailto:chaozhou@hunaas.cn">chaozhou@hunaas.cn</email></corresp>
<fn fn-type="equal" id="fn0001"><label>&#x2020;</label><p>These authors share first authorship</p></fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-21">
<day>21</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>9</volume>
<elocation-id>1734639</elocation-id>
<history>
<date date-type="received">
<day>29</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>08</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>15</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Yu, Bao, Wang, Ni, Xiao and Zhou.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Yu, Bao, Wang, Ni, Xiao and Zhou</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-21">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The marginal land area in Hunan Province was about 638,800 hectares. Due to the limitation of acid sticky red soil and steep slope, the low grain production capacity was suitable for tea planting. However, outbreak pests such as <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) and other outbreak pests seriously threatened the yield and quality of tea, and the traditional manual inspection and chemical control were difficult to meet the needs of green and high-quality tea industry. Traditional manual inspection and chemical control methods were increasingly inadequate, which would lead to ecological risks (e.g., soil and water pollution) and fail to meet the demands of sustainable agroecosystem management for the green and high-quality tea industry. In this study, a GET-YOLO recognition method for marginal tea garden was proposed, which was based on YOLO11m depth model, embedding ECA attention mechanism to enhance key feature extraction, introducing GhostConv to achieve model lightweight, and integrating transfer learning to improve small sample adaptability. The precision, recall and <italic>mAP</italic><sub>50</sub> of the improved model were improved by 0.42%,4.36% and 1.83%, respectively. The <italic>mAP</italic><sub>50</sub> was up to 87.94%, the number of parameters was reduced by 28%, and the inference time of a single image was shortened to 23&#x202F;ms, which achieved the balance of &#x201C;high precision, lightweight and high real-time.&#x201D; The model could automatically obtain the field growth and decline dynamics of Basilepta theobromae, provide real-time data support and key basis for the green prevention and control of diseases and insect pests in marginal tea gardens and the identification of other pests in tea plants, and has important theoretical and practical value for promoting the green and efficient upgrading of tea industry in marginal land.</p>
</abstract>
<kwd-group>
<kwd>green prevention and control</kwd>
<kwd>identification method</kwd>
<kwd>insect pest</kwd>
<kwd>intelligent identification</kwd>
<kwd>marginal land</kwd>
<kwd>tea</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This research was supported by Key Research and Development Program of Hunan Province of China (2023NK2013), Agricultural Science and Technology Innovation Fund Project of Hunan Province (2024CX57) and National Key Research and Development Program of China (2022YFD1600803).</funding-statement>
</funding-group>
<counts>
<fig-count count="7"/>
<table-count count="3"/>
<equation-count count="4"/>
<ref-count count="29"/>
<page-count count="11"/>
<word-count count="8031"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Agroecology and Ecosystem Services</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>Marginal land is a special category of agricultural land characterized by low productivity and ecological fragility due to soil constraints, water and heat limitations, and topographical restrictions (<xref ref-type="bibr" rid="ref1">Cao et al., 2021</xref>). Compared to conventional farmland, its core differences are showed in inherently poor soil quality (as exemplified by the southern lowland red soils dominated by marginal land in Hunan, which exhibit strong acidity, low organic matter, nutrient deficiency, and high clay content), complex topography (9.35% of the province&#x2019;s arable land lies on slopes steeper than 15&#x00B0;, concentrated in remote mountainous areas of central-southern, western, and southern Hunan with scattered plots), and fragile ecosystems (strong geographical isolation and weak natural regulatory capacity in remote regions). Although unsuitable for grain cultivation, 638,800 hectares of marginal land has become the optimal path for efficient utilization in Hunan. The tea industry in Hunan has developed mature cultivation models and established well-known brands, laying the foundation for large-scale development of marginal tea plantations.</p>
<p>However, the fundamental differences are result in significantly higher pest risks in marginal tea plantations between marginal land and conventional soil directly, which become a core bottleneck for sustainable green tea production. Firstly, tea plants exhibits weak stress tolerance. The poor fertility of red soil leads to stunted growth, making them far less resistant to pests and diseases compared to high-quality farmland plantations. Secondly, monitoring is extremely challenging. The rugged mountainous terrain and poor transport infrastructure hinder traditional manual inspections, which is failing to cover scattered plots and mask early symptoms of pests, resulting in delayed control measures. Thirdly, ecological pest management capabilities are insufficient. The surrounding ecosystems of marginal tea plantations remains unstable, with limited natural pest control due to scarce natural predators. Finally, the risk of a vicious cycle is emerging. During pest outbreak events, farmers&#x2019; concerns regarding yield reduction drived a substantial increase in the application of chemical pesticides. This practice not only exacerbated the risks associated with pesticide residues but also further impaired the fragile soil conditions and ecological systems of marginal lands. A self-reinforcing cycle of &#x201C;pest outbreaks&#x2014;excessive pesticide use&#x2014;soil degradation&#x2014;weaker tea plant stress tolerance&#x201D; would be created. Currently, the tea plantations are facing 58 types of annual pests in Hunan, with the <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) being the primary threat (<xref ref-type="bibr" rid="ref13">Ou et al., 2018</xref>). The tea plantations is average 5.5&#x2013;6.0 pesticide applications/year, using about 6,000 grams/ha of chemical pesticides. The unique conditions of marginal plantations exacerbate this issue, making precise pest monitoring and green control measures an urgent necessity for sustainable development.</p>
<p>The existing technical framework struggle to adapt to the unique challenges of marginal tea plantations. Inadequate supply of non-chemical control technologies and incomplete prevention systems has led to excessive reliance on chemical pesticides among tea farmers. Traditional manual inspections are characterized by low efficiency and limited coverage, which fail to meet the real-time monitoring demands of these complex terrains. With advancements in deep learning and computer vision technologies, significant progress has been made in crop pest detection. <xref ref-type="bibr" rid="ref29">Zhu et al. (2025)</xref> systematically reviewed cutting-edge developments in crop leaf disease detection using CNN (convolutional neural network) architectures, with particular emphasis on YOLO-based detection methods. <xref ref-type="bibr" rid="ref2">Cao et al. (2025)</xref> conducted comprehensive analyses of various YOLO algorithm versions, evaluating their practical applications in crop recognition and pest identification. <xref ref-type="bibr" rid="ref16">Su et al. (2025)</xref> developed the Pest-YOLOv8 detection model, enhancing precise pest localization through Triple-CA modules and Wise-IoU loss functions. <xref ref-type="bibr" rid="ref26">Zhang et al. (2023)</xref> proposed an improved DCF-YOLOv8 model by optimizing the C2F module and adopting Mish functions to boost accuracy. <xref ref-type="bibr" rid="ref7">Li et al. (2024)</xref> introduced the YOLOv8-SDPS model, which replaces standard convolutions with SD_Conv modules, incorporates GSConv and VoV-GSCSP architectures, and adds PSA self-attention mechanisms. This enhanced model achieves dual capabilities: not only identifying six major rice pests but also detecting and locating infested rice plants. <xref ref-type="bibr" rid="ref22">Xu et al. (2022)</xref> developed a rapid identification method for small green leafhoppers using a color plate induction counting pattern combined with an improved deep learning model and sliding window algorithm, achieving an accuracy rate exceeding 89.29%. <xref ref-type="bibr" rid="ref28">Zhu et al. (2022)</xref> enhanced the recognition accuracy of micro-insect pests and small sample pests by implementing a fine-tuned two-stage small sample learning strategy with the TPH-YOLO v5 recognition algorithm. <xref ref-type="bibr" rid="ref6">Li (2024)</xref> successfully identified pests on Yunnan large-leaf tea plants using the &#x201C;low-level freezing, high-level training&#x201D; transfer learning approach with the Xception network model, reaching an accuracy of 99.17%. <xref ref-type="bibr" rid="ref11">Liu et al. (2025)</xref> proposed an improved YOLO v8s-CFW model for detecting five types of tea pests, incorporating a Convolutional Attention Module (CBAM) to enhance feature representation. The model utilized FasterNet as the backbone network to maintain high precision while improving computational efficiency, and employed the WIoU loss function to boost detection accuracy. <xref ref-type="bibr" rid="ref8">Li J. et al. (2025)</xref> proposed an improved YOLOv11n model called YOLO-MSLP to enhance multi-scale rice pest recognition accuracy. The model integrates the AP_BiFPN module and triplet attention mechanism MS-TAM, while employing RepViT and knowledge distillation techniques for model lightweighting. It achieves an average precision of 94.5%. Through optimizations of the YOLO series algorithm, research has improved recognition accuracy in crop pest detection and fruit surface defect identification across various crops including strawberries (<xref ref-type="bibr" rid="ref19">Wang et al., 2025</xref>), grapes (<xref ref-type="bibr" rid="ref20">Wu et al., 2025</xref>), tomatoes (<xref ref-type="bibr" rid="ref3">Fan et al., 2025</xref>; <xref ref-type="bibr" rid="ref21">Wu and Xu, 2025</xref>), citrus (<xref ref-type="bibr" rid="ref12">Luo et al., 2023</xref>), corn (<xref ref-type="bibr" rid="ref25">Yang et al., 2023</xref>; <xref ref-type="bibr" rid="ref5">Guo et al., 2024</xref>), soybeans (<xref ref-type="bibr" rid="ref10">Linqi et al., 2024</xref>; <xref ref-type="bibr" rid="ref4">Fang et al., 2023</xref>), pears (<xref ref-type="bibr" rid="ref9">Li C. et al., 2025</xref>), sugarcane (<xref ref-type="bibr" rid="ref23">Xu et al., 2023</xref>), and flowers (<xref ref-type="bibr" rid="ref24">Yang, 2023</xref>). However, existing studies still exhibited notable limitations. Most models were designed for conventional farmland crops without adequately addressing special scenarios in marginal tea gardens, such as complex terrain, pest species variations, and challenging sample collection. Some high-precision models suffered from excessive parameter complexity and computational costs, making them impractical for real-time deployment in edge devices at marginal tea gardens. There was a lack of specialized optimization models targeting dominant pests in marginal tea gardens, resulting in insufficient adaptability. This study addresses the core challenges of the tea horn beetle outbreak in Hunan&#x2019;s marginal tea plantations, including inefficient manual inspections and excessive chemical control. By leveraging the fundamental differences between marginal land and conventional soil to tackle pest management complexities, we develop a high-precision, lightweight, and edge-deployable intelligent pest recognition framework. Through implementing the ECA attention mechanism, GhostConv lightweight convolution, and transfer learning improvements to the YOLO11m model, this framework enabled real-time dynamic monitoring of pests in marginal tea plantations. It would replace traditional empirical observation methods, reduce chemical pesticide usage, and resolve the conflict between ecological protection and industrial efficiency in marginal tea cultivation. The green and efficient transformation of marginal tea industries are drived, which provide a universal technical paradigm for ecological conservation and pest identification in hilly fragile areas and other small-scale crops, demonstrating significant theoretical and practical value.</p>
</sec>
<sec sec-type="materials|methods" id="sec2">
<label>2</label>
<title>Materials and methods</title>
<sec id="sec3">
<label>2.1</label>
<title>Image acquisition</title>
<p>The image acquisition site was located in Gaoqiao Base Tea Garden, located in Gaoqiao Town, Changsha County, Changsha City, Hunan Province, the geographical coordinates of this base are &#x201C;113&#x00B0;4&#x2032;33.632&#x2033; East longitude and 28&#x00B0;15&#x2032;40.903&#x2033; North latitude.&#x201D; The soil type here was typical hilly red soil. Excellent varieties such as &#x201C;Zhuyeqi&#x201D; and &#x201C;Bixiangzao&#x201D; were planted, which would provide raw materials for making famous and high-quality green teas like &#x201C;Gaoqiao Yinfeng.&#x201D; The main pests included <italic>Empoasca pirisuga Matumura</italic> which sucked the juice from tender shoots leading to curled leaves and withered buds, <italic>Ectropis grisescens</italic> that voraciously eats leaves, and <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) that feed on leaves creating holes, and the image acquisition methods were divided into two types: (1) Cannon EOS 700D camera with an image resolution of 3,456 pixels &#x00D7; 5,184 pixels was used to capture the yellow trap board image hanging on the tea tree (as shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>); and (2) Carry out real-time image transmission under that tea tree for experimental identification through a special online monitor device of the insect situation (as shown in <xref ref-type="fig" rid="fig2">Figure 2</xref>). Through two collection methods, 806 pictures were collected. The data set was divided into the training set and the validation set in a ratio of 9:1, and the data set sample was shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Image of yellow trap plate of <italic>Basilepta melanopus (Lef&#x00E8;vre).</italic></p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Based on the phototaxis of insects, yellow sticky traps exhibit strong attractiveness to insect pests. Hanging yellow sticky traps in pest-infested areas can achieve green prevention and control of diseases and insect pests to a certain extent.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>On-line monitoring device for pest situation. <bold>(a)</bold> Presents the design drawing of the online pest monitoring device, while <bold>(b)</bold> shows the physical prototype of the device manufactured in accordance with the design drawing.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">The diagram labeled &#x201C;a&#x201D; presents a 3D design model of an online pest monitoring device. The photograph labeled &#x201C;b&#x201D; shows the physical prototype of the device manufactured in accordance with the design model in &#x201C;a&#x201D;, which consists of core components including a yellow sticky trap, an image acquisition module, a data transmission unit, and a distribution box.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>Network structure of ECA attention mechanism.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">This is an ECA attention mechanism structure: First, global average pooling is applied to the input feature map, compressing each channel&#x2019;s spatial info into a value to get a channel descriptor. Then, 1D convolution models cross-channel dependencies&#x2014;avoiding info loss from sudden dimensionality reduction and cutting computation. Finally, Sigmoid generates channel attention weights, which multiply element-wise with the original map to enhance key features and suppress minor ones.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec4">
<label>2.2</label>
<title>Image preprocessing</title>
<p>Labelimage was used for label annotation, and the position and category of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) in each image were manually annotated. The labeled data are cleaned and verified to ensure the quality and accuracy of data. This study used three data enhancement techniques to enhance the dataset: vertical flipping, random resizing, and random rotation. Each image in the data set was enhanced twice, and there were a total of 2,410 images after data enhancement (including 2,169 images in the training set and 241 images in the validation set). The number of tags <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) in the image data set were 20,569 in the training set and 2063 in the validation set, respectively. By applying these data augmentation techniques, this study was able to train a more stable and accurate model to detect pests.</p>
</sec>
<sec id="sec5">
<label>2.3</label>
<title>YOLO11 network model</title>
<p>YOLO algorithm had become one of the most important algorithms in the field of target detection because of its real-time, one-step, multi-scale feature fusion, prediction box design and multi-task learning. In this experiment, YOLO11m was used as the identification algorithm of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) in the image, and it was improved. YOLO11 was a model launched by Ultralytics team on September 30, 2024. As the latest iteration of YOLO series, it had made many innovations and optimizations on the basis of its predecessor. These improvements significantly improved its performance and adaptability, making it the preferred tool in many computer vision applications, such as object detection and tracking, instance segmentation, image classification and pose estimation. YOLO11 offered the most exceptional detection accuracy, speed and efficiency, replacing the original C2f module with a C3k2 module in the backbone network and neck structure compared to YOLOv8. At the same time, a C2PSA module similar to the attention mechanism was added after the SPPF module to further enhance the ability of image feature extraction. In addition, by introducing the header design concept of YOLOv10, YOLO11 used a decoupling header based on anchor-free, in which the regression branch used ordinary convolution blocks and the classification branch uses depth-separable convolution (DWConv), which effectively reduced redundant computation and improves efficiency.</p>
</sec>
<sec id="sec6">
<label>2.4</label>
<title>EGT-YOLO11 model design</title>
<sec id="sec7">
<label>2.4.1</label>
<title>Improvement 1: add ECA attention mechanism</title>
<p>In order to enhance the model&#x2019;s ability to focus on key features and improve the detection accuracy, this chapter introduces the ECA (Efficient Channel Attention) attention mechanism into YOLO11m and embeds it into the feature extraction stage of the backbone network. The core idea of ECA attention mechanism is to adaptively adjust the weights of different feature channels through efficient channel attention learning, so that the model pays more attention to the features that are more important to the detection task. The structure of the ECA attention mechanism is shown in <xref ref-type="fig" rid="fig3">Figure 3</xref>.</p>
<p>The implementation process of the ECA attention mechanism is simple and efficient. Firstly, the input feature map is globally averaged and pooled, and the spatial information of each channel is compressed into a numerical value to obtain the channel description vector; then, the dependency relationship between different channels is calculated by one-dimensional convolution, which avoids the information loss caused by the sudden drop of dimensions and significantly reduces the amount of calculation; Finally, the attention weight of each channel is generated by Sigmoid activation function, and multiplied with the original feature map channel by channel to enhance the important features and suppress the secondary features. For example, assuming that the shape of the input feature map is 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;32, a 32&#x202F;&#x00D7;&#x202F;1 channel description vector is obtained after global average pooling, and then a 32&#x202F;&#x00D7;&#x202F;1 attention weight is generated by capturing the local correlation between channels through a one-dimensional convolution with kernel _ size of 3. Finally, the weights are multiplied by the original 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;32 feature map to obtain the weighted feature map.</p>
<p>This method could effectively learn channel attention without dimensionality reduction, and significantly improve the sensitivity of the model to key features and enhance the ability of feature expression with little increase in parameters and computation. The ECA attention mechanism provided a more accurate feature screening capability for the model, especially in complex background or small target detection scenarios, which helped to further improve the detection performance of YOLO11m.</p>
</sec>
<sec id="sec8">
<label>2.4.2</label>
<title>Improvement 2: add GhostConv</title>
<p>In order to reduce the amount of computation and the number of parameters while maintaining the performance of the model, GhostConv was added to YOLO11m in this chapter, replacing Conv in the backbone and neck networks with GhostConv. The process of GhostConv was shown in <xref ref-type="fig" rid="fig4">Figure 4</xref>. The core idea of GhostConv was to use a two-stage strategy for feature extraction. First, it used a small number of 1&#x202F;&#x00D7;&#x202F;1standard convolution kernels to generate basic feature maps; then, it used 5&#x202F;&#x00D7;&#x202F;5 deep convolution to further enrich these feature maps; the final output was the feature map spliced by these two stages. For example, assuming that the shape of the input feature map was 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;32, first use 1&#x202F;&#x00D7;&#x202F;1 convolution to reduce the number of channels of the input feature map, and the shape becomes 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;16; then used 5&#x202F;&#x00D7;&#x202F;5 depth convolution to extract features from each channel feature map, and the shape remains unchanged, which was still 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;16, which could be seen as a series of linear transformations of the previous layer; Finally, the output feature maps of the two convolutions were stacked in the channel dimension, and the shape was changed to 64&#x202F;&#x00D7;&#x202F;64&#x202F;&#x00D7;&#x202F;32, so as to achieve efficient feature representation with low computational cost. This approach not only significantly reduced the number of convolution kernels required, but also maintained the expressiveness of the model. GhostConv module provided an effective solution, which could significantly reduce the consumption of computing resources while ensuring or even improving the performance of the model, and opened up a new path for building a more efficient and flexible deep learning model. The method was particularly suitable for application scenes with higher requirements on real-time performance and efficiency.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>GhostConv.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Workflow Diagram of GhostConv: Its core idea is a two-stage feature extraction strategy. First, a small number of 1&#x00D7;1 standard convolution kernels generate base feature maps. Then, 5&#x00D7;5 depthwise convolutions further enrich these maps. The final output is the concatenation of feature maps from the two stages.</alt-text>
</graphic>
</fig>
</sec>
<sec id="sec9">
<label>2.4.3</label>
<title>Improvement three: transfer learning</title>
<p>Deep learning needs a lot of data to drive, so it was a very time-consuming and costly method to train a model from scratch. A common training technique in deep learning was transfer learning, which transplanted the model weight parameters trained on the source domain to the model that needed new target learning, which could not only improve the convergence speed of the model, but also reduce the occurrence of model overfitting to a certain extent. This study uses transfer learning to further improve the accuracy of the model. The process is shown in <xref ref-type="fig" rid="fig5">Figure 5</xref>. The pest part of the Pest24 data set had many similar characteristics with the self-made pest data set, so the Pest24 data set was used as the source domain to train the model, so that the model could fully learn the general characteristics of pests, and then the self-made <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) data set was used as the target domain. Then the GET-YOLO11 model was trained to make the model have the ability to identify pests in the field, and finally the GET-YOLO11 model with higher accuracy was obtained. The GET-YOLO11 model structure diagram was shown in <xref ref-type="fig" rid="fig6">Figure 6</xref>.</p>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Transfer learning.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Transfer Learning Process: The Pest24 dataset is used as the source domain to train the model for learning general pest features. Then, the self-built Basilepta melanopus (Lef&#x00E8;vre) dataset serves as the target domain to fine-tune the GET-YOLO11 model, enabling it to identify field pests and yielding a more accurate GET-YOLO11 model.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Network structure of EGT-YOLO11.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g006.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Network structure of EGT-YOLO11,This model consists of three parts: Backbone, Neck, and Detection Head. The Backbone uses GhostConv, CSk2, SPPF, ECA, and C2PSA for feature extraction. The Neck employs Contact and Upsample to fuse features. The Detection Head (Head1-3) outputs detection results via CSk2 and GhostConv.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="sec10">
<label>2.5</label>
<title>Model evaluation index</title>
<p>The test was conducted under the 64-bit operating system Ubuntu 22.04.4 LTS, and the server configuration was as follows: GPU was RTX4090, memory was 64GB, and CPU is AMD EPYC 9354. It was completed in the deep learning environment built by Python-3.10, PyTorch-2.2.2 and CUDA-12.8. In the training process, the number of training batches was 32, and the total number of training rounds was 100. SGD was used to optimize the parameters, the learning rate was 0.01, the weight attenuation coefficient was 0.0005, and the learning rate momentum was 0.937.</p>
<p>In this study, precision (<italic>P</italic>), recall (<italic>R</italic>), average precision meant (mean average precisoin, <italic>mAP</italic>), required floating-point operations [floating-point operations per second, FLOPS, Params and FPS (frames per seconds)] to evaluate the performance of the model. The calculation formula of precision is shown in <xref ref-type="disp-formula" rid="E1">Equation 1</xref>, that of recall in <xref ref-type="disp-formula" rid="E2">Equation 2</xref>, and that of mean average precision in <xref ref-type="disp-formula" rid="E3 E4">Equations 3,4</xref>.</p><disp-formula id="E1">
<mml:math id="M1">
<mml:mi>P</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FP</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(1)</label>
</disp-formula><disp-formula id="E2">
<mml:math id="M2">
<mml:mi>R</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mrow>
<mml:mi mathvariant="italic">TP</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi mathvariant="italic">FN</mml:mi>
</mml:mrow>
</mml:mfrac>
</mml:math>
<label>(2)</label>
</disp-formula><disp-formula id="E3">
<mml:math id="M3">
<mml:mi mathvariant="italic">AP</mml:mi>
<mml:mo>=</mml:mo>
<mml:msubsup>
<mml:mo>&#x222B;</mml:mo>
<mml:mn>0</mml:mn>
<mml:mn>1</mml:mn>
</mml:msubsup>
<mml:mi>P</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>R</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mi mathvariant="italic">dR</mml:mi>
</mml:math>
<label>(3)</label>
</disp-formula><disp-formula id="E4">
<mml:math id="M4">
<mml:mi mathvariant="italic">mAP</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:msubsup>
<mml:mi>A</mml:mi>
<mml:msub>
<mml:mi>P</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:mfrac>
</mml:math>
<label>(4)</label>
</disp-formula>
<p><italic>TP</italic> was a true positive sample, indicating the number of correctly identified <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>), the prediction box was the same as the label box and the intersection ratio was greater than 0.5; <italic>FP</italic> was a false positive sample, indicating the number of false identification of <italic>Basilepta melanopus(Lef&#x00E8;vre)</italic>; <italic>FN</italic> was a false-negative samples, indicating the number of unrecognized <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>); <italic>N</italic> was the number of detection categories, which was 1.</p>
</sec>
</sec>
<sec sec-type="results|discussion" id="sec11">
<label>3</label>
<title>Results and discussion</title>
<sec id="sec12">
<label>3.1</label>
<title>Comparative test and analysis of ECA and other attention mechanisms</title>
<p>In order to verify the effectiveness of ECA attention mechanism, this study analyzed the relationship between ECA attention mechanism and global attention (global attention mechanism, GAM), convolutional attention module [convolutional block attention module, CBAM and SE (squeeze and excitation networks)]. The comparison results are shown in <xref ref-type="table" rid="tab1">Table 1</xref>.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Comparison of results of different attention mechanisms.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model</th>
<th align="center" valign="top"><italic>P</italic>/%</th>
<th align="center" valign="top"><italic>R</italic>/%</th>
<th align="center" valign="top"><italic>mAP</italic><sub>50</sub>/%</th>
<th align="center" valign="top">Floating-point operations per second/G</th>
<th align="center" valign="top">Params/M</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">YOLO11m</td>
<td align="char" valign="top" char=".">82.10</td>
<td align="char" valign="top" char=".">82.84</td>
<td align="char" valign="top" char=".">86.11</td>
<td align="char" valign="top" char="."><bold>67.60</bold></td>
<td align="char" valign="top" char="."><bold>20.03</bold></td>
</tr>
<tr>
<td align="left" valign="middle">YOLO11m + GAM</td>
<td align="char" valign="top" char=".">81.50</td>
<td align="char" valign="top" char=".">86.14</td>
<td align="char" valign="top" char=".">84.14</td>
<td align="char" valign="top" char=".">72.90</td>
<td align="char" valign="top" char=".">26.57</td>
</tr>
<tr>
<td align="left" valign="middle">YOLO11m + CBAM</td>
<td align="char" valign="top" char=".">83.39</td>
<td align="char" valign="top" char=".">83.69</td>
<td align="char" valign="top" char=".">86.19</td>
<td align="char" valign="top" char=".">67.90</td>
<td align="char" valign="top" char=".">20.29</td>
</tr>
<tr>
<td align="left" valign="middle">YOLO11m + SE</td>
<td align="char" valign="top" char=".">80.29</td>
<td align="char" valign="top" char="."><bold>86.86</bold></td>
<td align="char" valign="top" char=".">86.53</td>
<td align="char" valign="top" char=".">67.70</td>
<td align="char" valign="top" char=".">20.06</td>
</tr>
<tr>
<td align="left" valign="middle">YOLO11m + ECA</td>
<td align="char" valign="top" char="."><bold>83.95</bold></td>
<td align="char" valign="top" char=".">84.96</td>
<td align="char" valign="top" char="."><bold>87.05</bold></td>
<td align="char" valign="top" char=".">67.70</td>
<td align="char" valign="top" char="."><bold>20.03</bold></td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Values in bold represent the optimal performance data among the comparative datasets.</p>
</table-wrap-foot>
</table-wrap>
<p>The <italic>mAP</italic><sub>50</sub> of YOLO11m&#x202F;+&#x202F;ECA reached 87.05%, which was 0.94% higher than that of the baseline model YOLO11m, ranking first among all the comparison models. Although the recall rate of YOLO11m&#x202F;+&#x202F;SE is high, ECA further improves the precision rate to 83. 95% while maintaining a high recall rate through efficient channel attention learning, which achieves the collaborative optimization of precision and recall rate. The data set of this study is based on the yellow plate collection of tea garden in Hunan, and the image of the thorax beetle is easily disturbed by the shadow of leaves and the background of yellow plate, and some samples have the characteristics of small target size. In this case, YOLO11m&#x202F;+&#x202F;GAM greatly increases the amount of calculation and the number of parameters, resulting in the imbalance of model performance, and the <italic>mAP</italic><sub>50</sub> is reduced to 84.14%. Although the computational cost of YOLO11m&#x202F;+&#x202F;CBAM was close to the baseline, the <italic>mAP</italic><sub>50</sub> was only slightly improved to 86.19% due to the simple stacking channels and spatial attention module, which did not adapt to the complex characteristics of field targets, while the accuracy of YOLO11m&#x202F;+&#x202F;SE was reduced to 80.29% due to the loss of information in the process of dimensional processing. In sharp contrast, ECA can adaptively adjust the weight of feature channels without dimensionality reduction by means of the design of one-dimensional convolution channel dependence, which not only avoids the loss of information, but also does not significantly increase the computational cost, and can accurately capture the key features of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>), effectively avoid the problem of feature redundancy, and avoid the problem of feature redundancy. This method thereby shows better detection performance on a field-collected dataset of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>).</p>
<p>The experimental results of the ECA attention mechanism aligned with those reported by <xref ref-type="bibr" rid="ref17">Tang et al. (2023)</xref>. Their study implemented the ECA attention mechanism to enhance key feature selection on agricultural pest detection, achieving a 73.4% <italic>mAP</italic><sub>0.5</sub> score on the Pest24 dataset. By applying YOLO11m&#x202F;+&#x202F;ECA, we improved the <italic>mAP</italic><sub>0.5</sub> to 87.05%, further demonstrating the feasibility of the ECA attention mechanism in identifying small agricultural pests. In their tea disease detection study <xref ref-type="bibr" rid="ref15">Song et al. (2025)</xref>, achieved 98.0% accuracy by integrating the EMA attention mechanism with deformable attention. However, since the dataset primarily contains single-leaf images and disease leaves exhibit large-scale contiguous features, which differ from the small target pests in this research, the EMA attention mechanism struggles to improve detection performance for small target pests.</p>
</sec>
<sec id="sec13">
<label>3.2</label>
<title>Ablation experiment</title>
<p>In order to verify the effectiveness of each improvement proposed in this study, the original model YOLO11m is used as the baseline model, and a series of ablation tests are carried out for each improvement proposed. The ablation tests are carried out through different combinations of each improved module, and the results are shown in <xref ref-type="table" rid="tab2">Table 2</xref>.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Ablation experiment.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">ECA</th>
<th align="center" valign="top">GhostConv</th>
<th align="center" valign="top">Transfer learning</th>
<th align="center" valign="top"><italic>P</italic>/%</th>
<th align="center" valign="top"><italic>R</italic>/%</th>
<th align="center" valign="top"><italic>mAP</italic><sub>50</sub>/%</th>
<th align="center" valign="top">Floating-point operations per second/G</th>
<th align="center" valign="top">Params/M</th>
<th align="center" valign="top">Frames per seconds /frames&#x00B7;s<sup>&#x2212;1</sup></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">&#x00D7;</td>
<td align="center" valign="middle">&#x00D7;</td>
<td align="center" valign="middle">&#x00D7;</td>
<td align="char" valign="top" char=".">82.10</td>
<td align="char" valign="top" char=".">82.84</td>
<td align="char" valign="top" char=".">86.11</td>
<td align="char" valign="top" char=".">67.60</td>
<td align="char" valign="top" char=".">20.03</td>
<td align="char" valign="middle" char=".">161.29</td>
</tr>
<tr>
<td align="left" valign="middle">&#x2713;</td>
<td align="center" valign="middle">&#x00D7;</td>
<td align="center" valign="middle">&#x00D7;</td>
<td align="char" valign="top" char="."><bold>83.95</bold></td>
<td align="char" valign="top" char=".">84.96</td>
<td align="char" valign="top" char=".">87.05</td>
<td align="char" valign="top" char=".">67.70</td>
<td align="char" valign="top" char=".">20.03</td>
<td align="char" valign="top" char=".">159.25</td>
</tr>
<tr>
<td align="left" valign="middle">&#x2713;</td>
<td align="center" valign="middle">&#x2713;</td>
<td align="center" valign="middle">&#x00D7;</td>
<td align="char" valign="middle" char=".">83.20</td>
<td align="char" valign="middle" char=".">85.70</td>
<td align="char" valign="middle" char=".">87.71</td>
<td align="char" valign="middle" char="."><bold>54.50</bold></td>
<td align="char" valign="middle" char="."><bold>14.65</bold></td>
<td align="char" valign="middle" char="."><bold>185.16</bold></td>
</tr>
<tr>
<td align="left" valign="middle">&#x2713;</td>
<td align="center" valign="middle">&#x2713;</td>
<td align="center" valign="middle">&#x2713;</td>
<td align="char" valign="middle" char=".">82.52</td>
<td align="char" valign="middle" char="."><bold>87.20</bold></td>
<td align="char" valign="middle" char="."><bold>87.94</bold></td>
<td align="char" valign="middle" char="."><bold>54.50</bold></td>
<td align="char" valign="middle" char="."><bold>14.65</bold></td>
<td align="char" valign="middle" char=".">182.34</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Values in bold represent the optimal performance data among the comparative datasets.</p>
</table-wrap-foot>
</table-wrap>
<p>When only ECA attention mechanism was introduced, the precision of the model increases by 83.95%, the recall increased to 84.96%, <italic>mAP</italic><sub>50</sub> increases to 87.05%, and the number of floating-point operations increased only slightly to 67.70&#x202F;G. This showed that the ECA module captures the channel dependence through one-dimensional convolution, and could adaptively enhance the key features of the thorax beetle without dimensionality reduction, such as the edge and texture details of the beetle, which effectively reduced the false detection and missed detection caused by the background interference of the yellow board, and would not significantly increase the complexity of the model. When the combination of ECA and GhostConv was introduced, the number of model parameters was reduced from 20.03&#x202F;M to 14.65&#x202F;M, the number of floating-point operations was reduced from 67.70&#x202F;G to 54.50&#x202F;G, and the frame rate was increased to 185.16 frames. At the same time, <italic>mAP</italic><sub>50</sub> further increased to 87.71%, and the recall rate increased to 85.70%. Although the accuracy was slightly reduced to 83.20%, the significant improvement of recall and computational efficiency showed that GhostConv achieved the initial balance of &#x201C;accuracy-efficiency&#x201D; by reducing redundant computation while retaining small target features through the two-stage strategy of &#x201C;1&#x202F;&#x00D7;&#x202F;1 convolution dimension reduction + 5&#x202F;&#x00D7;&#x202F;5 deep convolution complementary features.&#x201D; When ECA, GhostConv and transfer learning steps were integrated at the same time, the <italic>mAP</italic><sub>50</sub> of the model reaches 87.94%, the recall rate was significantly improved to 87.20%, and a further breakthrough in detection accuracy was achieved, which proved that transfer learning could learn the general morphological characteristics of pests through the pre-training of Pest24 source domain data set. The overfitting caused by sparse samples in the field was effectively alleviated and the slight accuracy loss of ECA&#x202F;+&#x202F;GhostConv combination was compensated by adapting the target domain data of B. theobromae, while GhostConv continued to play a lightweight role to control the computational cost and ensure the real-time of the model. To sum up, ECA is responsible for feature selection, GhostConv optimizes computational efficiency, and transfer learning improves sample adaptability. The three modules work together to promote the improved model to achieve a comprehensive optimization of &#x201C;high precision-lightweight-high real-time&#x201D; in the identification task of the <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>), which verifies the effectiveness and rationality of the improvement strategy in this study.</p>
<p>In their tomato detection study, Tian et al. addressed challenges including small targets, severe occlusion, and complex background interference by integrating GhostConv and C2fGhost modules into a neck network, achieving a 96.31% <italic>mAP</italic><sub>0.5</sub> and 2.7&#x202F;M ultra-lightweight parameter count (<xref ref-type="bibr" rid="ref18">Tian et al., 2024</xref>). This validated GhostConv&#x2019;s core value in agricultural small target detection. For the micro-sized tea leaf beetle, the study further fused ECA attention with lightweight GhostConv convolution at the module level, while employing transfer learning to mitigate overfitting. These enhancements significantly improved feature extraction precision and anti-interference capabilities, demonstrating the advantages of fine-tuning models for specific tea pest morphological characteristics.</p>
</sec>
<sec id="sec14">
<label>3.3</label>
<title>Comparative test</title>
<p>RT-DETR, YOLOv5, YOLOv8, YOLOv9, YOLOv10, YOLO11, YOLO12 and GET-YOLO were successively used in the detection of Thorax cornutus, and the experimental results were shown in <xref ref-type="table" rid="tab3">Table 3</xref>.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Comparative experiment.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model</th>
<th align="center" valign="top"><italic>P</italic>/%</th>
<th align="center" valign="top"><italic>R</italic>/%</th>
<th align="center" valign="top"><italic>mAP</italic><sub>50</sub>/%</th>
<th align="center" valign="top">Floating-point operations per second /G</th>
<th align="center" valign="top">Params/M</th>
<th align="center" valign="top">Frames per seconds/frames&#x00B7;s<sup>&#x2212;1</sup></th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="middle">RT-DETR</td>
<td align="char" valign="top" char=".">70.79</td>
<td align="char" valign="top" char=".">76.44</td>
<td align="char" valign="top" char=".">74.48</td>
<td align="char" valign="top" char=".">103.40</td>
<td align="char" valign="top" char=".">31.99</td>
<td align="char" valign="top" char=".">92.91</td>
</tr>
<tr>
<td align="left" valign="middle">YOLOv5m</td>
<td align="char" valign="top" char=".">80.84</td>
<td align="char" valign="top" char=".">81.77</td>
<td align="char" valign="top" char=".">84.59</td>
<td align="char" valign="top" char=".">64.00</td>
<td align="char" valign="top" char=".">20.06</td>
<td align="char" valign="top" char=".">155.87</td>
</tr>
<tr>
<td align="left" valign="middle">YOLOv8m</td>
<td align="char" valign="top" char=".">81.80</td>
<td align="char" valign="top" char=".">85.41</td>
<td align="char" valign="top" char=".">85.17</td>
<td align="char" valign="top" char=".">78.70</td>
<td align="char" valign="top" char=".">25.84</td>
<td align="char" valign="top" char=".">151.29</td>
</tr>
<tr>
<td align="left" valign="middle">YOLOv9c</td>
<td align="char" valign="top" char="."><bold>82.68</bold></td>
<td align="char" valign="top" char=".">85.50</td>
<td align="char" valign="top" char=".">85.42</td>
<td align="char" valign="top" char=".">102.30</td>
<td align="char" valign="top" char=".">25.32</td>
<td align="char" valign="top" char=".">117.34</td>
</tr>
<tr>
<td align="left" valign="middle">YOLOv10m</td>
<td align="char" valign="top" char=".">80.68</td>
<td align="char" valign="top" char=".">81.92</td>
<td align="char" valign="top" char=".">83.83</td>
<td align="char" valign="top" char=".">58.90</td>
<td align="char" valign="top" char=".">15.31</td>
<td align="char" valign="top" char="."><bold>185.19</bold></td>
</tr>
<tr>
<td align="left" valign="middle">YOLO11m</td>
<td align="char" valign="top" char=".">82.10</td>
<td align="char" valign="top" char=".">82.84</td>
<td align="char" valign="top" char=".">86.11</td>
<td align="char" valign="top" char=".">67.60</td>
<td align="char" valign="top" char=".">20.03</td>
<td align="char" valign="top" char=".">161.29</td>
</tr>
<tr>
<td align="left" valign="middle">YOLO12m</td>
<td align="char" valign="top" char=".">82.34</td>
<td align="char" valign="top" char=".">84.05</td>
<td align="char" valign="top" char=".">85.95</td>
<td align="char" valign="top" char=".">67.10</td>
<td align="char" valign="top" char=".">20.11</td>
<td align="char" valign="top" char=".">159.43</td>
</tr>
<tr>
<td align="left" valign="top">GET-YOLO</td>
<td align="char" valign="middle" char=".">82.52</td>
<td align="char" valign="middle" char="."><bold>87.20</bold></td>
<td align="char" valign="middle" char="."><bold>87.94</bold></td>
<td align="char" valign="middle" char="."><bold>54.50</bold></td>
<td align="char" valign="middle" char="."><bold>14.65</bold></td>
<td align="char" valign="middle" char=".">182.34</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Values in bold represent the optimal performance data among the comparative datasets.</p>
</table-wrap-foot>
</table-wrap>
<p>In the aspect of target detection core index <italic>mAP</italic><sub>50</sub>, GET-YOLO was superior to all comparison models with 87.94% detection accuracy, which was 1.83% higher than baseline model YOLO11m, 2.77% higher than YOLOv8m and 4.11% higher than YOLOv10m. Compared with the traditional two-stage model RT-DETR, it was improved by 13.46%, which fully verified the improvement effect of the improved strategy in this study on the identification accuracy of the thorax tea beetle. On the premise of maintaining the highest <italic>mAP</italic><sub>50</sub>, GET-YOLO showed excellent precision-recall balance: its recall rate reached 87.20%, which was the highest among all models, and effectively reduced the missed detection rate of the sparse tea pod beetle in the field. The accuracy is 82.52%, which is slightly lower than that of YOLOv9c, but higher than that of most models such as YOLOv5m and YOLOv10m. Compared with YOLO12m, GET-YOLO improves the recall by 3.15% and the precision by 0.18%, which indicates that the improved structure can effectively reduce the false detection rate caused by the background interference of yellow boards and enhance the adaptability of complex field scenes. In terms of model lightweight, GET-YOLO is the lightest scheme among all the comparison models with a parameter quantity of 14.65&#x202F;M, which is 26.9% lower than benchmark YOLO11m and 54.2% lower than RT-DETR. The lightweight advantage is significant, and it has high practicability in the field deployment of tea garden. In terms of frame rate, GET-YOLO reaches 182.34 frames&#x00B7;s<sup>&#x2212;1</sup>, which is 89.43 and 65.00 frames&#x00B7;s<sup>&#x2212;1</sup> higher than RT-DETR and YOLOv9c, respectively. Although it is 2.85 frames&#x00B7;s<sup>&#x2212;1</sup> lower than fastest YOLOv10m, it still fully meets the speed requirements of real-time pest monitoring in tea gardens on the premise of maintaining higher detection accuracy, reflecting a good balance between &#x201C;speed and accuracy.&#x201D; To sum up, through the collaborative optimization of ECA attention mechanism, GhostConv lightweight convolution and transfer learning, GET-YOLO is superior to the comparison model in terms of detection accuracy, model efficiency and adaptability to complex field scenes, which verifies the effectiveness and advancement of the improved strategy proposed in this study. And a bet technical scheme is provided for field real-time monitor of that Sterculia theae.</p>
<p>In order to visually verify the key role of the improved module in the identification of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>), this study shows the real bounding box through visual comparison experiments, and chooses YOLOv8m, YOLOv9c, YOLO11m and YOLO12m with better effect for visual comparison with GET-YOLO. The comparison results are shown in <xref ref-type="fig" rid="fig7">Figure 7</xref>. It can be seen from <xref ref-type="fig" rid="fig7">Figure 7</xref> that GET-YOLO shows a lower missed detection rate and a more stable detection confidence in the detection of <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>), especially in small target recognition and complex background adaptation scenarios, and can accurately detect <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) with sparse distribution, dense arrangement and multi-scale. In <xref ref-type="fig" rid="fig7">Figure 7a</xref>, the detection confidence of YOLOv8m and YOLOv9c was low, which reflected the lack of feature characterization ability. Although the confidence of YOLO11m and YOLO12m was improved, the confidence of most of the beetles was lower than that of GET-YOLO. GET-YOLO strengthens the key features such as texture and edge of the worm body through the ECA attention module, which effectively improves the confidence of prediction and makes the detection more accurate. In the type of images shown in <xref ref-type="fig" rid="fig7">Figure 7b</xref>, some individuals of the <italic>Basilepta melanopus</italic> (<italic>Lef&#x00E8;vre</italic>) are distributed in the top area of the sticky traps. After being stuck, the pests tend to struggle, which easily leads to the damage or incompleteness of their body morphology. This makes the identification of incomplete pest individuals located at the image boundaries a technical difficulty in practical monitoring: YOLOv8m and YOLOv9c miss the target; By virtue of GhostConv&#x2019;s ability to retain small target features, combined with the generalization optimization of transfer learning, GET-YOLO not only successfully detects the target, but also has a higher confidence than YOLO11m and YOLO12m. In <xref ref-type="fig" rid="fig7">Figure 7c</xref>, YOLOv8m, YOLOv9c, YOLO11m and YOLO12m were not detected in the lower left, which was close to the background noise. GET-YOLO accurately captures this goal through the synergy of feature enhancement and lightweight of ECA&#x202F;+&#x202F;GhostConv. To sum up, GET-YOLO is superior to the comparison model in visualization, which intuitively verifies the synergistic value of ECA attention, GhostConv and transfer learning.</p>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Comparison of detection effects of different models. <bold>(a-c)</bold> Present the experimental results of visual comparison performed by five models&#x2014;YOLOv8m, YOLOv9c, YOLO11m, YOLO12m and GET-YOLO&#x2014;on three images, where each row corresponds to one group of comparison results.</p>
</caption>
<graphic xlink:href="fsufs-09-1734639-g007.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Comparison of object detection outputs from five models: YOLOv8m, YOLOv9c, YOLO11m, YOLO12m, and GET-YOLO. Three rows labeled a, b, and c display images with marked Basilepta melanopus detections and their confidence scores. Each column shows results from different models, highlighting variations in detection accuracy and precision across the images.</alt-text>
</graphic>
</fig>
<p>In this study, 806 images of marginal tea plantations in the main producing areas of Hunan were collected, and the YOLO11m model was optimized by combining ECA attention, GhostConv lightweight convolution and transfer learning, and then the SPXY method was used to divide the training set and the validation set, and the identification model of the population density of the tea pod beetle was established by deep learning method. The Results showed that the model optimized by GET-TOLO had the best performance in pest identification, the <italic>mAP</italic><sub>50</sub>, precision and recall of the validation set were 87.94%, 89.13% and 88.47%, respectively, the single inference time was 23&#x202F;ms, and the number of parameters was reduced by 28%. It showed that the lightweight deep learning technology could be used to quickly and accurately identify pests in marginal tea plantations.</p>
<p>However, due to the great differences in terrain, illumination and shelter degree of marginal tea plantations, the existing general pest detection model cannot be directly used, so a lightweight identification model suitable for regional characteristics must be established. There were still some deviations in the identification results of this study, which might be due to the fact that the samples were not cleaned when they were labeled, and the extreme light outliers were not eliminated in the data analysis, so the role of outlier elimination method in improving the performance of the model needs further study. Currently, integrated pest management (IPM) strategies for the <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) remain predominantly chemical-based, with limited research on green control technologies. <xref ref-type="bibr" rid="ref27">Zhang et al. (2025)</xref> systematically summarized existing control methods, categorizing them into six approaches: agricultural management and cultivation practices, pathogen control, plant-derived insecticides, natural enemy insect utilization, chemical control, and color trap trapping. <xref ref-type="bibr" rid="ref14">Qiu et al. (2025)</xref> conducted leaf immersion tests to evaluate the efficacy of 8 chemical agents and 6 biological agents against adult <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) in laboratory settings, followed by field trials to identify the optimal formulation for practical application.</p>
<p>The research would promote: (1) Iteratively optimize the existing identification algorithm, improve the identification accuracy and generalization ability of the model in the complex environment of marginal land, and expand to other categories of tea diseases and insect pests, such as tea aphids and tea leafhoppers, to build an intelligent identification system of tea diseases and insect pests covering marginal land tea gardens. (2) Taking the tea garden of typical marginal land such as steep slope and mountain in hilly red soil area of Xiangxi Prefecture, Hunan Province as the study area, the field application and verification of the model were carried out to clarify the adaptability and improvement direction of the model in the special habitat of marginal land in Hunan Province (acid clay red soil, steep mountain slope); (3) Integrating the multi-source data of environmental factors (such as red soil fertility, steep slope light and water), tea physiological and ecological indicators (such as stress resistance related indicators, new shoot growth rate) and other data of marginal land tea garden, exploring and establishing an information management model suitable for the whole production process of marginal land tea garden in Hunan Province. It provides technical support and key foundation for the efficient utilization, precise management and sustainable development of marginal land tea plantations (such as steep slopes and mountain tea plantations in hilly red soil areas of Xiangxi Prefecture, Hunan Province).</p>
<p>Current research in intelligent identification systems and real-time population monitoring of the <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) remains underdeveloped, posing significant challenges to implementing precise pest management strategies. The proposed intelligent identification method not only enhances technical accuracy and efficiency in pest detection but also holds practical significance for ecological conservation and sustainable management of tea plantations. Traditional tea garden pest control heavily relies on broad-spectrum chemical pesticides, whose excessive use leads to soil contamination, excessive pesticide residues in tea leaves, and disruption of natural ecosystem regulation. The GET-TOLO model developed in this study enables accurate monitoring of <italic>Basilepta melanopus</italic> (Lef&#x00E8;vre) population dynamics and identifies critical density thresholds for pest outbreaks. Tea farmers can implement targeted control measures based on this information, significantly reducing chemical pesticide usage frequency and dosage. Notably, this approach aligns with the perspective proposed by <xref ref-type="bibr" rid="ref27">Zhang et al. (2025)</xref>, who emphasized the urgent need to develop intelligent identification technologies and related products for automated population monitoring to support timely and precise pest control decisions. Therefore, this research holds substantial application value in advancing green pest management practices and promoting ecological sustainability in tea cultivation systems.</p>
</sec>
</sec>
<sec sec-type="conclusions" id="sec15">
<label>4</label>
<title>Conclusion</title>
<p>This study showed the pest threats faced by tea cultivation on approximately 638,800 hectares of marginal land in Hunan Province. It proposed an innovative GET-YOLO lightweight model to overcome the limitations of traditional manual detection and chemical control methods, thereby meeting the demands of a green and high-quality tea industry. The GET-YOLO model significantly enhanced key feature extraction for tea tree pests by embedding an ECA attention mechanism which built upon the YOLO11m deep learning framework, thereby improving both detection accuracy and recall rate. The model achieved an <italic>mAP</italic><sub>50</sub> score of 87.94%.</p>
<p>To achieve model lightweight, the research introduced GhostConv modle, which reduced the model parameters by 28%. Meanwhile, the transfer learning strategy improved the model&#x2019;s adaptability to small sample datasets and shortened the inference time of a single image to 23&#x202F;ms, achieving a balance of &#x201C;high precision, lightweight, and high real-time performance.&#x201D;</p>
<p>The results demonstrated that the GET-YOLO model outperforms traditional methods by achieving 0.42% improvement in precision, 4.36% increase in recall rate, and 1.83% enhancement in <italic>mAP</italic><sub>50</sub>, which showed superior performance in pest detection. The GET-YOLO model automatically tracks pest population dynamics, including the Empoasca onukii Matsuda (tea green leafhopper), providing real-time data for green pest control in tea plantations. This system holds significant theoretical and practical value for advancing the green transformation of the tea industry and optimizing marginal agricultural land use, effectively driving the sustainable and efficient upgrading of marginal tea cultivation.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec16">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="sec17">
<title>Author contributions</title>
<p>CY: Formal analysis, Methodology, Data curation, Writing &#x2013; original draft, Software, Visualization, Funding acquisition. QB: Writing &#x2013; review &#x0026; editing, Data curation. YW: Conceptualization, Writing &#x2013; original draft, Funding acquisition. XN: Writing &#x2013; original draft, Investigation, Project administration. LX: Writing &#x2013; original draft, Investigation, Data curation. CZ: Writing &#x2013; review &#x0026; editing, Writing &#x2013; original draft, Methodology.</p>
</sec>
<sec sec-type="COI-statement" id="sec18">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec19">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec20">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cao</surname><given-names>X. M.</given-names></name> <name><surname>Sun</surname><given-names>B.</given-names></name> <name><surname>Chen</surname><given-names>H. B.</given-names></name> <name><surname>Zhou</surname><given-names>J. M.</given-names></name> <name><surname>Song</surname><given-names>X.</given-names></name> <name><surname>Liu</surname><given-names>X. W.</given-names></name> <etal/></person-group>. (<year>2021</year>). <article-title>Approaches and research progress on expanding marginal land productivity and enhancing ecological benefits in China</article-title>. <source>Chin. Acad. Sci. J.</source> <volume>36</volume>, <fpage>336</fpage>&#x2013;<lpage>348</lpage>. doi: <pub-id pub-id-type="doi">10.16418/j.issn.1000-3045.20201228002</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cao</surname><given-names>L.</given-names></name> <name><surname>Xiao</surname><given-names>W.</given-names></name> <name><surname>Li</surname><given-names>X.</given-names></name></person-group> (<year>2025</year>) <article-title>Overview of the YOLO algorithm and its application in crop recognition and pest detection</article-title>. <source>J. Zhongkai Univ. Agric. Eng.</source> doi: <pub-id pub-id-type="doi">10.3969/j.issn.1674-5663.2025.06.010</pub-id></mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fan</surname><given-names>X. P.</given-names></name> <name><surname>Zhang</surname><given-names>Y. Q.</given-names></name> <name><surname>Zhou</surname><given-names>S.</given-names></name> <name><surname>Ren</surname><given-names>M. F.</given-names></name> <name><surname>Wang</surname><given-names>Y. W.</given-names></name> <name><surname>Cai</surname><given-names>X. L.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Recognition and position method of tomato picking robot based on improved YOLOv8s and RGB-D information fusion</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>41</volume>, <fpage>106</fpage>&#x2013;<lpage>116</lpage>. <comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202503181</pub-id></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Fang</surname><given-names>W. B.</given-names></name> <name><surname>Guo</surname><given-names>Y. G.</given-names></name> <name><surname>Guan</surname><given-names>F. C.</given-names></name> <name><surname>Zhang</surname><given-names>W.</given-names></name> <name><surname>Liu</surname><given-names>Q. Q.</given-names></name> <name><surname>Wang</surname><given-names>S. W.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Soybean leaf wormhole recognition based on improved YOLO v5s algorithm</article-title>. <source>J. Hunan Agric. Univ.</source> <volume>49</volume>, <fpage>127</fpage>&#x2013;<lpage>132</lpage>. doi: <pub-id pub-id-type="doi">10.13331/j.cnki.jhau.2023.01.018</pub-id></mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Guo</surname><given-names>J.</given-names></name> <name><surname>Wang</surname><given-names>R.</given-names></name> <name><surname>Nan</surname><given-names>J.</given-names></name> <name><surname>Xiaohu</surname><given-names>L. I.</given-names></name> <name><surname>Changzhe</surname><given-names>J.</given-names></name></person-group> (<year>2024</year>). <article-title>YOLOv5 model integrated with GRN attention mechanism for insect pest recognition</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>40</volume>, <fpage>159</fpage>&#x2013;<lpage>170</lpage>. <comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202310226</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Li</surname><given-names>Z.</given-names></name></person-group> (<year>2024</year>). <source>Research on tea disease and pest identification based on transfer learning and convolutional neural networks</source>. <publisher-loc>China</publisher-loc>: <publisher-name>Yunnan Agricultural University</publisher-name>.</mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>L.</given-names></name> <name><surname>Li</surname><given-names>M.</given-names></name> <name><surname>Li</surname><given-names>Z.</given-names></name></person-group> (<year>2024</year>). <article-title>Method for rice pest identification based on improved YOLO v8</article-title>. <source>Jiangsu Agric. Sci.</source> <volume>52</volume>, <fpage>209</fpage>&#x2013;<lpage>219</lpage>. doi: <pub-id pub-id-type="doi">10.15889/j.issn.1002-1302.2024.20.025</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>J. H.</given-names></name> <name><surname>Xu</surname><given-names>Y. L.</given-names></name> <name><surname>Lu</surname><given-names>J.</given-names></name> <name><surname>Li</surname><given-names>S. H.</given-names></name> <name><surname>Cai</surname><given-names>X.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Lightweight multi-scale rice pest recognition model based on improved YOLOv11n</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>41</volume>, <fpage>175</fpage>&#x2013;<lpage>183</lpage><comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202504222</pub-id></mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname><given-names>C. X.</given-names></name> <name><surname>Zhao</surname><given-names>C. J.</given-names></name> <name><surname>Zhang</surname><given-names>C.</given-names></name> <name><surname>Huang</surname><given-names>W. Q.</given-names></name> <name><surname>Li</surname><given-names>J. Q.</given-names></name> <name><surname>He</surname><given-names>X.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Multi-surface defect detection method for Dangshan pears based on AICYOLOv11n model</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>41</volume>, <fpage>320</fpage>&#x2013;<lpage>328</lpage>. <comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202504201</pub-id></mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Linqi</surname><given-names>Z.</given-names></name> <name><surname>Xiaoming</surname><given-names>L.</given-names></name> <name><surname>Hongmin</surname><given-names>S.</given-names></name> <name><surname>Yingpeng</surname><given-names>H.</given-names></name></person-group> (<year>2024</year>). <article-title>Research on CBF-YOLO detection model for common soybean pests in complex environment</article-title>. <source>Comput. Electron. Agric.</source> <volume>216</volume>:<fpage>108515</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2023.108515</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Liu</surname><given-names>R.</given-names></name> <name><surname>Wang</surname><given-names>L. J.</given-names></name> <name><surname>Wang</surname><given-names>Q. H.</given-names></name> <name><surname>Lin</surname><given-names>X. D.</given-names></name> <name><surname>Guo</surname><given-names>Q. H.</given-names></name> <name><surname>Xu</surname><given-names>D. L.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Detection of tea pests and diseases based on improved YOLOv8s</article-title>. <source>Acta Agric. Zhejiangensis</source> <volume>37</volume>, <fpage>1933</fpage>&#x2013;<lpage>1942</lpage>. doi: <pub-id pub-id-type="doi">10.3969/j.issn.1004-1524.20240836</pub-id></mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Luo</surname><given-names>D.</given-names></name> <name><surname>Xue</surname><given-names>Y.</given-names></name> <name><surname>Deng</surname><given-names>X.</given-names></name> <name><surname>Yang</surname><given-names>B.</given-names></name> <name><surname>Chen</surname><given-names>H.</given-names></name> <name><surname>Mo</surname><given-names>Z.</given-names></name></person-group> (<year>2023</year>). <article-title>Citrus diseases and pests detection model based on self-attention YOLOV8</article-title>. <source>IEEE Access</source> <volume>11</volume>, <fpage>139872</fpage>&#x2013;<lpage>139881</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ACCESS.2023.3340148</pub-id></mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ou</surname><given-names>G. C.</given-names></name> <name><surname>Wang</surname><given-names>Y. J.</given-names></name> <name><surname>Yang</surname><given-names>F. C.</given-names></name> <name><surname>Li</surname><given-names>G. H.</given-names></name> <name><surname>Bao</surname><given-names>Q.</given-names></name> <name><surname>Xiao</surname><given-names>L.</given-names></name> <etal/></person-group>. (<year>2018</year>). <article-title>Integration and application of green prevention and control techniques for tea pests and diseases in Hunan Province</article-title>. <source>China Plant Protect. Guide</source> <volume>38</volume>, <fpage>78</fpage>&#x2013;<lpage>80</lpage>.</mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Qiu</surname><given-names>H.</given-names></name> <name><surname>Tan</surname><given-names>R. R.</given-names></name> <name><surname>Chen</surname><given-names>X.</given-names></name> <name><surname>Wang</surname><given-names>H.</given-names></name> <name><surname>Huang</surname><given-names>D.</given-names></name> <name><surname>Wu</surname><given-names>P.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Study on safe chemical control of <italic>Basilepta melanopus</italic></article-title>. <source>J. Tea Sci.</source>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi: <pub-id pub-id-type="doi">10.13305/j.cnki.jts.20251103.002</pub-id></mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Song</surname><given-names>J.</given-names></name> <name><surname>Zhang</surname><given-names>Y.</given-names></name> <name><surname>Lin</surname><given-names>S.</given-names></name> <name><surname>Han</surname><given-names>H.</given-names></name> <name><surname>Yu</surname><given-names>X.</given-names></name></person-group> (<year>2025</year>). <article-title>TLDDM: an enhanced tea leaf pest and disease detection model based on YOLOv8</article-title>. <source>Agronomy</source> <volume>15</volume>:<fpage>727</fpage>. doi: <pub-id pub-id-type="doi">10.3390/agronomy15030727</pub-id></mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Su</surname><given-names>B.</given-names></name> <name><surname>Zhu</surname><given-names>Y. Y.</given-names></name> <name><surname>Lin</surname><given-names>Y.</given-names></name> <name><surname>Zhu</surname><given-names>Y.</given-names></name></person-group> (<year>2025</year>). <article-title>Pest-YOLOv8: enhanced detection for small and complex agricultural pests using triple attention and wise-IoU</article-title>. <source>J. Plant Dis. Protect.</source> <volume>132</volume>:<fpage>135</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s41348-025-01134-w</pub-id></mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tang</surname><given-names>Z.</given-names></name> <name><surname>Lu</surname><given-names>J.</given-names></name> <name><surname>Chen</surname><given-names>Z.</given-names></name> <name><surname>Qi</surname><given-names>F.</given-names></name> <name><surname>Zhang</surname><given-names>L.</given-names></name></person-group> (<year>2023</year>). <article-title>Improved Pest-YOLO: real-time pest detection based on efficient channel attention mechanism and transformer encoder</article-title>. <source>Ecol. Inform.</source> <volume>78</volume>:<fpage>102340</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.ecoinf.2023.102340.</pub-id></mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tian</surname><given-names>Z.</given-names></name> <name><surname>Hao</surname><given-names>H.</given-names></name> <name><surname>Dai</surname><given-names>G.</given-names></name> <name><surname>Li</surname><given-names>Y.</given-names></name></person-group> (<year>2024</year>). <article-title>Optimizing tomato detection and counting in smart greenhouses: a lightweight YOLOv8 model incorporating high-and low-frequency feature transformer structures</article-title>. <source>Network</source>, <fpage>1</fpage>&#x2013;<lpage>37</lpage>. doi: <pub-id pub-id-type="doi">10.1080/0954898X.2024.2428713</pub-id></mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname><given-names>S. C.</given-names></name> <name><surname>Wang</surname><given-names>H. Q.</given-names></name> <name><surname>Ding</surname><given-names>X. M.</given-names></name> <name><surname>Du</surname><given-names>X. P.</given-names></name> <name><surname>Yin</surname><given-names>Y. L.</given-names></name> <name><surname>Cui</surname><given-names>J. Y.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Pest monitoring in strawberry greenhouses using improved YOLOv8n</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>41</volume>, <fpage>184</fpage>&#x2013;<lpage>193</lpage><comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202411131</pub-id></mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname><given-names>S. C.</given-names></name> <name><surname>Mao</surname><given-names>Y. M.</given-names></name> <name><surname>Hu</surname><given-names>H. Z.</given-names></name> <name><surname>Zhang</surname><given-names>W. K.</given-names></name> <name><surname>Xing</surname><given-names>S. Y.</given-names></name> <name><surname>Duan</surname><given-names>H. Q.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Detecting grape leaf diseases using improved YOLO11n</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source> <volume>41</volume>, <fpage>140</fpage>&#x2013;<lpage>147</lpage>. <comment>(in Chinese with English abstract)</comment>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202503032</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname><given-names>L.</given-names></name> <name><surname>Xu</surname><given-names>X.</given-names></name></person-group> (<year>2025</year>). <article-title>Lightweight tomato leaf disease and pest detection method based on improved YOLOv10n</article-title>. <source>Smart Agric</source> <volume>7</volume>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. doi: <pub-id pub-id-type="doi">10.12133/j.smartag.SA202410023</pub-id> <comment>(in Chinese with English abstract)</comment></mixed-citation></ref>
<ref id="ref22"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname><given-names>R.</given-names></name> <name><surname>Jin</surname><given-names>Z. H.</given-names></name> <name><surname>Luo</surname><given-names>L. W.</given-names></name> <name><surname>Feng</surname><given-names>H. Q.</given-names></name> <name><surname>Fang</surname><given-names>H. H.</given-names></name> <name><surname>Wang</surname><given-names>X. C.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Monitoring and precise prevention and control of small green leafhopper with sound and light</article-title>. <source>China Tea Process.</source> <volume>3</volume>, <fpage>28</fpage>&#x2013;<lpage>33</lpage>. doi: <pub-id pub-id-type="doi">10.15905/j.cnki.33-1157/ts.2022.03.005</pub-id></mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xu</surname><given-names>W.</given-names></name> <name><surname>Xu</surname><given-names>T.</given-names></name> <name><surname>Thomasson</surname><given-names>J. A.</given-names></name> <name><surname>Chen</surname><given-names>W.</given-names></name> <name><surname>Karthikeyan</surname><given-names>R.</given-names></name> <name><surname>Tian</surname><given-names>G.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>A lightweight SSV2-YOLO based model for detection of sugarcane aphids in unstructured natural environments</article-title>. <source>Comput. Electron. Agric.</source> <volume>211</volume>:<fpage>107961</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2023.107961</pub-id></mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>H.</given-names></name></person-group> (<year>2023</year>). <source>Research on the identification method of clivia shell worm pests based on deep learning</source>. <publisher-loc>China</publisher-loc>: <publisher-name>Guizhou University</publisher-name>.</mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yang</surname><given-names>S.</given-names></name> <name><surname>Xing</surname><given-names>Z.</given-names></name> <name><surname>Wang</surname><given-names>H.</given-names></name> <name><surname>Dong</surname><given-names>X.</given-names></name> <name><surname>Gao</surname><given-names>X.</given-names></name> <name><surname>Liu</surname><given-names>Z.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Maize-YOLO: a new high-precision and real-time method for maize pest detection</article-title>. <source>Insects</source> <volume>14</volume>:<fpage>278</fpage>. doi: <pub-id pub-id-type="doi">10.3390/insects14030278</pub-id>, <pub-id pub-id-type="pmid">36975962</pub-id></mixed-citation></ref>
<ref id="ref26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>L.</given-names></name> <name><surname>Ding</surname><given-names>G.</given-names></name> <name><surname>Li</surname><given-names>C.</given-names></name> <name><surname>Li</surname><given-names>D.</given-names></name></person-group> (<year>2023</year>). <article-title>DCF-Yolov8: an improved algorithm for aggregating low-level features to detect agricultural pests and diseases</article-title>. <source>Agronomy</source> <volume>13</volume>. doi: <pub-id pub-id-type="doi">10.3390/agronomy13082012</pub-id></mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname><given-names>Y. P.</given-names></name> <name><surname>Pan</surname><given-names>Z. P.</given-names></name> <name><surname>Yu</surname><given-names>X. Q.</given-names></name> <name><surname>Huang</surname><given-names>S. H.</given-names></name> <name><surname>Li</surname><given-names>C. Y.</given-names></name> <name><surname>Liao</surname><given-names>Z. X.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Research progress on biology, ecology and control of <italic>Basilepta melanopus</italic> (Coleoptera: Chrysomelidae)</article-title>. <source>J. Environ. Entomol.</source>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. Available online at: <ext-link xlink:href="https://link.cnki.net/urlid/44.1640.q.20250314.1714.004" ext-link-type="uri">https://link.cnki.net/urlid/44.1640.q.20250314.1714.004</ext-link></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname><given-names>X.</given-names></name> <name><surname>Nie</surname><given-names>H.</given-names></name> <name><surname>Zhou</surname><given-names>X.</given-names></name></person-group> (<year>2022</year>). <article-title>Pest recognition method based on TPH-YOLOv5 and small sample learning</article-title>. <source>Computer Science</source> <volume>49</volume>, <fpage>257</fpage>&#x2013;<lpage>263</lpage>. doi: <pub-id pub-id-type="doi">10.11896/jsjkx.221000203</pub-id></mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhu</surname><given-names>R.</given-names></name> <name><surname>Zhang</surname><given-names>J. Y.</given-names></name> <name><surname>Huang</surname><given-names>J. C.</given-names></name> <name><surname>Kang</surname><given-names>R.</given-names></name> <name><surname>Chen</surname><given-names>K. J.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Research progress of crop leaf disease detection based on convolutional neural network</article-title>. <source>Trans. Chin. Soc. Agric. Eng.</source><comment>(in Chinese with English abstract)</comment> <volume>41</volume>, <fpage>15</fpage>&#x2013;<lpage>28</lpage>. doi: <pub-id pub-id-type="doi">10.11975/j.issn.1002-6819.202502117</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0002">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2995372/overview">Tongcheng Fu</ext-link>, Hunan Agricultural University, China</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0003">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/455919/overview">Chaochen Tang</ext-link>, Guangdong Academy of Agricultural Sciences (GDAAS), China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3260428/overview">Feng Chen</ext-link>, Anhui Agricultural University, China</p>
</fn>
</fn-group>
</back>
</article>