<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2024.1468676</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Plant Science</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Integrating AI detection and language models for real-time pest management in Tomato cultivation</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes" corresp="yes">
<name>
<surname>&#x15e;ahin</surname>
<given-names>Yavuz Selim</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<xref ref-type="author-notes" rid="fn003">
<sup>&#x2020;</sup>
</xref>
<xref ref-type="author-notes" rid="fn004">
<sup>&#x2021;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2798284"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>Gen&#xe7;er</surname>
<given-names>Nimet Sema</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn003">
<sup>&#x2020;</sup>
</xref>
<xref ref-type="author-notes" rid="fn004">
<sup>&#x2021;</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name>
<surname>&#x15e;ahin</surname>
<given-names>Hasan</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn003">
<sup>&#x2020;</sup>
</xref>
<xref ref-type="author-notes" rid="fn004">
<sup>&#x2021;</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2798376"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/visualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Bursa Uluda&#x11f; University, Faculty of Agriculture, Department of Plant Protection</institution>, <addr-line>Bursa</addr-line>, <country>T&#xfc;rkiye</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Bursa Technical University, Faculty of Engineering and Natural Sciences, Department of Industrial Engineering</institution>, <addr-line>Bursa</addr-line>, <country>T&#xfc;rkiye</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Yongliang Qiao, University of Adelaide, Australia</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Yinyan Shi, Nanjing Agricultural University, China</p>
<p>Jian Zhang, University of British Columbia, Canada</p>
<p>Elio Romano, Centro di ricerca per l&#x2019;Ingegneria e le Trasformazioni agroalimentari (CREA-IT), Italy</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Yavuz Selim &#x15e;ahin, <email xlink:href="mailto:yavuzselimsahin@uludag.edu.tr">yavuzselimsahin@uludag.edu.tr</email>
</p>
</fn>
<fn fn-type="equal" id="fn003">
<p>&#x2020;These authors have contributed equally to this work</p>
</fn>
<fn fn-type="other" id="fn004">
<p>&#x2021;ORCID: Yavuz Selim &#x15e;ahin, <uri xlink:href="https://orcid.org/0000-0001-6848-1849">orcid.org/0000-0001-6848-1849</uri>; Nimet Sema Gen&#xe7;er, <uri xlink:href="https://orcid.org/0009-0007-2435-2384">orcid.org/0009-0007-2435-2384</uri>; Hasan &#x15e;ahin, <uri xlink:href="https://orcid.org/0000-0002-8915-000X">orcid.org/0000-0002-8915-000X</uri>
</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>21</day>
<month>02</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>15</volume>
<elocation-id>1468676</elocation-id>
<history>
<date date-type="received">
<day>22</day>
<month>07</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>10</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2025 &#x15e;ahin, Gen&#xe7;er and &#x15e;ahin</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>&#x15e;ahin, Gen&#xe7;er and &#x15e;ahin</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Tomato (<italic>Solanum lycopersicum</italic> L.) cultivation is crucial globally due to its nutritional and economic value. However, the crop faces significant threats from various pests, including <italic>Tuta absoluta</italic>, <italic>Helicoverpa armigera</italic>, and <italic>Leptinotarsa decemlineata</italic>, among others. These pests not only reduce yield but also increase production costs due to the heavy reliance on pesticides. Traditional pest detection methods are labor-intensive and prone to errors, necessitating the exploration of advanced techniques. This study aims to enhance pest detection in tomato cultivation using AI-based detection and language models. Specifically, it integrates YOLOv8 for detection and segmentation tasks and ChatGPT-4 for generating detailed, actionable insights on the detected pests. YOLOv8 was chosen for its superior performance in agricultural pest detection, capable of processing large volumes of data in real-time with high accuracy. The methodology involved training the YOLOv8 model with images of various pests and plant damage. The model achieved a precision of 98.91%, recall of 98.98%, mAP50 of 98.75%, and mAP50-95 of 97.72% for detection tasks. For segmentation tasks, precision was 97.47%, recall 98.81%, mAP50 99.38%, and mAP50-95 95.99%. These metrics demonstrate significant improvements over traditional methods, indicating the model&#x2019;s effectiveness. The integration of ChatGPT-4 further enhances the system by providing detailed explanations and recommendations based on detected pests. This approach facilitates real-time expert consultation, making pest management accessible to untrained producers, especially in remote areas. The study&#x2019;s results underscore the potential of combining AI-based detection and language models to revolutionize agricultural practices. Future research should focus on training these models with domain-specific data to improve accuracy and reliability. Additionally, addressing the computational limitations of personal devices will be crucial for broader adoption. This integration promises to democratize information access, promoting a more resilient, informed, and environmentally conscious approach to farming.</p>
</abstract>
<kwd-group>
<kwd>pest detection</kwd>
<kwd>precision agriculture</kwd>
<kwd>ChatGPT</kwd>
<kwd>YOLOv8</kwd>
<kwd>sustainable agriculture</kwd>
</kwd-group>
<counts>
<fig-count count="7"/>
<table-count count="4"/>
<equation-count count="0"/>
<ref-count count="54"/>
<page-count count="11"/>
<word-count count="4180"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Sustainable and Intelligent Phytoprotection</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<title>Introduction</title>
<p>Tomato (<italic>Solanum lycopersicum</italic> L.) is a globally significant vegetable crop, essential for both nutritional value and economic stability. However, tomato cultivation faces substantial threats from various pests. Key pests include <italic>Tuta absoluta</italic> (Lepidoptera: Gelechiidae), which has a significant socioeconomic impact in Eastern Africa due to its widespread distribution and increased costs and pesticide use among farmers (<xref ref-type="bibr" rid="B32">Pereyra and S&#xe1;nchez, 2006</xref>; <xref ref-type="bibr" rid="B42">Shaltiel-Harpaz et al., 2015</xref>; <xref ref-type="bibr" rid="B1">Aigbedion-Atalor et&#xa0;al., 2019</xref>). <italic>Helicoverpa armigera</italic> (Lepidoptera: Noctuidae) is another critical pest, highlighting the low adoption of biological control measures and underscoring the need for improved farmer knowledge and extension programs (<xref ref-type="bibr" rid="B5">Balipoor and Ommani, 2014</xref>). <italic>Leptinotarsa decemlineata</italic> (Coleoptera: Chrysomelidae) and <italic>Bemisia tabaci</italic> (Hemiptera: Aleyrodidae) also pose substantial threats. <italic>Myzus persicae</italic> (Hemiptera: Aphididae), along with <italic>Dolycoris baccarum</italic> (Hemiptera: Pentatomidae), <italic>Phyllotreta</italic> spp. (Coleoptera: Chrysomelidae), and <italic>Nezara viridula</italic> (Hemiptera: Pentatomidae), further complicate tomato cultivation. Additionally, <italic>Tetranychus urticae</italic> (Trombidiformes: Tetranychidae) is a significant allergen, particularly among greenhouse workers and asthmatics living near orchards (<xref ref-type="bibr" rid="B24">Jee et&#xa0;al., 2000</xref>). <italic>Frankliniella occidentalis</italic> (Thysanoptera: Thripidae) is another pest impacting tomato crops. Insecticide use patterns among tomato farmers in Ghana reveal a mix of recommended and non-recommended, persistent insecticides, highlighting the need for better regulation and education (<xref ref-type="bibr" rid="B13">Danquah et&#xa0;al., 2010</xref>).</p>
<p>Efficient and timely identification of pests is essential for maintaining crop health and optimizing yield. Traditionally, this process has relied heavily on human observation, which is labor-intensive, time-consuming, and susceptible to errors (<xref ref-type="bibr" rid="B13">Danquah et&#xa0;al., 2010</xref>). Artificial Intelligence (AI) models, which use algorithms and computational power to simulate human intelligence, offer a promising alternative. There are various types of AI models for data processing: some models process images by converting them into matrices (detection models), while others process text by converting characters or tokens into vectors (language models) (<xref ref-type="bibr" rid="B48">Vaswani et&#xa0;al., 2017</xref>). Detection models, such as Mask R-CNN, Faster R-CNN, SSD, and YOLO (You Only Look Once), provide rapid and accurate pest detection, significantly reducing the need for manual labor and enhancing precision. They are capable of processing large volumes of data in real-time, thereby greatly improving agricultural efficiency and sustainability (<xref ref-type="bibr" rid="B29">Liu and Wang, 2020</xref>; <xref ref-type="bibr" rid="B45">Swinburne et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B25">Jin et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B49">Wen et al., 2022</xref>; <xref ref-type="bibr" rid="B34">Rajamohanan and Latha, 2023</xref>).</p>
<p>YOLO excels due to its real-time processing, high detection accuracy, and versatility in both detection and segmentation tasks. Unlike traditional AI-based pest management systems, this study introduces a novel integration of real-time detection with YOLOv8 and language-based decision support via ChatGPT-4, offering both precision in pest detection and actionable, context-specific recommendations for farmers. This combination allows not only for accurate detection but also for informed decision-making, making the system accessible and practical for real-world agricultural applications. By reducing the reliance on manual expertise and providing timely insights, this system improves both the efficiency and sustainability of pest management practices. It is particularly effective for detecting small, densely packed objects like agricultural pests, making it ideal for real-time applications (<xref ref-type="bibr" rid="B37">Redmon et&#xa0;al., 2016</xref>). Its adaptability to various scales and high mean Average Precision (mAP) scores further justify its use in training and detecting agricultural pests, effectively managing multiple pest species with diverse morphologies (<xref ref-type="bibr" rid="B50">Yang et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B20">Hashimoto et&#xa0;al., 2020</xref>). These features make YOLO an excellent choice for pest detection and segmentation in this study.</p>
<p>However, while detection models like YOLO have the potential to analyze pests more accurately and quickly than humans, they lack the capability to interpret the findings and provide actionable recommendations to farmers (<xref ref-type="bibr" rid="B45">Swinburne et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B25">Jin et&#xa0;al., 2022</xref>). This gap, which requires knowledge and experience, can be filled by language models. Language models, like ChatGPT, are a type of AI designed to understand and generate human language. They process data by converting characters into vectors, which allows the model to recognize and predict patterns in text (<xref ref-type="bibr" rid="B48">Vaswani et&#xa0;al., 2017</xref>). ChatGPT-4, developed by OpenAI, was trained on approximately 1.3 trillion tokens, providing it with a vast knowledge base (<xref ref-type="bibr" rid="B35">Rao et&#xa0;al., 2023</xref>). Therefore, while YOLO is used for accurate and real-time pest detection, ChatGPT was chosen as the language model for this study due to its extensive training and ability to generate relevant, insightful responses to interpret the detected tomato pests.</p>
<p>Accurate and real-time identification of agricultural pests necessitates education, knowledge, and experience (<xref ref-type="bibr" rid="B45">Swinburne et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B13">Danquah et&#xa0;al., 2010</xref>). Once pests are detected, it is essential to have detailed information about them to devise effective management strategies (<xref ref-type="bibr" rid="B5">Balipoor and Ommani, 2014</xref>; <xref ref-type="bibr" rid="B1">Aigbedion-Atalor et&#xa0;al., 2019</xref>). Accessing this information can be time-consuming and costly. However, language models can provide detailed commentary on detected pests in agricultural applications, thus informing farmers who may lack expertise. By facilitating access to accurate information and analyzing large datasets more quickly than humans, these models can save time and costs while enhancing the quality of education. In this study, detection models and language models are integrated through an API (Application Programming Interface, a set of rules and protocols for building and interacting with software applications, allowing different systems to communicate and share data) to analyze and interpret pest data, providing a valuable guide for future similar research endeavors.</p>
</sec>
<sec id="s2" sec-type="materials|methods">
<title>Materials and methods</title>
<sec id="s2_1">
<title>Definition of the research area and dataset</title>
<p>Turkey is one of the top five tomato-producing countries in the world. About 10% of Turkey&#x2019;s tomato production occurs in Bursa, where tomatoes were the most produced vegetable in 2020, with 13.2 million tons (<xref ref-type="bibr" rid="B27">Kumbasaro&#x11f;lu et&#xa0;al., 2021</xref>). This study was conducted from March 2022 to September 2023 in Bakirk&#xf6;y village, located in the Karacabey district of Bursa province in the northwest of Turkey, lying between latitudes 40&#xb0;7&#x2019;17.53&#x201d;N and 40&#xb0;10&#x2019;40.36&#x201d;N and longitudes 28&#xb0;21&#x2019;14.12&#x201d;E and 28&#xb0;26&#x2019;2.37&#x201d;E. Field campaigns were conducted from June to July 2023. The site covers an area of 47.16 km&#xb2;, and a total of 96 tomato fields were investigated. The identification of pests observed in the field photographs was carried out according to the morphological diagnostic keys available in the literature (<xref ref-type="bibr" rid="B9">Blackman and Eastop, 2000</xref>; <xref ref-type="bibr" rid="B22">Hoebeke and Carter, 2003</xref>; <xref ref-type="bibr" rid="B14">Desneux et&#xa0;al., 2010</xref>; <xref ref-type="bibr" rid="B3">Ashbrook et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B28">Li et&#xa0;al., 2023</xref>).</p>
<p>Detection models excel in identifying the presence and location of pests quickly and efficiently (<xref ref-type="bibr" rid="B6">Barbedo, 2016</xref>). However, segmentation models are more suitable when detailed morphological features or comprehensive damage maps are necessary (<xref ref-type="bibr" rid="B2">Arockia et&#xa0;al., 2023</xref>). The YOLOv8 model integrates both detection and segmentation capabilities. In this study, the yolov8s.pt model was employed for detection tasks, while the yolov8n-seg.pt model was utilized for segmentation tasks. The images used for detection and segmentation in this study encompass various pests and damage types affecting tomato crops. These include <italic>Dolycoris baccarum</italic> (<italic>Hemiptera: Pentatomidae</italic>), <italic>Phyllotreta</italic> spp. (<italic>Coleoptera: Chrysomelidae</italic>), <italic>Nezara viridula (Hemiptera: Pentatomidae), Myzus persicae (Hemiptera: Aphididae), Bemisia tabaci (Hemiptera: Aleyrodidae), Leptinotarsa decemlineata (Coleoptera: Chrysomelidae), Tuta absoluta (Lepidoptera: Gelechiidae), Helicoverpa armigera (Lepidoptera: Noctuidae), Liriomyza bryoniae (Diptera: Agromyzidae) damage, Frankliniella occidentalis (Thysanoptera: Thripidae)</italic> damage, and <italic>Tetranychus urticae (Trombidiformes: Tetranychidae)</italic> damage. These pests and damage types were systematically photographed and used to train the YOLOv8 model for accurate detection and segmentation tasks, aiming to enhance the model&#x2019;s ability to identify and manage multiple pest species effectively.</p>
<p>From March 2023 to September 2024, high-resolution images of tomato plant diseases and pests were captured using a Canon EOS 700D camera with a resolution of 768 &#xd7; 1024 pixels. To ensure consistency in image quality, all photographs were taken using cameras set to identical resolution settings. Images were taken at distances of 1 meter and 0.2 meters from the leaves, from various angles (<xref ref-type="bibr" rid="B52">Yong et&#xa0;al., 2020</xref>). A comprehensive dataset of over 1,000 images was compiled for each pest, documenting different angles and features. These images were then divided into three subsets: 80% for training, 18% for validation, and 2% for testing, as outlined in <xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref>.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Distribution of the image dataset for model training.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="center">Method</th>
<th valign="top" align="center">Image Type</th>
<th valign="top" align="center">Total Images</th>
<th valign="top" align="center">Training %80</th>
<th valign="top" align="center">Validation %18</th>
<th valign="top" align="center">Test %2</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" rowspan="3" align="left">
<bold>Detection</bold>
</td>
<td valign="top" align="left">Pest adult images</td>
<td valign="top" align="center">7000</td>
<td valign="top" align="center">5600</td>
<td valign="top" align="center">1260</td>
<td valign="top" align="center">140</td>
</tr>
<tr>
<td valign="top" align="left">Pest nymph images</td>
<td valign="top" align="center">2000</td>
<td valign="top" align="center">1600</td>
<td valign="top" align="center">360</td>
<td valign="top" align="center">40</td>
</tr>
<tr>
<td valign="top" align="left">Pest larva images</td>
<td valign="top" align="center">2000</td>
<td valign="top" align="center">1600</td>
<td valign="top" align="center">360</td>
<td valign="top" align="center">40</td>
</tr>
<tr>
<td valign="top" colspan="2" align="left">
<bold>Total Images</bold>
</td>
<td valign="top" align="center">11000</td>
<td valign="bottom" align="center">8800</td>
<td valign="bottom" align="center">1980</td>
<td valign="bottom" align="center">220</td>
</tr>
<tr>
<td valign="top" rowspan="2" align="left">
<bold>Segmentation</bold>
</td>
<td valign="top" align="left">Tomato leaf images</td>
<td valign="top" align="center">5000</td>
<td valign="top" align="center">4000</td>
<td valign="top" align="center">900</td>
<td valign="top" align="center">100</td>
</tr>
<tr>
<td valign="top" align="left">Tomato fruit images</td>
<td valign="top" align="center">4000</td>
<td valign="top" align="center">3200</td>
<td valign="top" align="center">720</td>
<td valign="top" align="center">80</td>
</tr>
<tr>
<td valign="top" colspan="2" align="left">
<bold>Total Images</bold>
</td>
<td valign="top" align="center">9000</td>
<td valign="top" align="center">7200</td>
<td valign="top" align="center">1620</td>
<td valign="top" align="center">180</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s2_2">
<title>Data preprocessing techniques and applications</title>
<p>Data preprocessing refers to a series of steps undertaken to prepare raw data for analysis or modeling. It is commonly used in data mining, machine learning, statistics, and data analysis to address data deficiencies, noise, and inconsistencies, thereby enabling more effective analysis (<xref ref-type="bibr" rid="B41">&#x15e;ahin and Topal, 2016</xref>; <xref ref-type="bibr" rid="B4">Atalan et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B7">Ba&#x15f;t&#xfc;rk and &#x15e;ahin, 2022</xref>). For image processing models, preprocessing the dataset involves three steps: image labeling, resizing, and augmentation. Augmentation provides a large amount of training data to learn features and achieve accurate classification on unseen data, preventing issues like overfitting and poor generalization (<xref ref-type="bibr" rid="B38">Redmon and Farhadi, 2018</xref>; <xref ref-type="bibr" rid="B39">Rubanga et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B19">G&#xfc;ven and &#x15e;ahin, 2022</xref>).</p>
<p>In this study, Python (version 3.11.8) was used for data preprocessing due to its extensive library ecosystem. To enhance the quality of model training, the original images taken under field conditions were augmented using the OpenCV library. During augmentation, transformations were applied to each image, including rotation, cropping, flipping, adding noise, adjusting lighting, and zooming out (<xref ref-type="bibr" rid="B43">Shorten and Khoshgoftaar, 2019</xref>). All images were resized to 600 &#xd7; 600 pixels as required for model training (<xref ref-type="bibr" rid="B21">He et&#xa0;al., 2017</xref>). The analyses were conducted in the Spyder IDE, part of the Anaconda distribution, which offers various libraries for scientific computing and data science (<xref ref-type="bibr" rid="B40">&#x15e;ahin et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B51">Y&#x131;lmaz et&#xa0;al., 2021</xref>). Prior to any augmentation, the dataset was divided into training, validation, and test sets (80%, 18%, and 2%, respectively) to ensure that no data leakage occurred during the augmentation process. Augmentation was only applied to the training set to avoid introducing artificial examples into the validation and test sets, which could result in overly optimistic performance estimates (<xref ref-type="bibr" rid="B43">Shorten &amp; Khoshgoftaar, 2019</xref>). Specifically, transformations such as rotation, cropping, flipping, adding noise, adjusting lighting, and zooming out were applied only to the training data after the initial dataset split. Labeling was performed using the LabelMe tool (<ext-link ext-link-type="uri" xlink:href="https://github.com/wkentaro/labelme">https://github.com/wkentaro/labelme</ext-link>), with two main approaches: pixel-based segmentation for precise boundary definitions and rectangular bounding for approximate location and size.</p>
</sec>
<sec id="s2_3">
<title>Model setup and training</title>
<p>YOLOv8 was selected for this study due to its superior speed and efficiency compared to slower yet more accurate models like Faster R-CNN, making it particularly well-suited for real-time agricultural pest detection, where timely decisions are essential for effective pest management (<xref ref-type="bibr" rid="B46">Tang et&#xa0;al., 2021</xref>). The YOLOv8 models were trained using the ultralytics library for model loading and training, and the google.colab library for accessing the dataset via Google Drive. Training parameters included over 100,000 epochs (with patience set to 50 to prevent overfitting), a batch size of 16, and an image size of 640 (<xref ref-type="table" rid="T2">
<bold>Table&#xa0;2</bold>
</xref>). The model&#x2019;s hyperparameters, including the number of epochs, batch size, and learning rate, were optimized through an iterative process. Early stopping (patience) was employed to prevent overfitting, while cross-validation was used to fine-tune the learning rate and batch size. The optimal values for these parameters were selected based on the model&#x2019;s performance on the validation set, ensuring robustness and preventing overfitting. Training was conducted on Google Colaboratory, utilizing an Intel Xeon CPU, 12.68 GB RAM, and a Tesla K80 GPU. Both detection and segmentation models were trained in Python on a custom dataset. Instance segmentation models were chosen to precisely identify damage caused by multiple pest species on tomato plants, which is crucial for accurately identifying specific damages on leaves and fruits (<xref ref-type="bibr" rid="B31">Mirhaji et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B53">Zhang et&#xa0;al., 2023a</xref>, <xref ref-type="bibr" rid="B54">2023</xref>). The YOLO framework used in this study is illustrated in <xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1</bold>
</xref>.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Key parameters were set in Google Colab for the training of the Ultralytics YOLOv8.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="center">Task</th>
<th valign="top" align="center">Mode</th>
<th valign="top" align="center">model</th>
<th valign="top" align="center">epochs</th>
<th valign="top" align="center">batch</th>
<th valign="top" align="center">imgsz</th>
<th valign="top" align="center">patience</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">segment</td>
<td valign="top" align="center">train</td>
<td valign="top" align="center">yolov8s.pt</td>
<td valign="top" align="center">10000</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">640</td>
<td valign="top" align="center">50</td>
</tr>
<tr>
<td valign="top" align="center">detect</td>
<td valign="top" align="center">train</td>
<td valign="top" align="center">yolov8n-seg.pt</td>
<td valign="top" align="center">10000</td>
<td valign="top" align="center">16</td>
<td valign="top" align="center">640</td>
<td valign="top" align="center">50</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>YOLO Framework used in this study. Input: Raw images fed into the model. Backbone: Extracts features from images. Neck: Combines and enhances features. Prediction: Predicts pests&#x2019; presence and location. Output: Provides detection and segmentation results.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g001.tif"/>
</fig>
</sec>
<sec id="s2_4">
<title>Model evaluation methodology and testing process</title>
<p>The model&#x2019;s generalization capability was assessed using a pre-allocated dataset: 80% for training, 18% for validation, and 2% for testing. Key performance metrics, including precision, recall, mAP50, and mAP50-95, were calculated during training on both training and validation datasets. Detection and segmentation performances were evaluated at various IOU thresholds using precision (P), recall (R), and mean average precision (mAP). The mAP50 metric refers to the mean average precision at a 50% IOU threshold, indicating the accuracy of the model in identifying objects with at least 50% overlap with ground truth labels. On the other hand, mAP50-95 averages precision over IOU thresholds from 50% to 95%, offering a more comprehensive view of the model&#x2019;s accuracy across different overlap scenarios, which is particularly important in agricultural contexts where pests may be occluded or vary in size (<xref ref-type="bibr" rid="B28">Li et&#xa0;al., 2023</xref>). These metrics provide insights into the model&#x2019;s ability to handle various sizes and overlaps in real-world agricultural environments. Loss metrics&#x2014;box_loss, cls_loss, and dfl_loss&#x2014;were analyzed to identify areas for improvement. The confusion matrix summarized predictions across classes, highlighting correct and incorrect classifications. This comprehensive analysis provided a clear understanding of the model&#x2019;s strengths and weaknesses. Metrics P, R, mAP50, and mAP50-95 are defined by <xref ref-type="table" rid="T3">
<bold>Table&#xa0;3</bold>
</xref>.</p>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Formulas of key performance metrics for evaluating YOLO models in object detection.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="left">Performance metrics</th>
<th valign="top" align="left">Formula</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Precision (P)</td>
<td valign="top" align="left">
<inline-formula>
<mml:math display="inline" id="im1">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td valign="top" align="left">Recall (R)</td>
<td valign="top" align="left">
<inline-formula>
<mml:math display="inline" id="im2">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mi>P</mml:mi>
<mml:mo>+</mml:mo>
<mml:mi>F</mml:mi>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td valign="top" align="left">mAP50</td>
<td valign="top" align="left">
<inline-formula>
<mml:math display="inline" id="im3">
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>Q</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>Q</mml:mi>
</mml:msubsup>
<mml:mi>P</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>Rq</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td valign="top" align="left">mAP50-95</td>
<td valign="top" align="left">
<inline-formula>
<mml:math display="inline" id="im4">
<mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>Q</mml:mi>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>95</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mi>P</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>q</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:msubsup>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>q</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:mi>P</mml:mi>
</mml:mstyle>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>R</mml:mi>
<mml:mi>q</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>TP (True Positives): The count of correct positive predictions. FP (False Positives): The count of incorrect positive predictions (actual negatives predicted as positives). FN (False Negatives): The count of incorrect negative predictions (actual positives predicted as negatives). Q: Number of query points. P(Rq): Interpolated precision at recall level Rq.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="s2_5">
<title>Prompt creation and OpenAI GPT-4 integration</title>
<p>The study used Python and open-source libraries to integrate detection models with OpenAI&#x2019;s GPT-4 via an API key. Initially, models identified trained objects, which were then linked to GPT-4. A good prompt should be clear, specific, and provide context to guide the AI&#x2019;s response. Labels were defined as &#x2018;det_labels_str&#x2019; and &#x2018;seg_labels_str&#x2019;. The prompt used in the study was: prompt_str = f&#x201d;Could you provide a detailed explanation, in academic English, on the methods for controlling {det_labels_str} or {seg_labels_str} and the potential damage they can inflict on related plants, including preventative measures and integrated pest management strategies?&#x201d;. Text outputs, limited to 250-400 tokens, were visualized with detection results. A ten-step coding sequence enabled the simultaneous operation of segmentation and detection models (<xref ref-type="table" rid="T4">
<bold>Table&#xa0;4</bold>
</xref>). The workflow, from image capture to ChatGPT-4 output, is depicted in <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2</bold>
</xref>.</p>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Integration of Ultralytics YOLO and OpenAI GPT-4 using API key.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" colspan="2" align="center">Steps</th>
<th valign="top" align="center">Description</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="center">1</td>
<td valign="top" align="center">Model Loading</td>
<td valign="top" align="left">Pre-trained segmentation and detection models are loaded using the YOLO framework.</td>
</tr>
<tr>
<td valign="top" align="center">2</td>
<td valign="top" align="center">Image Loading</td>
<td valign="top" align="left">The image to be analysed has been uploaded.</td>
</tr>
<tr>
<td valign="top" align="center">3</td>
<td valign="top" align="center">Segmentation</td>
<td valign="top" align="left">The segmentation model is run on the image, and the predicted masks are obtained.</td>
</tr>
<tr>
<td valign="top" align="center">4</td>
<td valign="top" align="center">Application of Masks</td>
<td valign="top" align="left">The predicted masks are applied to the original image to identify specific areas.</td>
</tr>
<tr>
<td valign="top" align="center">5</td>
<td valign="top" align="center">Detection</td>
<td valign="top" align="left">The detection model is executed on the combined image, predicting bounding boxes and class names for objects.</td>
</tr>
<tr>
<td valign="top" align="center">6</td>
<td valign="top" align="center">Visualization of Results</td>
<td valign="top" align="left">Both segmentation and detection outcomes are visualized and saved for further analysis.</td>
</tr>
<tr>
<td valign="top" align="center">7</td>
<td valign="top" align="center">Labelling</td>
<td valign="top" align="left">The segmentation and detection results are converted into labels and formatted as strings.</td>
</tr>
<tr>
<td valign="top" align="center">8</td>
<td valign="top" align="center">Integration with Natural Language Processing</td>
<td valign="top" align="left">By connecting to OpenAI&#x2019;s GPT-4 through its API, a question is formulated regarding the detected labels, and a response is subsequently obtained.</td>
</tr>
<tr>
<td valign="top" align="center">9</td>
<td valign="top" align="center">Prompt Creation for GPT-4</td>
<td valign="top" align="left">prompt_str = f&#x201d;Could you provide a detailed explanation, in academic English, on the methods for controlling {det_labels_str} or {seg_labels_str} and the potential damage they can inflict on related plants, including preventative measures and integrated pest management strategies?&#x201d;</td>
</tr>
<tr>
<td valign="top" align="center">10</td>
<td valign="top" align="center">Combination and Visualization of Results</td>
<td valign="top" align="left">The detection results and the text response from GPT-4 are visualized and saved.</td>
</tr>
</tbody>
</table>
</table-wrap>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>ChatGPT-4 integration process.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g002.tif"/>
</fig>
</sec>
</sec>
<sec id="s3" sec-type="results">
<title>Results</title>
<sec id="s3_1">
<title>Training and validation loss graphs</title>
<p>In this study, significant improvements were observed during YOLOv8 model training for pest detection. For training metrics, the box_loss decreased from 1.84 to 0.54, cls_loss from 3.48 to 0.37, and dfl_loss from 1.54 to 0.86. Similarly, validation metrics showed a decline: val/box_loss reduced from 1.38 to 0.53, val/cls_loss from 3.45 to 0.31, and val/dfl_loss from 1.30 to 0.90. Training was halted at 749 epochs to prevent overfitting, demonstrating effective learning and performance. For segmentation, the train/box_loss decreased from 1.97 to 0.57, train/cls_loss from 4.04 to 0.37, and train/dfl_loss from 1.56 to 0.86, while validation metrics also improved, with val/box_loss reducing from 1.99 to 0.74, val/cls_loss from 2.65 to 0.37, and val/dfl_loss from 1.45 to 0.92. Training stopped at 372 epochs to avoid overfitting, indicating robust model performance (<xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>).</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Changes in training and validation loss values over epochs for the YOLOv8 model trained for pest detection and segmentation.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g003.tif"/>
</fig>
</sec>
<sec id="s3_2">
<title>Performance evaluation metrics</title>
<p>During YOLOv8 model training, significant improvements were noted across key metrics: precision increased from 0% to 98.91%, recall from 0% to 98.98%, mAP50 from 0% to 98.75%, and mAP50-95 from 0% to 97.72%. Training was halted at 749 epochs to prevent overfitting, demonstrating enhanced accuracy and reliability in object detection (<xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4A</bold>
</xref>). For segmentation, precision improved from 0% to 97.47%, recall from 0% to 98.81%, mAP50 from 0% to 99.38%, and mAP50-95 from 0% to 95.99%, with training stopping at 372 epochs to avoid overfitting (<xref ref-type="fig" rid="f4">
<bold>Figure&#xa0;4B</bold>
</xref>).</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Improvements in performance metrics during YOLOv8 model training for object detection <bold>(A)</bold> and object segmentation <bold>(B)</bold>.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g004.tif"/>
</fig>
</sec>
<sec id="s3_3">
<title>Confusion matrix analysis</title>
<p>A confusion matrix, essential for evaluating a model&#x2019;s performance, pinpoints misclassifications and highlights areas for potential improvement. The test set, comprising images of adult insects, nymphs, and larvae across 11 classes, facilitated the computation of the confusion matrix. The YOLOv8 model exhibited high accuracy in detection tasks. Specifically, <italic>D. baccarum</italic> adults were correctly classified 890 times with 15 misclassifications, <italic>N. viridula</italic> adults were accurately identified 695 times with 15 errors, and <italic>M. persicae</italic> adults were correctly classified 998 times. Additionally, <italic>B. tabaci</italic> adults achieved 1025 correct identifications, and <italic>L. decemlineata</italic> adults were correctly identified 666 times with no errors (<xref ref-type="fig" rid="f5">
<bold>Figure&#xa0;5</bold>
</xref>). The model&#x2019;s performance for segmentation on the test set revealed notable outcomes across 8 classes: 550 correct detections of <italic>L. bryoniae</italic> damage, 345 accurate detections of <italic>T. absoluta</italic> damage on fruit, 450 precise detections of <italic>T. urticae</italic> damage on leaves, and 565 correct identifications of healthy tomato leaves (<xref ref-type="fig" rid="f6">
<bold>Figure&#xa0;6</bold>
</xref>).</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Confusion matrix illustrating YOLOv8 model&#x2019;s performance in pest detection across 11 classes.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g005.tif"/>
</fig>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Confusion matrix illustrating YOLOv8 model&#x2019;s performance in pest segmentation across eight classes.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g006.tif"/>
</fig>
</sec>
<sec id="s3_4">
<title>Prompt creation and real-time textual response to visual data</title>
<p>The integration of YOLOv8 and ChatGPT-4 showcases the powerful combination of computer vision and natural language processing, enabling expert feedback on visual data. This integration was tested on five pictures from the test set, which were not included during the training phase. The responses to the crafted prompt, [prompt_str = f&#x201d;Could you provide a detailed explanation, in academic English, on the methods for controlling {det_labels_str} or {seg_labels_str} and the potential damage they can inflict on related plants, including preventative measures and integrated pest management strategies?&#x201d;] are presented in <xref ref-type="fig" rid="f7">
<bold>Figure&#xa0;7</bold>
</xref>. The trained detection and segmentation models processed the test images in approximately 0.10 seconds, while the integration with ChatGPT-4 provided textual responses within 3.5 seconds via the API. Despite being limited to 250-400 tokens, the ChatGPT-4 responses, while not always fully comprehensive, demonstrated the potential to offer key information.</p>
<fig id="f7" position="float">
<label>Figure&#xa0;7</label>
<caption>
<p>Feedback from ChatGPT-4 based on object labels detected by YOLOv8.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-15-1468676-g007.tif"/>
</fig>
</sec>
</sec>
<sec id="s4" sec-type="discussion">
<title>Discussion</title>
<p>In the literature, numerous detection models such as Mask R-CNN, SSD, Detectron, and MobileNet are capable of identifying objects in photographs using image processing techniques (<xref ref-type="bibr" rid="B21">He et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B30">Liu et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B17">Girshick et&#xa0;al., 2014</xref>; <xref ref-type="bibr" rid="B23">Howard, 2017</xref>). However, among these models, YOLOv8 is preferred in this study due to its superior performance in agricultural pest detection (<xref ref-type="bibr" rid="B37">Redmon et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B10">Bochkovskiy et&#xa0;al., 2020</xref>). These superior results can be attributed to several factors, including the large and diverse dataset used for training, YOLOv8&#x2019;s advanced architecture which allows for real-time processing with high accuracy, and the application of optimized hyperparameters specific to agricultural pest detection. For the detection task of the YOLOv8, precision increased to 98.91%, recall to 98.98%, mAP50 to 98.75%, and mAP50-95 to 97.72%. For segmentation tasks, precision increased to 97.47%, recall to 98.81%, mAP50 to 99.38%, and mAP50-95 to 95.99%. These results are consistent with other studies, such as the Pest-YOLO model achieving 69.59% mAP and 77.71% recall, and another study using YOLOv8 for small pest detection in field crops reporting an mAP of 84.7% (<xref ref-type="bibr" rid="B26">Khalid et&#xa0;al., 2023</xref>). Additionally, a study on pest detection in strawberries using segmented image datasets achieved a pest detection rate of 91.93% and detection reliability of 83.41% (<xref ref-type="bibr" rid="B12">Choi et&#xa0;al., 2022</xref>).</p>
<p>The integration of AI-based detection models with language models like ChatGPT offers significant benefits in pest detection and environmentally friendly pest control (<xref ref-type="bibr" rid="B18">Gu et&#xa0;al., 2021</xref>). Traditional methods, such as literature reviews, are resource-intensive, whereas language models provide rapid interpretations within 3.5 seconds, as demonstrated in this research. Researchers emphasize ChatGPT&#x2019;s potential to train producers and improve information access (<xref ref-type="bibr" rid="B36">Ray, 2023</xref>; <xref ref-type="bibr" rid="B44">Siche and Siche, 2023</xref>). However, challenges exist regarding output accuracy, which depends on the training data (<xref ref-type="bibr" rid="B16">Gaddikeri et&#xa0;al., 2023</xref>). Inaccurate training data can compromise response precision, highlighting the need for training with credible sources.</p>
<p>Open-source language models like LLAMA (Meta) (<xref ref-type="bibr" rid="B47">Touvron et&#xa0;al., 2023</xref>), GPT-Neo and GPT-J (EleutherAI) (<xref ref-type="bibr" rid="B8">Black et&#xa0;al., 2022</xref>), BERT (Hugging Face) (<xref ref-type="bibr" rid="B15">Devlin et&#xa0;al., 2019</xref>), and GPT-2 (OpenAI) (<xref ref-type="bibr" rid="B33">Radford et&#xa0;al., 2019</xref>) allow for training on local computers with specific, reliable, and targeted datasets. Nevertheless, even with these advanced models, their effectiveness is contingent upon the quality of the input data and their ability to generalize across diverse agricultural environments. Furthermore, the computational power of personal computers may be insufficient for effectively using these models (<xref ref-type="bibr" rid="B11">Brown et&#xa0;al., 2020</xref>). While this study employed the YOLOv8 model integrated with the broadly-informed GPT-4 via an API, utilizing models trained with domain-specific, reliable data could enhance the accuracy and reliability of outputs. Future work should focus on training with domain-specific, trustworthy sources to improve accuracy and applicability across various sectors.</p>
</sec>
<sec id="s5" sec-type="conclusion">
<title>Conclusion</title>
<p>The integration of AI-based detection and language models in this study demonstrates a significant advancement in agricultural practices. By embedding these models into common devices like smartphones, even untrained producers can access real-time expert consultation, enabling immediate pest detection and sustainable pest control. This technology holds the potential to revolutionize agriculture, particularly in remote areas, by reducing costs and facilitating integration with unmanned vehicles for continuous monitoring.</p>
<p>The study&#x2019;s results, showing substantial improvements in detection and segmentation precision, recall, and mAP metrics, underscore the efficacy of YOLOv8 in agricultural applications. Additionally, integrating language models like ChatGPT enhances the system&#x2019;s capability by providing detailed explanations and recommendations based on detected pests. This combination allows for rapid, informed decision-making, improving pest management strategies.</p>
<p>Future work should focus on training these models with domain-specific, reliable data to further enhance their accuracy and applicability. Moreover, addressing the computational limitations of personal devices for running advanced models will be crucial for broader adoption. To fully realize the potential of this technology in low-income and remote agricultural settings, future work should focus on the development of energy-efficient models that can run on low-power devices and operate under limited connectivity conditions. Additionally, partnerships with local agricultural cooperatives could facilitate the dissemination and training required for widespread adoption. Ultimately, this integration promises to democratize information access, promoting a more resilient, informed, and environmentally conscious approach to farming.</p>
</sec>
</body>
<back>
<sec id="s6" sec-type="data-availability">
<title>Data availability statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.</p>
</sec>
<sec id="s7" sec-type="author-contributions">
<title>Author contributions</title>
<p>Y&#x15e;: Writing &#x2013; review &amp; editing, Writing &#x2013; original draft. NG: Writing &#x2013; original draft, Visualization, Validation, Supervision, Software, Resources, Project administration, Methodology, Investigation, Funding acquisition, Formal analysis, Data curation, Conceptualization. H&#x15e;: Writing &#x2013; review &amp; editing, Visualization, Validation, Supervision, Software, Resources, Project administration, Methodology, Investigation, Funding acquisition, Formal analysis, Data curation, Conceptualization.</p>
</sec>
<sec id="s8" sec-type="funding-information">
<title>Funding</title>
<p>The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.</p>
</sec>
<ack>
<title>Acknowledgments</title>
<p>We extend our deepest gratitude to Necmettin Kuruhan for graciously allowing us to use his land for the study, and to undergraduate students Do&#x11f;ukan Akkanat and Emre &#xc7;ak&#x131;r for their invaluable assistance in data collection.</p>
</ack>
<sec id="s9" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s10" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aigbedion-Atalor</surname> <given-names>P. O.</given-names>
</name>
<name>
<surname>Hill</surname> <given-names>M. P.</given-names>
</name>
<name>
<surname>Zalucki</surname> <given-names>M. P.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>The socioeconomic impact of Tuta absoluta (Lepidoptera: Gelechiidae) in Eastern Africa</article-title>. <source>J. Economic Entomology</source> <volume>112</volume>, <fpage>1111</fpage>&#x2013;<lpage>1122</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/jee/toz220</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arockia</surname> <given-names>V.</given-names>
</name>
<name>
<surname>Epsy</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Radha</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Pest detection using image denoising and cascaded Unet segmentation for pest images</article-title>. <source>Tuijin Jishu/J. Propulsion Technol</source>. <volume>44</volume> (<issue>4</issue>), <fpage>1359</fpage>&#x2013;<lpage>1371</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.52783/tjjpt.v44.i4.1040</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ashbrook</surname> <given-names>A. R.</given-names>
</name>
<name>
<surname>Mikaelyan</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Schal</surname> <given-names>C.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Comparative efficacy of a fungal entomopathogen with a broad host range against two human-associated pests</article-title>. <source>Insects</source> <volume>13</volume>, <elocation-id>774</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/insects13090774</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Atalan</surname> <given-names>A.</given-names>
</name>
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Atalan</surname> <given-names>Y. A.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Integration of machine learning algorithms and discrete-event simulation for the cost of healthcare resources</article-title>. <source>Healthcare</source> <volume>10</volume>, <elocation-id>1920</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/healthcare10101920</pub-id>
</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Balipoor</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Ommani</surname> <given-names>A. R.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Adoption of biological control measures for Helicoverpa armigera in tomato cultivation</article-title>. <source>J. Agric. Sci.</source> <volume>6</volume>, <fpage>56</fpage>&#x2013;<lpage>62</lpage>.</citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Barbedo</surname> <given-names>J. G. A.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A review on the main challenges in automatic plant disease identification based on visible range images</article-title>. <source>Biosyst. Eng.</source> <volume>144</volume>, <fpage>52</fpage>&#x2013;<lpage>60</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.biosystemseng.2016.01.017</pub-id>
</citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ba&#x15f;t&#xfc;rk</surname> <given-names>F.</given-names>
</name>
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Makine &#xf6;&#x11f;renmesi s&#x131;n&#x131;fland&#x131;rma algoritmalar&#x131;n kar&#x15f;&#x131;la&#x15f;t&#x131;rmas&#x131;: Metinden dil tan&#x131;ma &#xf6;rne&#x11f;i</article-title>. <source>Electronic Lett. Sci. Eng.</source> <volume>18</volume>, <fpage>68</fpage>&#x2013;<lpage>78</lpage>.</citation>
</ref>
<ref id="B8">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Black</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Biderman</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Hallahan</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Anthony</surname> <given-names>Q.</given-names>
</name>
<name>
<surname>Gao</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Golding</surname> <given-names>L.</given-names>
</name>
<etal/>
</person-group>. (<year>2022</year>). <article-title>Gpt-neox-20b: An open-source autoregressive language model</article-title>. <source>arXiv [Preprint]. arXiv:2204.06745</source>.</citation>
</ref>
<ref id="B9">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Blackman</surname> <given-names>R. L.</given-names>
</name>
<name>
<surname>Eastop</surname> <given-names>V. F.</given-names>
</name>
</person-group> (<year>2000</year>). <source>Aphids on the world&#x2019;s crops: An identification and information guide</source>. <publisher-loc>Chichester</publisher-loc>: <publisher-name>Wiley</publisher-name>, <fpage>466</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/0022-2011(95)90067-5</pub-id>
</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bochkovskiy</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>C. Y.</given-names>
</name>
<name>
<surname>Liao</surname> <given-names>H. Y. M.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>YOLOv4: Optimal speed and accuracy of object detection</article-title>. <source>arXiv preprint arXiv:2004.10934</source>.</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Brown</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Mann</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Ryder</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Subbiah</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Kaplan</surname> <given-names>J. D.</given-names>
</name>
<name>
<surname>Dhariwal</surname> <given-names>P.</given-names>
</name>
<etal/>
</person-group>. (<year>2020</year>). <article-title>Language models are few-shot learners</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>33</volume>, <fpage>1877</fpage>&#x2013;<lpage>1901</lpage>.</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Choi</surname> <given-names>Y.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Paudel</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Strawberry pests and diseases detection technique optimized for symptoms using deep learning algorithm</article-title>. <source>J. Bio-Environment Control</source> <volume>31</volume>, <fpage>255</fpage>&#x2013;<lpage>260</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.12791/ksbec.2022.31.3.255</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Danquah</surname> <given-names>O. B.</given-names>
</name>
<name>
<surname>Asare</surname> <given-names>D. K.</given-names>
</name>
<name>
<surname>Asante</surname> <given-names>K. P.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Insecticide use patterns among tomato farmers in Ghana</article-title>. <source>Int. J. Pest Manage.</source> <volume>56</volume>, <fpage>343</fpage>&#x2013;<lpage>351</lpage>.</citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Desneux</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Wajnberg</surname> <given-names>E.</given-names>
</name>
<name>
<surname>Wyckhuys</surname> <given-names>K. A. G.</given-names>
</name>
<name>
<surname>Burgio</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Arpaia</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Narv&#xe1;ez-Vasquez</surname> <given-names>C. A.</given-names>
</name>
<etal/>
</person-group>. (<year>2010</year>). <article-title>Biological invasion of European tomato crops by Tuta absoluta: Ecology, geographic expansion and prospects for biological control</article-title>. <source>J. Pest Sci.</source> <volume>83</volume>, <fpage>197</fpage>&#x2013;<lpage>215</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jip.2009.09.011</pub-id>
</citation>
</ref>
<ref id="B15">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Devlin</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Chang</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Lee</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Toutanova</surname> <given-names>K.</given-names>
</name>
</person-group> (<year>2019</year>). &#x201c;<article-title>BERT: Pre-training of deep bidirectional transformers for language understanding</article-title>.&#x201d; in: <conf-name>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</conf-name>, eds. <person-group person-group-type="editor">
<name>
<surname>Burstein</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Doran</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Solorio</surname> <given-names>T.</given-names>
</name>
</person-group> (<publisher-name>NAACL</publisher-name>). <volume>1</volume>, <fpage>4171</fpage>&#x2013;<lpage>4186</lpage>.</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gaddikeri</surname> <given-names>V.</given-names>
</name>
<name>
<surname>Jatav</surname> <given-names>M. S.</given-names>
</name>
<name>
<surname>Rajput</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Revolutionizing agriculture: Unlocking the potential of ChatGPT in agriculture</article-title>. <source>Food Sci. Rep.</source> <volume>4</volume>, <fpage>20</fpage>&#x2013;<lpage>25</lpage>.</citation>
</ref>
<ref id="B17">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Girshick</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Donahue</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Darrell</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Malik</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Rich feature hierarchies for accurate object detection and semantic segmentation</article-title>. In <conf-name>Proceedings of the IEEE conference on computer vision and pattern recognition</conf-name>, pp. <fpage>580</fpage>&#x2013;<lpage>587</lpage>.</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gu</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Lin</surname> <given-names>T. Y.</given-names>
</name>
<name>
<surname>Kuo</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Cui</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Open-vocabulary object detection via vision and language knowledge distillation</article-title>. <source>arXiv preprint arXiv:2104.13921</source>.</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>G&#xfc;ven</surname> <given-names>&#xd6;.</given-names>
</name>
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Predictive maintenance based on machine learning in public transportation vehicles</article-title>. <source>M&#xfc;hendislik Bilimleri ve Ara&#x15f;t&#x131;rmalar&#x131; Dergisi</source> <volume>4</volume>, <fpage>89</fpage>&#x2013;<lpage>98</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.46387/bjesr.1093519</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hashimoto</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Ohtsu</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Kudo</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Adaptability of YOLO model for agricultural pest detection</article-title>. <source>Comput. Electron. Agric.</source> <volume>168</volume>, <fpage>105120</fpage>.</citation>
</ref>
<ref id="B21">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Gkioxari</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Doll&#xe1;r</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Girshick</surname> <given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Mask R-CNN</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE International Conference on Computer Vision</conf-name>. <fpage>2961</fpage>&#x2013;<lpage>2969</lpage>.</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hoebeke</surname> <given-names>E. R.</given-names>
</name>
<name>
<surname>Carter</surname> <given-names>M. E.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Halyomorpha halys (St&#xe5;l) (Heteroptera: Pentatomidae): A new adventive pest of some crops in North America</article-title>. <source>Proc. Entomological Soc. Washington</source> <volume>105</volume>, <fpage>225</fpage>&#x2013;<lpage>237</lpage>.</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Howard</surname> <given-names>A. G.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Mobilenets: Efficient convolutional neural networks for mobile vision applications</article-title>. <source>arXiv preprint arXiv:1704.04861</source>.</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jee</surname> <given-names>Y. H.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>Y. C.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>J. Y.</given-names>
</name>
</person-group> (<year>2000</year>). <article-title>The allergenic impact of Tetranychus urticae among greenhouse workers</article-title>. <source>Ann. Allergy Asthma Immunol.</source> <volume>84</volume>, <fpage>543</fpage>&#x2013;<lpage>547</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/S1081-1206(10)62520-3</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jin</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Gao</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>L.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Application of AI in pest detection for sustainable agriculture</article-title>. <source>Agric. Syst.</source> <volume>193</volume>, <fpage>103233</fpage>.</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khalid</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Oqaibi</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Aqib</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Hafeez</surname> <given-names>Y.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Small pests detection in field crops using deep learning object detection</article-title>. <source>Sustainability</source> <volume>15</volume>, <elocation-id>6815</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/su15086815</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kumbasaro&#x11f;lu</surname> <given-names>H.</given-names>
</name>
<name>
<surname>&#xd6;z&#xe7;elik</surname> <given-names>A.</given-names>
</name>
<name>
<surname>&#xc7;elikkol</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Growing tomato under protected cultivation conditions: Overall effects on productivity, nutritional yield, and pest incidences</article-title>. <source>Crops</source> <volume>1</volume>, <fpage>97</fpage>&#x2013;<lpage>110</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/crops1020010</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>H. Y.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>J. R.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>Y. J.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>S. X.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>S. S.</given-names>
</name>
<etal/>
</person-group>. (<year>2023</year>). <article-title>Plant resistance against whitefly and its engineering</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1232735</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>X.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Tomato diseases and pests detection based on improved YOLO V3 convolutional neural network</article-title>. <source>Front. Plant Sci.</source> <volume>11</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2020.00898</pub-id>
</citation>
</ref>
<ref id="B30">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>W.</given-names>
</name>
<name>
<surname>Anguelov</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Erhan</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Szegedy</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Reed</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Fu</surname> <given-names>C. Y.</given-names>
</name>
<etal/>
</person-group>. (<year>2016</year>). &#x201c;<article-title>SSD: Single shot multibox detector</article-title>,&#x201d; in <conf-name>Computer Vision&#x2013;ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11&#x2013;14, 2016, Proceedings, Part I</conf-name>, vol. <volume>14</volume>. (<publisher-name>Springer International Publishing</publisher-name>), p. <fpage>21</fpage>&#x2013;<lpage>37</lpage>.</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mirhaji</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Soleymani</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Asakereh</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Mehdizadeh</surname> <given-names>S. A.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Fruit detection and load estimation of an orange orchard using the YOLO models through simple approaches in different imaging and illumination conditions</article-title>. <source>Comput. Electron. Agric.</source> <volume>191</volume>, <elocation-id>106533</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2021.106533</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pereyra</surname> <given-names>P. C.</given-names>
</name>
<name>
<surname>S&#xe1;nchez</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>Effect of two solanaceous plants on developmental and population parameters of the tomato leaf miner, Tuta absoluta (Meyrick) (Lepidoptera: Gelechiidae)</article-title>. <source>Neotropical Entomology</source> <volume>35</volume>, <fpage>671</fpage>&#x2013;<lpage>676</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1590/S1519-566X2006000500016</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Radford</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Child</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Luan</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Amodei</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Sutskever</surname> <given-names>I.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Language models are unsupervised multitask learners</article-title>. <source>OpenAI Blog</source> <volume>1</volume>, <fpage>9</fpage>.</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rajamohanan</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Latha</surname> <given-names>B. C.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>An optimized YOLO v5 model for tomato leaf disease classification with field dataset</article-title>. <source>Engineering Technol. Appl. Sci. Res.</source> <volume>13</volume>, <fpage>12033</fpage>&#x2013;<lpage>12038</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.48084/etasr.6377</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rao</surname> <given-names>D.</given-names>
</name>
<name>
<surname>McCann</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>ChatGPT-4: training and applications</article-title>. <source>AI Magazine</source> <volume>44</volume>, <fpage>22</fpage>&#x2013;<lpage>30</lpage>.</citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ray</surname> <given-names>P.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>AI-assisted sustainable farming: harnessing the power of ChatGPT in modern agricultural sciences and technology</article-title>. <source>ACS Agric. Sci. Technol.</source> <volume>6</volume>, <fpage>460</fpage>&#x2013;<lpage>462</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1021/acsagscitech.3c00145</pub-id>
</citation>
</ref>
<ref id="B37">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Redmon</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Divvala</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Girshick</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Farhadi</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>You only look once: Unified, real-time object detection</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</conf-name>. (<conf-loc>Las Vegas, NV, USA</conf-loc>), <fpage>779</fpage>&#x2013;<lpage>788</lpage>.</citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Redmon</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Farhadi</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>YOLOv3: An incremental improvement</article-title>. <source>arXiv</source>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1804.02767</pub-id>
</citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rubanga</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Loyani</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Richard</surname> <given-names>M.</given-names>
</name>
<name>
<surname>Shimada</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>A deep learning approach for determining effects of Tuta absoluta in tomato plants</article-title>. <source>arXiv</source>.</citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
<name>
<surname>G&#xfc;nt&#xfc;rk&#xfc;n</surname> <given-names>R.</given-names>
</name>
<name>
<surname>H&#x131;z</surname> <given-names>O.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Design and application of PLC controlled robotic arm choosing objects according to their color</article-title>. <source>Electronic Lett. Sci. Eng.</source> <volume>16</volume>, <fpage>52</fpage>&#x2013;<lpage>62</lpage>.</citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Topal</surname> <given-names>B.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>The effect of the use of information technologies in businesses on cost and financial performance</article-title>. <source>Int. J. Eng. Innov. Res. (IJEIR)</source> <volume>5</volume>, <fpage>394</fpage>&#x2013;<lpage>402</lpage>.</citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shaltiel-Harpaz</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Gerling</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Graph</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Kedoshim</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Azolay</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Rozenberg</surname> <given-names>T.</given-names>
</name>
<etal/>
</person-group>. (<year>2015</year>). <article-title>Control of the tomato leafminer, Tuta absoluta (Lepidoptera: Gelechiidae), in open-field tomatoes by indigenous natural enemies occurring in Israel</article-title>. <source>J. Economic Entomology</source> <volume>109</volume>, <fpage>120</fpage>&#x2013;<lpage>131</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/jee/tov309</pub-id>
</citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shorten</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Khoshgoftaar</surname> <given-names>T. M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A survey on image data augmentation for deep learning</article-title>. <source>Journal of big data</source> <volume>6</volume> (<issue>1</issue>), <fpage>1</fpage>&#x2013;<lpage>48</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s40537-019-0197-0</pub-id>
</citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Siche</surname> <given-names>R.</given-names>
</name>
<name>
<surname>Siche</surname> <given-names>N.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>The language model based on sensitive artificial intelligence - ChatGPT: bibliometric analysis and possible uses in agriculture and livestock</article-title>. <source>Scientia Agropecuaria</source> <volume>1</volume>, <fpage>111</fpage>&#x2013;<lpage>116</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.17268/sci.agropecu.2023.010</pub-id>
</citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Swinburne</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Yadav</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Kim</surname> <given-names>J.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Efficiency of AI models in agricultural pest management</article-title>. <source>J. Agric. Technol.</source> <volume>32</volume>, <fpage>245</fpage>&#x2013;<lpage>260</lpage>.</citation>
</ref>
<ref id="B46">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Tang</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Qi</surname> <given-names>F.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2021</year>). &#x201c;<article-title>Pest-YOLO: deep image mining and multi-feature fusion for real-time agriculture pest detection</article-title>,&#x201d; in <conf-name>2021 IEEE International Conference on Data Mining (ICDM)</conf-name>. (<publisher-name>IEEE</publisher-name>). <fpage>1348</fpage>&#x2013;<lpage>1353</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICDM51629.2021.00169</pub-id>
</citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Touvron</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Martin</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Stone</surname> <given-names>K.</given-names>
</name>
<name>
<surname>Albert</surname> <given-names>P.</given-names>
</name>
<name>
<surname>Almahairi</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Babaei</surname> <given-names>Y.</given-names>
</name>
<etal/>
</person-group>. (<year>2023</year>). <article-title>Llama 2: Open foundation and fine-tuned chat models</article-title>. <source>arXiv preprint</source> <volume>arXiv</volume>:<fpage>2307.09288</fpage>.</citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vaswani</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Shazeer</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Parmar</surname> <given-names>N.</given-names>
</name>
<name>
<surname>Uszkoreit</surname> <given-names>J.</given-names>
</name>
<name>
<surname>Jones</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Gomez</surname> <given-names>A. N.</given-names>
</name>
<etal/>
</person-group>. (<year>2017</year>). <article-title>Attention is all you need</article-title>. <source>Adv. Neural Inf. Process. Syst.</source> <volume>30</volume>, <fpage>5998</fpage>&#x2013;<lpage>6008</lpage>.</citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wen</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Ma</surname> <given-names>Z.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>T.</given-names>
</name>
<name>
<surname>Su</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>H.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Pest-YOLO: A model for large-scale multi-class dense and tiny pest detection and counting</article-title>. <source>Front. Plant Sci.</source> <volume>13</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2022.973985</pub-id>
</citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname> <given-names>X.</given-names>
</name>
<name>
<surname>Zhao</surname> <given-names>B.</given-names>
</name>
<name>
<surname>Lin</surname> <given-names>F.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>High mean average precision scores for agricultural pest detection using YOLO</article-title>. <source>Comput. Electron. Agric.</source> <volume>175</volume>, <fpage>105456</fpage>.</citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Y&#x131;lmaz</surname> <given-names>M.</given-names>
</name>
<name>
<surname>&#x15e;ahin</surname> <given-names>H.</given-names>
</name>
<name>
<surname>Y&#x131;ld&#x131;z</surname> <given-names>A.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Sectoral application analysis of studies made with deep learning models</article-title>. <source>Electronic Lett. Sci. Eng.</source> <volume>17</volume>, <fpage>126</fpage>&#x2013;<lpage>140</lpage>.</citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yong</surname> <given-names>A.</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Jun</surname> <given-names>T.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Research on recognition model of crop diseases and insect pests based on deep learning in harsh environments</article-title>. <source>IEEE Access</source> <volume>8</volume>, <fpage>171686</fpage>&#x2013;<lpage>171693</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ACCESS.2020.3025325</pub-id>
</citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname> <given-names>L.</given-names>
</name>
<name>
<surname>Ding</surname> <given-names>G.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Li</surname> <given-names>D.</given-names>
</name>
</person-group> (<year>2023</year>a). <article-title>DCF-Yolov8: An improved algorithm for aggregating low-level features to detect agricultural pests and diseases</article-title>. <source>Agronomy</source> <volume>13</volume>, <elocation-id>2012</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy13082012</pub-id>
</citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname> <given-names>S.</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>C.</given-names>
</name>
<name>
<surname>Park</surname> <given-names>D.</given-names>
</name>
<name>
<surname>Yoon</surname> <given-names>S.</given-names>
</name>
</person-group> (<year>2023</year>b). <article-title>Editorial: Machine learning and artificial intelligence for smart agriculture, volume II</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1166209</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>