<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2026.1747657</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Estimation of Kenaf seedling canopy coverage in saline soil using semantic segmentation of UAV RGB images</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name><surname>Wang</surname><given-names>Wei</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3189252/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Cao</surname><given-names>Kunzhi</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3328787/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project-administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Luo</surname><given-names>Guiying</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Huang</surname><given-names>Ruohan</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Nie</surname><given-names>Jihao</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Zhang</surname><given-names>Junyu</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Lu</surname><given-names>Jianning</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Cui</surname><given-names>Guoxian</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>An</surname><given-names>Xia</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>*</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/1664975/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>She</surname><given-names>Wei</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>*</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2800188/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>College of Agriculture &amp; Ramie Research Institute, Hunan Agricultural University, Yuelushan Laboratory</institution>, <city>Changsha</city>,&#xa0;<country country="cn">China</country></aff>
<aff id="aff2"><label>2</label><institution>Zhejiang Xiaoshan Institute of Cotton &amp; Bast Fiber Crops, Zhejiang Institute of Landscape Plants and Flowers, Zhejiang Academy of Agricultural Sciences</institution>, <city>Hangzhou</city>,&#xa0;<country country="cn">China</country></aff>
<author-notes>
<corresp id="c001"><label>*</label>Correspondence: Xia An, <email xlink:href="mailto:anxia@zaas.ac.cn">anxia@zaas.ac.cn</email>; Wei She, <email xlink:href="mailto:weishe@hunau.edu.cn">weishe@hunau.edu.cn</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-10">
<day>10</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>17</volume>
<elocation-id>1747657</elocation-id>
<history>
<date date-type="received">
<day>16</day>
<month>11</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>22</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="rev-recd">
<day>13</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Wang, Cao, Luo, Huang, Nie, Zhang, Lu, Cui, An and She.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Wang, Cao, Luo, Huang, Nie, Zhang, Lu, Cui, An and She</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-10">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>The growth status of Kenaf (<italic>Hibiscus cannabinus</italic> L.) seedlings directly impacts its yield and quality. Addressing the challenges of inefficient monitoring and quantitative assessment of Kenaf seedlings under saline-alkali conditions, this study developed an automated method for plant identification and canopy coverage estimation during the seedling stage. This approach leverages high-resolution visible light imagery captured by unmanned aerial vehicles (UAVs) combined with deep learning semantic segmentation techniques. First, a UAV imagery dataset of Kenaf seedlings was constructed through geometric and radiometric calibration, image cropping, and sample annotation. Subsequently, three classical semantic segmentation models&#x2014;FCN, U-Net, and DeepLabV3+&#x2014;were trained and compared using image enhancement strategies. Model performance was quantitatively evaluated using metrics including Intersection over Union (IoU), accuracy, precision, and F1 score. Results indicate that all three models effectively segmented Kenaf plants from soil backgrounds. U-Net demonstrated optimal overall accuracy and detail retention, DeepLabV3+ exhibited advantages in small-scale object recognition, while FCN offered high computational efficiency, making it suitable for applications demanding real-time processing. Building upon this, the U-Net architecture was enhanced by incorporating a Self-Attention (SE) channel mechanism, further improving model performance to achieve an IoU of 85.99% and an Dice of 92.44%. Based on segmentation results from the enhanced UNet, plant canopy coverage during the Kenaf seedling stage was calculated. Combined with measured dry bark yield per mu, this enabled analysis of growth performance across varieties under saline-alkali conditions, identifying Xiao 3, K5, and Xiao 2 as materials exhibiting strong saline-alkali adaptability. The study demonstrates that this method enables high-precision identification and quantitative analysis of Kenaf seedlings, providing effective technical support for monitoring seedling growth and variety selection in saline-alkali soils.</p>
</abstract>
<kwd-group>
<kwd>deep learning</kwd>
<kwd>Kenaf</kwd>
<kwd>saline soil tolerance</kwd>
<kwd>UAV remote sensing</kwd>
<kwd>U-net</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This research was supported by the Ministry of Finance and Ministry of Agriculture and Rural Affairs: National Modern Agricultural Industry Technology System (CARS-16-E11), Yuelushan Laboratory Seed Industry Special Project (YLS-2025-ZY02056) and the Scientific Research Fund of Hunan Provincial Education Department (23A0178).</funding-statement>
</funding-group>
<counts>
<fig-count count="7"/>
<table-count count="6"/>
<equation-count count="1"/>
<ref-count count="32"/>
<page-count count="15"/>
<word-count count="6871"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Technical Advances in Plant Science</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Kenaf (Hibiscus <italic>cannabinus L</italic>.), belonging to the genus Hibiscus within the Malvaceae family, is a significant annual bast fiber crop widely utilized in hemp textiles, papermaking, and construction materials (<xref ref-type="bibr" rid="B29">Webber et&#xa0;al., 2002</xref>). In recent years, soil salinization has become increasingly severe, emerging as one of the primary factors constraining sustainable agricultural development (<xref ref-type="bibr" rid="B20">Rahman et al., 2002</xref>). Therefore, the rational utilization of saline-alkali soils and the selection of salt-tolerant crop varieties are crucial for ensuring agricultural production (<xref ref-type="bibr" rid="B9">Hassani et&#xa0;al., 2021</xref>). The adaptability of Kenaf to saline-alkali conditions during the seedling stage directly impacts its growth, development, and ultimate yield, with the emergence rate being particularly sensitive to soil salinity levels (<xref ref-type="bibr" rid="B1">Chen et&#xa0;al., 2024</xref>). In modern agricultural production, the rapid and accurate acquisition of crop seedling condition information is a critical step in achieving modernized and precision management of farmland (<xref ref-type="bibr" rid="B15">Lamichhane et&#xa0;al., 2018</xref>). The survey methods of traditional crop condition primarily rely on manual measurements and visual interpretation, which are time-consuming and labor-intensive, making them ill-suited for efficient monitoring of large-scale farmland. With the rapid advancement of unmanned aerial vehicle (UAV) remote sensing technology, the acquisition of high-resolution visible light imagery using lightweight, low-cost drones has become a vital method for crop seedling condition monitoring (<xref ref-type="bibr" rid="B31">Zeng et&#xa0;al., 2025</xref>). This approach provides new technical support for monitoring the early growth stages of economic crops like Kenaf and studying their adaptability to saline-alkali conditions.</p>
<p>Accurate extraction of crop information from imagery, enabling automated recognition and segmentation, is a critical step in achieving precise monitoring. Image segmentation primarily relies on features such as color, texture, and shape to segment and identify regions of interest (<xref ref-type="bibr" rid="B24">Sadashivan et&#xa0;al., 2021</xref>). Color threshold-based segmentation methods utilize differences in the distribution of subjects and backgrounds across grayscale, RGB, and HSV histograms (<xref ref-type="bibr" rid="B5">Feng et&#xa0;al., 2020</xref>). achieved high accuracy of 97.4% by segmenting plants and backgrounds using HSV and CieLab (luminance, green-red channel, blue-yellow channel) color spaces (<xref ref-type="bibr" rid="B21">Riehle et&#xa0;al., 2020</xref>). Another approach employs unsupervised clustering based on pixel color or texture features (<xref ref-type="bibr" rid="B11">Jia et&#xa0;al., 2020</xref>), dividing images into distinct categories from which vegetation regions are extracted. These methods can achieve certain results when the soil background is relatively simple. However, in the face of complex and variable field environments and factors such as weed interference, the color difference between plants and their background is often not clearly discernible (<xref ref-type="bibr" rid="B8">Guo et&#xa0;al., 2025</xref>). Relying solely on manual interpretation or traditional image processing methods often leads to misclassification or missed segmentation, making it difficult to achieve precise crop identification.</p>
<p>In recent years, with the rapid advancement of deep learning technology, image segmentation methods based on convolutional neural networks (CNNs) have been widely applied in the field of remote sensing (<xref ref-type="bibr" rid="B14">Kattenborn et&#xa0;al., 2021</xref>). Compared to traditional approaches, deep learning models can automatically learn multi-level features from massive datasets, significantly enhancing the accuracy and robustness of crop identification and segmentation. Common semantic segmentation networks such as FCN (Fully Convolutional Network), UNet, SegNet, and the DeepLab (<xref ref-type="bibr" rid="B22">Ronneberger et&#xa0;al., 2015</xref>; <xref ref-type="bibr" rid="B26">Wang et&#xa0;al., 2022a</xref>; <xref ref-type="bibr" rid="B27">Wang et&#xa0;al., 2022b</xref>) series have demonstrated outstanding performance in tasks including crop identification (<xref ref-type="bibr" rid="B16">Lapkovskis et&#xa0;al., 2025</xref>), pest and disease detection (<xref ref-type="bibr" rid="B32">Zhang et&#xa0;al., 2022</xref>), and canopy cover estimation (<xref ref-type="bibr" rid="B30">Yao et&#xa0;al., 2025</xref>). Alkhudaydi employed FCN and transfer learning for semantic segmentation of wheat ear images, providing an effective method for yield trait assessment with an average accuracy exceeding 76% (<xref ref-type="bibr" rid="B500">Alkhudaydi et al., 2019</xref>). Jiang combined K-means clustering with an augmented UNet network to perform dual segmentation of rapeseed and weeds, significantly improving recognition accuracy in complex field environments (<xref ref-type="bibr" rid="B12">Jiang et&#xa0;al., 2025</xref>). Furthermore, Ding achieved precise segmentation of apple leaf lesions by refining the DeepLabv3+ model, attaining a segmentation accuracy of 98.45% and significantly improving disease detection accuracy (<xref ref-type="bibr" rid="B3">Ding et al., 2025</xref>).</p>
<p>Beyond CNN-based models, Transformer-based models and hybrid CNN&#x2013;Transformer architectures have garnered increasing attention in recent years due to their robust capabilities in modelling long-range dependencies and capturing multi-scale contextual features. For instance, segmentation frameworks based on Vision Transformers (ViT) (<xref ref-type="bibr" rid="B4">Dosovitskiy et&#xa0;al., 2020</xref>) and their improved variants, such as EPEMMSA-ViT (<xref ref-type="bibr" rid="B28">Wang et&#xa0;al., 2024</xref>) and MMVT (<xref ref-type="bibr" rid="B18">Lucas et&#xa0;al., 2025</xref>), have demonstrated outstanding performance in agricultural remote sensing applications for pest and disease identification and soil health assessment. Concurrently, hybrid models integrating CNN&#x2019;s local perception with Transformer&#x2019;s global modelling advantages continue to emerge. Examples include FTransUNet (<xref ref-type="bibr" rid="B19">Ma et&#xa0;al., 2024</xref>), the lightweight TinySegformer (<xref ref-type="bibr" rid="B501">Zhang and Lv, 2024</xref>), and CTFFNet (<xref ref-type="bibr" rid="B24">Jin et&#xa0;al., 2024)</xref>, which have made significant strides in enhancing sensitivity to small targets and robustness in complex backgrounds. Nevertheless, despite their outstanding performance across multiple agricultural vision tasks, these methods remain under-systematically applied and validated in UAV image segmentation during the crop seedling stage. The small scale and indistinct boundaries of seedlings during this stage impose heightened demands on model stability and generalisation capabilities, areas where existing research remains insufficient. In summary, the evolution from CNNs to Transformers and their hybrid architectures provides powerful tools for accurately extracting crop spatial distribution and early growth information from UAV imagery.</p>
<p>This study aims to utilize drone-captured visible light imagery and semantic segmentation technology to achieve automated identification and precise segmentation of Kenaf seedlings in saline-alkali soils, thereby constructing an efficient and stable model for extracting seedling condition information. By analyzing imagery of young Kenaf seedlings from different varieties within saline-alkali plots, the research explores the response characteristics of Kenaf seedlings to saline-alkali stress during their early growth stage and assesses the salt tolerance adaptability of various cultivars. This provides technical support for selecting salt-tolerant Kenaf varieties and promoting the rational utilization of saline-alkali lands.</p>
<p>To achieve the aforementioned objectives, this paper primarily undertakes the following research activities:</p>
<list list-type="simple">
<list-item>
<p>(1)UAV Image Acquisition and Data Preprocessing: Acquire high-resolution visible light imagery of Kenaf seedlings across different saline-alkali plots via UAV flights. Generate complete orthorectified images through geometric correction, radiometric calibration, and subsequent steps. Construct a Kenaf seedling imagery dataset suitable for deep learning segmentation tasks via image cropping and feature annotation.</p></list-item>
<list-item>
<p>(2)Construction of an Image Semantic Segmentation Model for Kenaf Seedling Stages: Based on three deep learning segmentation networks&#x2014;FCN, UNet, and DeepLabv3+&#x2014;we designed and trained a model suitable for image segmentation of Kenaf seedling stages.</p></list-item>
<list-item>
<p>(3)Model Optimization and Evaluation: Enhance model performance through data augmentation techniques, and conduct quantitative analysis and model validation using precision evaluation metrics such as IoU, F1-score, and Precision.</p></list-item>
<list-item>
<p>(4)Extraction and Analysis of Kenaf Seedling Information in Saline-Alkali Soils: Utilizing the most effective segmentation model to extract seedling emergence data during the Kenaf seedling stage, investigating the adaptability of different Kenaf varieties to saline-alkali soils, analyzing the relationship between adaptability and final yield, and identifying Kenaf varieties with breeding potential.</p></list-item>
</list>
</sec>
<sec id="s2">
<label>2</label>
<title>Materials</title>
<sec id="s2_1">
<label>2.1</label>
<title>Experimental area</title>
<p>The experiment was conducted in Yancheng City, Jiangsu Province, China (32&#xb0;34&#x2032;N&#x2013;34&#xb0;28&#x2032;N, 119&#xb0;27&#x2032;E&#x2013;120&#xb0;54&#x2032;E). Located in the core region of the Yangtze River Delta, Yancheng features flat terrain and a temperate climate, transitioning from the northern subtropical to warm temperate zones. Bordering the East China Sea, the region encompasses approximately 7 million mu (473, 333 hectares) of tidal flats with widespread saline-alkali soils. The experimental area exhibits elevated soil salinity levels, imposing stress on crop germination and early growth (<xref ref-type="fig" rid="f1"><bold>Figure 1</bold></xref>).</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Experimental area map. The red box in <xref ref-type="fig" rid="f1"><bold>Figure&#xa0;1c</bold></xref> indicates the effective range ultimately selected from the sample plot.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g001.tif">
<alt-text content-type="machine-generated">Map panel (a) shows Jiangsu Province highlighted in red on a map of China, with administrative boundaries marked. Panel (b) depicts a detailed elevation map of Jiangsu, showing provinces and the Yellow Sea, with sampling locations indicated. Panel (c) features an aerial view of a field with visible crop patterns and coordinates marked.</alt-text>
</graphic></fig>
<p>Soil pH ranged from 8.88 to 9.07, with total water-soluble salts (TSW) content between 0.12% and 2.86%. Kenaf was sown using broadcast seeding, with seeds evenly distributed across each experimental plot. Each plot measured 40 m in length and 1.4 m in width. Eleven test varieties were included: H368, Xiao 1, Xiao 2, Xiao 3, K1, K2, K3, K4, K5, K6, and K7.</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>UAV image acquisition</title>
<p>Visible light imagery was captured using a quadcopter UAV (DJI Inspire 2, DJI Innovation, Guangdong, China) equipped with an RGB camera featuring a resolution of 5472 &#xd7; 3648 pixels. Multispectral imagery acquisition employed a quadcopter (DJI Phantom 4 Multispectral, DJI Innovation, Guangdong, China) equipped with one visible light sensor and five multispectral sensors. Flight operations occurred from 10:00&#x2013;11:00 on July 9, 2023, with aerial photography parameters set as follows: flight altitude 25 m, lens tilt angle &#x2212;90&#xb0;, flight speed 6 m/s, and both fore-and-aft and sideways overlap rates set to 75%. To minimize image distortion and target movement, data collection was conducted under clear, windless weather conditions. Ultimately, 335 high-resolution visible images and 1, 734 multispectral images were acquired during the Kenaf seedling stage. Image calibration and stitching were completed using DJI Terra software.</p>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Ground-based Kenaf growth data</title>
<p>Prior to harvesting, ground growth data were collected at the Kenaf maturity stage. Fresh plant weight was determined by randomly selecting 20 Kenaf plants and summing their fresh weights; plant height and stem diameter were each taken as the average value from the same batch of 20 plants; bark thickness measurements were obtained by stacking the bark layers from five selected plants to ensure more stable and representative data.</p>
</sec>
<sec id="s2_4">
<label>2.4</label>
<title>Hardware and software environment</title>
<p>All image processing and deep learning experiments were conducted on the same high-performance workstation. The hardware configuration includes a 13th-generation Intel Core i5-13490F processor (base frequency 2.5 GHz), 64 GB DDR4 memory, and an NVIDIA GeForce RTX 4070 graphics card. The software environment utilized Anaconda-Miniconda3 to build isolated virtual environments running Python 3.8, CUDA 11.3, and PyTorch 1.10. Development and debugging were performed using PyCharm Community Edition 2024. Remote sensing feature extraction was conducted with phenoAI air (AgriBrain, Nanjing, China). This hardware and software configuration provided stable and efficient computational support for model training and testing.</p>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Methods</title>
<sec id="s3_1">
<label>3.1</label>
<title>Construction of image dataset</title>
<p>Orthoimages were cropped based on neighborhood boundaries to uniform widths and lengths, yielding sub-images of 340 &#xd7; 512 pixels (Image A) and 340 &#xd7; 8880 pixels (Image B: single-image representation of a single Kenaf variety). From image A, 600 images were randomly selected and manually annotated using LabelMe software to generate binary semantic segmentation labels for Kenaf seedlings/background. To enhance model robustness and generalization, multiple data augmentation strategies were employed during training, including:</p>
<p>Scale change: Images were scaled at six levels (0.5&#xd7;, 0.75&#xd7;, 1.25&#xd7;, 1.5&#xd7;, 1.75&#xd7;, 2&#xd7;) and uniformly cropped to 512&#xd7;512 pixels;</p>
<p>Geometric flip: Including horizontal flipping and vertical flipping by 180&#xb0;;</p>
<p>Photometric distortion: Random adjustments to brightness, contrast, and saturation.</p>
<p>To ensure image quality and consistency, all images underwent mean and standard deviation normalization during dataset construction to minimize color bias caused by lighting variations. Background pixels were filled with 0, while segmented label regions were filled with 255 to ensure label consistency. Finally, the dataset was divided into training, validation, and test sets at an 8:1:1 ratio, providing ample data support for deep learning model construction and evaluation.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Semantic segmentation model construction and training methods</title>
<p>The objective of this study is to establish a standardised, reproducible workflow for monitoring and segmenting Kenaf seedlings using UAV. A systematic comparison will be made of the performance of representative classical semantic segmentation models on this task. In selecting architectures for this study, three were selected based on the following principles.</p>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Model selection rationale</title>
<p>FCN represents a pioneering model that initially accomplished end-to-end pixel-level prediction, thereby establishing the theoretical foundation for all subsequent networks (<xref ref-type="bibr" rid="B17">Long et&#xa0;al., 2015</xref>).</p>
<p>U-Net: This architecture is distinguished by its ability to achieve fine-grained structure recovery through skip connections, thus establishing itself as one of the most widely applied and consistently stable architectures in agricultural image segmentation. In particular, it has been demonstrated to be highly effective for small object detection via UAVs (<xref ref-type="bibr" rid="B2">Chen et&#xa0;al., 2018</xref>).</p>
<p>DeepLabv3+: The model utilises dilated convolutions and the Atrous Spatial Pyramid Pooling (ASPP) module to deliver robust multi-scale contextual modelling capabilities, thus representing a high-precision segmentation model in complex field environments (<xref ref-type="bibr" rid="B22">Ronneberger et&#xa0;al., 2015</xref>).</p>
<p>The second point to consider is the alignment with UAV agricultural imagery characteristics and research task requirements.</p>
<p>Kenaf seedlings in UAV imagery manifest characteristics of small volume, dispersed distribution, fine and irregular edges, while being significantly affected by variations in illumination, shadow interference, and complex background textures. In the domain of agricultural remote sensing, U-Net and DeepLab architectures have been extensively validated for the segmentation of small objects, while FCN serves as a natural baseline model for systematic comparison. Moreover, given that transformer-based models generally necessitate more substantial training datasets and greater computational resources, the present study concentrates on the aforementioned classical and efficient convolutional network architectures to guarantee reproducibility and stability.</p>
<p>In the domain of agricultural remote sensing, U-Net and DeepLab architectures have been extensively validated for the segmentation of small objects, while FCN serves as a natural baseline model for systematic comparison. Furthermore, given that transformer-based models generally necessitate larger training datasets and greater computational resources, the present study focuses on the aforementioned classical and efficient convolutional network architectures to ensure reproducibility, stability, and comparability of experimental results.</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Design of the SE channel attention module</title>
<p>To enhance U-Net&#x2019;s feature representation capabilities in complex field environments, this study introduces a squeeze-excitation (SE) channel attention module. The SE module adaptively recalibrates the importance of feature channels by explicitly modelling inter-channel dependencies (<xref ref-type="bibr" rid="B10">Hu et&#xa0;al., 2017</xref>).</p>
<p>As illustrated in <xref ref-type="fig" rid="f2"><bold>Figure&#xa0;2</bold></xref>, the SE module comprises three operations: squeeze, excitation, and scale. First, a global average pooling operation generates channel-level descriptors; subsequently, a two-layer fully connected network learns channel weights; Finally, it recalibrates the original features through per-channel multiplication, amplifying information-rich features while suppressing irrelevant ones.</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>SE-UNet network.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g002.tif">
<alt-text content-type="machine-generated">Diagram of a neural network architecture using convolutional layers, ReLU activations, a squeeze-and-excitation (SE) block, max pooling, and up-convolutions. Blue arrows indicate convolution with ReLU, yellow arrows indicate SE blocks, red arrows for max pooling, and green arrows for up-convolutions. The SE block processes the channel-wise features, enhancing network performance.</alt-text>
</graphic></fig>
<p>The SE module is embedded at the end of each stage within the U-Net encoder, enhancing feature recognition of minute seedling regions amidst complex backgrounds. The improved network, termed SE-UNet, undergoes quantitative performance evaluation against baseline U-Net and alternative models in Section 4.</p>
</sec>
<sec id="s3_2_3">
<label>3.2.3</label>
<title>Unified training strategy</title>
<p>In order to present the material clearly and intuitively whilst ensuring comparability between models, the core configurations of the study&#x2019;s unified training strategy are summarised in the following <xref ref-type="table" rid="T1"><bold>Table&#xa0;1</bold></xref>:</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Unified configuration for model training.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Config item</th>
<th valign="middle" align="center">Parameter setting</th>
<th valign="middle" align="center">Description</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">Maximum Iterations</td>
<td valign="middle" align="center">40, 000</td>
<td valign="middle" align="center">Validation is performed every 4, 000 iterations.</td>
</tr>
<tr>
<td valign="middle" rowspan="2" align="center">Optimizer</td>
<td valign="middle" align="center">SGD</td>
<td valign="middle" align="center">Learning rate: 0.0025, Momentum: 0.9, Weight decay: 0.0005</td>
</tr>
<tr>
<td valign="middle" align="center">AdamW</td>
<td valign="middle" align="center">Learning rate: 0.0001, Beta1: 0.9, Beta2: 0.999, Weight decay: 0.01</td>
</tr>
<tr>
<td valign="middle" align="center">Learning Rate Schedule</td>
<td valign="middle" align="center">Polynomial decay (PolyLR)</td>
<td valign="middle" align="center">SDG: 1e-4<break/>AdamW: 1e-6</td>
</tr>
<tr>
<td valign="middle" align="center">Pre-trained Weights</td>
<td valign="middle" align="center">ResNet-101</td>
<td valign="middle" align="center">The backbone network is initialized with weights pre-trained on ImageNet.</td>
</tr>
<tr>
<td valign="middle" align="center">Training Resume</td>
<td valign="middle" align="center">Checkpoint resume mode enabled</td>
<td valign="middle" align="center">Automatically loads the best saved weights to continue training after an interruption.</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>This standardised configuration serves to eliminate performance variations inherent to the training process itself, thereby ensuring fairness in subsequent model comparisons and the reliability of conclusions drawn.</p>
</sec>
<sec id="s3_2_4">
<label>3.2.4</label>
<title>Data augmentation strategy melting experiment</title>
<p>Image enhancement is crucial for improving the generalisation capability of semantic segmentation models, with different enhancement strategies contributing to segmentation performance to varying degrees. To systematically evaluate the contributions of various enhancement methods, this study designed image enhancement ablation experiments based on the SE-UNet architecture.</p>
<p>In each experiment, only one data enhancement strategy was activated, with all others disabled. This controlled experimental design enables the assessment of each enhancement method&#x2019;s independent impact on model performance. Three independent experiments were conducted:</p>
<list list-type="order">
<list-item>
<p>Scale change only: Random scale transformations (scaling to 0.5&#x2013;2 times the original image) applied to input images to simulate variations in ground sampling distance and plant size caused by differing flight altitudes.</p></list-item>
<list-item>
<p>Geometric inversion only: Random horizontal/vertical flipping (flipping probability 0.5) to enhance model robustness against perspective shifts and spatial layout variations.</p></list-item>
<list-item>
<p>Photometric distortion only: Employed luminance enhancement techniques (random brightness/contrast adjustments: &#xb1; 20% brightness, &#xb1; 15% contrast) to compensate for illumination variations caused by solar angle shifts and atmospheric conditions during UAV imaging.</p></list-item>
</list>
<p>By comparing SE-UNet&#x2019;s segmentation performance under these three independent augmentation strategies, this ablation study aims to reveal the relative effectiveness of scale, geometric, and luminance data augmentation for Kenaf seedling segmentation.</p>
</sec>
<sec id="s3_2_5">
<label>3.2.5</label>
<title>Model testing method</title>
<p>Model testing employs Test Time Augmentation (TTA) technology, generating augmented images through multi-scale transformations and horizontal flipping to enhance prediction accuracy. Given the extended length of the single Image B neighborhood image, a sliding window strategy is adopted for prediction, balancing computational efficiency with accuracy.</p>
<p>Subsequently, SegLocal Visualizer is used to visually compare segmentation results against ground truth labels, enabling intuitive assessment of model performance.</p>
</sec>
<sec id="s3_2_6">
<label>3.2.6</label>
<title>Model evaluation metric</title>
<p>Model performance is comprehensively evaluated using multiple metrics, including: Intersection over Union (IoU), accuracy, F1 score, precision, recall, as well as data loading time and model runtime. The calculation formulas are shown in <xref ref-type="table" rid="T2"><bold>Table&#xa0;2</bold></xref>.</p>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Formulae for model evaluation metrics.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Abbreviation</th>
<th valign="middle" align="center">Full Name or Calculation Formulas</th>
<th valign="middle" align="center">Serial Number</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">TP</td>
<td valign="middle" align="left">Positive samples predicted as positive by the model (correctly predicted positive samples)</td>
<td valign="middle" align="left">(1)</td>
</tr>
<tr>
<td valign="middle" align="left">TN</td>
<td valign="middle" align="left">Negative samples predicted as negative by the model (correctly predicted negative samples)</td>
<td valign="middle" align="left">(2)</td>
</tr>
<tr>
<td valign="middle" align="left">FP</td>
<td valign="middle" align="left">Negative samples predicted as positive by the model (positive samples with incorrect predictions)</td>
<td valign="middle" align="left">(3)</td>
</tr>
<tr>
<td valign="middle" align="left">FN</td>
<td valign="middle" align="left">Positive samples predicted as negative by the model (negatively classified samples with incorrect predictions)</td>
<td valign="middle" align="left">(4)</td>
</tr>
<tr>
<td valign="middle" align="left">Iou</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im1"><mml:mrow><mml:mtext>IoU</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>TP</mml:mtext></mml:mrow><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FN</mml:mtext></mml:mrow></mml:mfrac></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="left">(5)</td>
</tr>
<tr>
<td valign="middle" align="left">Accuracy</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im2"><mml:mrow><mml:mtext>Acc</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>TN</mml:mtext></mml:mrow><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>TN</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FN</mml:mtext></mml:mrow></mml:mfrac></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="left">(6)</td>
</tr>
<tr>
<td valign="middle" align="left">F1-score</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im3"><mml:mrow><mml:mtext>F</mml:mtext><mml:mn>1</mml:mn><mml:mo>=</mml:mo><mml:mn>2</mml:mn><mml:mo>*</mml:mo><mml:mfrac><mml:mrow><mml:mtext>P</mml:mtext><mml:mo>*</mml:mo><mml:mtext>R</mml:mtext></mml:mrow><mml:mrow><mml:mtext>P</mml:mtext><mml:mo>+</mml:mo><mml:mtext>R</mml:mtext></mml:mrow></mml:mfrac></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="left">(7)</td>
</tr>
<tr>
<td valign="middle" align="left">Precision</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im4"><mml:mrow><mml:mtext>P</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>TP</mml:mtext></mml:mrow><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FP</mml:mtext></mml:mrow></mml:mfrac></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="left">(8)</td>
</tr>
<tr>
<td valign="middle" align="left">Recall</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im5"><mml:mrow><mml:mtext>R</mml:mtext><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mtext>TP</mml:mtext></mml:mrow><mml:mrow><mml:mtext>TP</mml:mtext><mml:mo>+</mml:mo><mml:mtext>FN</mml:mtext></mml:mrow></mml:mfrac></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="left">(9)</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Calculation of plant canopy coverage during the seedling stage of Kenaf</title>
<p>Based on the optimal segmentation model trained, semantic segmentation was performed on the Image B dataset. The model automatically identified and distinguished the Kenaf canopy coverage from the background areas in the images, generating a binary segmentation result map where plant area pixels are valued as 1 and background pixels as 0.</p>
<p>Subsequently, using NumPy&#x2014;a Python module for image analysis&#x2014;the number of pixels within the plant canopy coverage was counted, and its proportion relative to the entire image was calculated. This proportion reflects the coverage of Kenaf in the field image, i.e., the percentage of plant area. The calculation formula is as follows:</p>
<disp-formula>
<mml:math display="block" id="M1"><mml:mrow><mml:mtext>Kenaf&#xa0;Canopy&#xa0;Coverage</mml:mtext><mml:mo>=</mml:mo><mml:mtext>Number&#xa0;of&#xa0;Pixels&#xa0;Occupied&#xa0;by&#xa0;Plants</mml:mtext><mml:mo stretchy="false">/</mml:mo><mml:mtext>Total&#xa0;Number&#xa0;of&#xa0;Pixels</mml:mtext><mml:mo>&#xd7;</mml:mo><mml:mn>100</mml:mn><mml:mo>%</mml:mo></mml:mrow></mml:math>
</disp-formula>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Obtaining Kenaf seedling-stage plant information using remote sensing features</title>
<p>To further investigate the relationship between Kenaf seedling growth information and remote sensing characteristics, this study employed PhenoAI Air software to systematically analyze and extract features from remote sensing data across all experimental plots. This software can remove soil background from images. For each Kenaf variety area, through information annotation, it automatically calculates and outputs 138 remote sensing feature values across three major categories: color features, texture features, and spectral indices (<xref ref-type="bibr" rid="B6">Fu et&#xa0;al., 2023</xref>).</p>
<p>Color features primarily reflect the color composition and brightness variations of plant canopies, indirectly indicating crop health status. Texture features describe the spatial distribution patterns of grayscale values in images, aiding in distinguishing plants from soil backgrounds or weeds. Spectral indices quantitatively reflect vegetation coverage, photosynthetic activity, and biomass changes, serving as crucial indicators for vegetation growth monitoring.</p>
<p>Following feature extraction, multiple feature selection methods, including Pearson correlation analysis, Lasso, Competitive Adaptive Reweighted Sampling (CARS), and Least Angle Regression (LARS), were conducted between various remote sensing feature values and the plant canopy coverage of Kenaf seedlings obtained through optimal model segmentation. By comparing the key features identified through different methods, determine the optimal set of remote sensing feature values. The complete data collection and processing, model construction, and analysis workflow diagram is shown in <xref ref-type="fig" rid="f3"><bold>Figure 3</bold></xref>.</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Data collection and processing, model construction and analysis flowchart.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g003.tif">
<alt-text content-type="machine-generated">This figure demonstrates a comparative analysis of aerial imagery of farmland processed through various semantic segmentation models. From left to right, the images represent: the original image, FCN, DeeplabV3+, UNet, and SE-UNet segmentation results. The red boxes highlight areas of missed and redundant detections within each model's segmentation output, providing a visual comparison of algorithmic performance in crop identification tasks.</alt-text>
</graphic></fig>
</sec>
</sec>
<sec id="s4" sec-type="results">
<label>4</label>
<title>Result</title>
<sec id="s4_1">
<label>4.1</label>
<title>Visualization results for model segmentation</title>
<p>To visually demonstrate the performance of different models in segmenting Kenaf seedlings, the segmentation results were visualized. Plant regions are represented in green, while soil background appears in blue. The overlay layer&#x2019;s opacity was set to 50% to enable simultaneous observation of the original image and segmentation results.</p>
<p>The segmentation visualizations reveal that all three models accurately distinguish Kenaf plants from the soil background, though differences exist in detail handling and boundary recognition. The FCN model achieves overall coherent segmentation with fewer missed detections, yet some soil background misclassification persists within plant regions&#x2014;particularly in areas with heavy plant shadows. Additionally, when plants densely cluster in small patches, FCN struggles to effectively separate adjacent individuals, leading to minor plant adhesion in small areas.</p>
<p>The DeepLabv3+ model excels at identifying small-scale individual plants, effectively preserving the morphological features of independent plants. However, it underperforms in recognizing some plants with lighter leaf colors, resulting in a few missed detections (<xref ref-type="fig" rid="f4"><bold>Figure 4</bold></xref>).</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Segmentation results of three different models: <bold>(a)</bold> original image; <bold>(b)</bold> FCN; <bold>(c)</bold> DeepLabv3+; <bold>(d)</bold> U-Net; <bold>(e)</bold> SE-UNet.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g004.tif">
<alt-text content-type="machine-generated">Three sets of aerial images in rows show farmland with green crops and soil. Each row progresses from an original image on the left to processed images with increasing red square markings highlighting specific areas on the crops, from (b) to (e). The squares represent detected items or features within the crops.</alt-text>
</graphic></fig>
<p>In contrast, the UNet model delivers the best overall segmentation accuracy and regional coherence. Within small areas, UNet preserves plant details more effectively. Although minor missed detections still occur in isolated cases, its overall visual consistency and recognition accuracy surpass those of the other two models.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Accuracy evaluation of three semantic segmentation models</title>
<p>The segmentation results of three models&#x2014;DeepLabv3+, FCN, and U-Net&#x2014;were compared (<xref ref-type="table" rid="T3"><bold>Table&#xa0;3</bold></xref>) to evaluate their performance in identifying Kenaf seedlings during the seedling stage. Overall, all three models effectively distinguished Kenaf from soil background areas, though differences existed in accuracy and computational efficiency.</p>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Segmentation results of three different models.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Model</th>
<th valign="middle" align="center">Type</th>
<th valign="middle" align="center">IoU</th>
<th valign="middle" align="center">Accuracy</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">Recall</th>
<th valign="middle" align="center">Data loading time</th>
<th valign="middle" align="center">Model run time</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" rowspan="3" align="center">DeepLabv3+</td>
<td valign="middle" align="center">Kenaf</td>
<td valign="middle" align="center">78.35</td>
<td valign="middle" align="center">84.34</td>
<td valign="middle" align="center">87.86</td>
<td valign="middle" align="center">91.7</td>
<td valign="middle" align="center">84.34</td>
<td valign="middle" rowspan="3" align="center">0.258</td>
<td valign="middle" rowspan="3" align="center">1.707</td>
</tr>
<tr>
<td valign="middle" align="center">Background</td>
<td valign="middle" align="center">86.69</td>
<td valign="middle" align="center">95.21</td>
<td valign="middle" align="center">92.87</td>
<td valign="middle" align="center">90.65</td>
<td valign="middle" align="center">95.21</td>
</tr>
<tr>
<td valign="middle" align="center">Average</td>
<td valign="middle" align="center">82.52</td>
<td valign="middle" align="center">89.77</td>
<td valign="middle" align="center">90.37</td>
<td valign="middle" align="center">91.17</td>
<td valign="middle" align="center">89.77</td>
</tr>
<tr>
<td valign="middle" rowspan="3" align="center">FCN</td>
<td valign="middle" align="center">Kenaf</td>
<td valign="middle" align="center">76.71</td>
<td valign="middle" align="center">88.13</td>
<td valign="middle" align="center">86.82</td>
<td valign="middle" align="center">85.54</td>
<td valign="middle" align="center">88.13</td>
<td valign="middle" rowspan="3" align="center">0.204</td>
<td valign="middle" rowspan="3" align="center">0.566</td>
</tr>
<tr>
<td valign="middle" align="center">Background</td>
<td valign="middle" align="center">84.38</td>
<td valign="middle" align="center">90.66</td>
<td valign="middle" align="center">91.53</td>
<td valign="middle" align="center">92.41</td>
<td valign="middle" align="center">90.66</td>
</tr>
<tr>
<td valign="middle" align="center">Average</td>
<td valign="middle" align="center">80.54</td>
<td valign="middle" align="center">89.39</td>
<td valign="middle" align="center">89.17</td>
<td valign="middle" align="center">88.98</td>
<td valign="middle" align="center">89.39</td>
</tr>
<tr>
<td valign="middle" rowspan="3" align="center">U-Net</td>
<td valign="middle" align="center">Kenaf</td>
<td valign="middle" align="center">81.3</td>
<td valign="middle" align="center">92.33</td>
<td valign="middle" align="center">89.68</td>
<td valign="middle" align="center">87.19</td>
<td valign="middle" align="center">92.33</td>
<td valign="middle" rowspan="3" align="center">0.244</td>
<td valign="middle" rowspan="3" align="center">0.653</td>
</tr>
<tr>
<td valign="middle" align="center">Background</td>
<td valign="middle" align="center">87.29</td>
<td valign="middle" align="center">91.49</td>
<td valign="middle" align="center">93.21</td>
<td valign="middle" align="center">95.0</td>
<td valign="middle" align="center">91.49</td>
</tr>
<tr>
<td valign="middle" align="center">Average</td>
<td valign="middle" align="center">84.29</td>
<td valign="middle" align="center">91.91</td>
<td valign="middle" align="center">91.45</td>
<td valign="middle" align="center">91.1</td>
<td valign="middle" align="center">91.91</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Among segmentation accuracy metrics, the U-Net model demonstrated the best performance, achieving an Intersection over Union (IoU) of 84.29%, an average precision of 91.91%, and an F1 score of 91.45%. This indicates that UNet effectively fuses shallow spatial information with deep semantic features through its skip-connection architecture, better preserving Kenaf leaf edge details and texture information, thereby enhancing overall segmentation quality.</p>
<p>The DeepLabv3+ model achieved an average IoU of 82.52%, slightly lower than U-Net, but its precision reached 91.17%, indicating fewer false positives in plant pixel classification. This high precision relates to the ASPP module&#x2019;s multiscale feature extraction capability, which effectively captures the complex morphology of Kenaf plants. However, DeepLabv3+&#x2019;s model execution time is approximately 2.5 times longer than the other two models, resulting in relatively low inference efficiency that is less suitable for large-scale rapid processing.</p>
<p>The FCN model achieved an average Intersection over Union (IoU) of 80.54%, performing slightly below the other two models overall. Nevertheless, it maintained a high accuracy rate and stable recall rate. Its simpler model architecture enables the fastest inference speed, making it suitable for applications requiring real-time processing or with limited computational resources.</p>
<p>A comprehensive comparison of the three models reveals that U-Net strikes a favorable balance between segmentation accuracy and computational efficiency. Notably, it achieves the highest recall rate of 92.33% in identifying Kenaf areas, demonstrating its most stable and reliable performance in the task of extracting Kenaf plants.</p>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>SE attention and AdamW optimizer synergistically enhance U-net segmentation performance</title>
<p>To enhance and optimise the model architecture, we compared UNet, SE-UNet, and SE-UNet with AdamW (<xref ref-type="table" rid="T4"><bold>Table&#xa0;4</bold></xref>), the SE module consistently achieved improvements across all metrics. mIoU increased by 1.03%, while mF1-score rose by 0.61%. Replacing SGD with AdamW further enhanced SE-UNet&#x2019;s performance, achieving an overall mIoU improvement of 1.7%, accuracy gain of 1.01%, and mF1-score increase of 0.99%.</p>
<table-wrap id="T4" position="float">
<label>Table&#xa0;4</label>
<caption>
<p>Performance comparison of the U-net model under different configurations.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Model</th>
<th valign="middle" align="center">Optimizer</th>
<th valign="middle" align="center">IoU</th>
<th valign="middle" align="center">Accuracy</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">UNet</td>
<td valign="middle" align="center">SDG</td>
<td valign="middle" align="center">84.29</td>
<td valign="middle" align="center">91.91</td>
<td valign="middle" align="center">91.45</td>
<td valign="middle" align="center">91.1</td>
<td valign="middle" align="center">91.91</td>
</tr>
<tr>
<td valign="middle" align="center">SE-UNet</td>
<td valign="middle" align="center">SDG</td>
<td valign="middle" align="center">85.32</td>
<td valign="middle" align="center">92.55</td>
<td valign="middle" align="center">92.06</td>
<td valign="middle" align="center">91.68</td>
<td valign="middle" align="center">92.55</td>
</tr>
<tr>
<td valign="middle" align="center">SE-UNet</td>
<td valign="middle" align="center">AdamW</td>
<td valign="middle" align="center">85.99</td>
<td valign="middle" align="center">92.92</td>
<td valign="middle" align="center">92.44</td>
<td valign="middle" align="center">92.08</td>
<td valign="middle" align="center">92.92</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Data augmentation strategies impact the segmentation performance of SE-UNet</title>
<p>To quantitatively assess the individual contribution of different data augmentation strategies, a series of ablation experiments were conducted using the SE-UNet model optimized with AdamW. In each experiment, only one augmentation strategy was applied while all other augmentation operations were disabled. The segmentation performance under different augmentation configurations is summarized in <xref ref-type="table" rid="T5"><bold>Table&#xa0;5</bold></xref>.</p>
<table-wrap id="T5" position="float">
<label>Table&#xa0;5</label>
<caption>
<p>Performance comparison of data augmentation strategy ablation experiments.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Model</th>
<th valign="middle" align="center">Optimizer</th>
<th valign="middle" align="center">Ablation</th>
<th valign="middle" align="center">IoU</th>
<th valign="middle" align="center">Accuracy</th>
<th valign="middle" align="center">F1-score</th>
<th valign="middle" align="center">Precision</th>
<th valign="middle" align="center">Recall</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">SE-UNet</td>
<td valign="middle" align="center">AdamW</td>
<td valign="middle" align="center">Scale change</td>
<td valign="middle" align="center">85.08</td>
<td valign="middle" align="center">92.06</td>
<td valign="middle" align="center">91.91</td>
<td valign="middle" align="center">91.77</td>
<td valign="middle" align="center">92.06</td>
</tr>
<tr>
<td valign="middle" align="center">SE-UNet</td>
<td valign="middle" align="center">AdamW</td>
<td valign="middle" align="center">Geometric inversion</td>
<td valign="middle" align="center">85.34</td>
<td valign="middle" align="center">92.82</td>
<td valign="middle" align="center">92.07</td>
<td valign="middle" align="center">91.59</td>
<td valign="middle" align="center">92.82</td>
</tr>
<tr>
<td valign="middle" align="center">SE-UNet</td>
<td valign="middle" align="center">AdamW</td>
<td valign="middle" align="center">Photometric distortion</td>
<td valign="middle" align="center">85.52</td>
<td valign="middle" align="center">92.93</td>
<td valign="middle" align="center">92.17</td>
<td valign="middle" align="center">91.69</td>
<td valign="middle" align="center">92.93</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>When scale change was applied, SE-UNet achieved an IoU of 85.08% and an F1-score of 91.91%, indicating that scale-based augmentation can moderately improve the model&#x2019;s robustness to variations in seedling size and spatial resolution. However, its contribution to overall segmentation accuracy was relatively limited compared with other augmentation strategies.The use of geometric inversion with IoU and F1-score increasing to 85.34% and 92.07%, respectively. Among the tested strategies, photometric distortion resulted in the highest segmentation performance. Under this configuration, SE-UNet achieved an IoU of 85.52%, an accuracy of 92.93%, and an F1-score of 92.17%. These results demonstrate that photometric augmentation is particularly effective in improving model robustness to illumination variability.</p>
</sec>
<sec id="s4_5">
<label>4.5</label>
<title>Relationship between plant canopy coverage and yield components in different varieties of Kenaf during the seedling stage</title>
<p>In the context of saline-alkali conditions, a notable degree of variability in growth patterns was observed among 11 Kenaf cultivars. The dry bark yield per mu fluctuated between 241.76&#x2013;538.65 kg/mu, with Xiao 3 and K5 yielding the highest, while K3 yielded the lowest.</p>
<p>During the seedling stage, the plant area exhibited marked inter-varietal variation, ranging from 0.2202 to 0.4743. Xiao 1 and Xiao 2 exhibited larger plant areas and vigorous early growth, whereas K2 and K3 displayed smaller plant areas, indicating greater susceptibility to salt stress inhibition during the seedling stage (<xref ref-type="fig" rid="f5"><bold>Figure 5</bold></xref>).</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Comparative analysis of yield-related traits among different kenaf varieties.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g005.tif">
<alt-text content-type="machine-generated">Four bar graphs display data on plant attributes for different samples labeled K1 to H368. The top left graph shows plant height in centimeters, and the top right graph shows stem thickness in millimeters. The bottom left graph displays skin thickness in millimeters, and the bottom right graph compares the weight in kilograms of fresh leaves, skin, and bones. Asterisks indicate significance levels, with one asterisk for p&lt;0.05 and two for p&lt;0.01. Each graph has bars of varying colors.</alt-text>
</graphic></fig>
<p>In comparison with conventional morphological indicators, the correlation between seedling-stage plant area and yield demonstrated enhanced stability. The sprout area exhibited a moderately strong positive correlation with dry bark yield (r &#x2248; 0.52), indicating that early canopy expansion plays a crucial role in subsequent biomass formation (<xref ref-type="fig" rid="f6"><bold>Figure 6</bold></xref>).</p>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>Comparison of plant canopy coverage and yield components of ramie seedlings.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g006.tif">
<alt-text content-type="machine-generated">Line chart comparing various attributes of kenaf varieties labeled K1 through H368. Attributes include plant height, stem thickness, skin thickness, leaf weight, skin weight, bone weight, and canopy coverage. Each attribute is represented by a different colored line with corresponding scales.</alt-text>
</graphic></fig>
</sec>
<sec id="s4_6">
<label>4.6</label>
<title>Results of remote sensing feature analysis</title>
<p>To further investigate the relationship between remote sensing characteristics and the growth status of Kenaf seedlings during the seedling stage, a comprehensive feature analysis framework was employed, integrating correlation analysis and multi-method feature selection. The three types of remote sensing characteristics used in this analysis were detailed in our previous study (<xref ref-type="bibr" rid="B7">Fu et&#xa0;al., 2024</xref>). The plant canopy coverage derived from optimal semantic segmentation was used as the target variable for subsequent analysis.</p>
<p>Initially, Pearson correlation Analysis results indicate that color characteristics and vegetation index features generally exhibit high correlations with plant area, demonstrating that spectral and color features extracted from imagery can effectively reflect differences in plant canopy coverage and biomass. Among these, 18 feature indicators showed extremely high correlations with plant area (correlation coefficient |r| &gt; 0.8) (<xref ref-type="fig" rid="f7"><bold>Figure 7</bold></xref>).</p>
<fig id="f7" position="float">
<label>Figure&#xa0;7</label>
<caption>
<p>Correlation analysis heatmap: <bold>(a)</bold> Color characteristics; <bold>(b)</bold> Texture characteristics; <bold>(c)</bold> Vegetation characteristics.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-17-1747657-g007.tif">
<alt-text content-type="machine-generated">Three heatmaps (a), (b) and (c) respectively present the correlation analysis results between the colour index, texture index, vegetation index and plant cover. Each heatmap employs a colour scale ranging from -1.0 (blue) to 1.0 (red) to visually represent the strength and spatial distribution patterns of correlations between indices and coverage across different locations. Red areas indicate a strong positive correlation between the index and plant coverage, while blue areas denote negative correlations or no relationship. This clearly reveals the varying contributions of different characteristics towards characterising vegetation conditions.</alt-text>
</graphic></fig>
<p>These indicators primarily originate from combinations of reflectance values in the visible and near-infrared bands. Vegetation indices such as NDRE, GNDVI, and GRNDVI effectively reflect chlorophyll content and canopy coverage in Kenaf leaves, while color features like ExG, ExGR, and RGRI demonstrate enhanced green characteristics in the visible spectrum, showing significant correlation with plant area (<xref ref-type="table" rid="T6"><bold>Table&#xa0;6</bold></xref>).</p>
<table-wrap id="T6" position="float">
<label>Table&#xa0;6</label>
<caption>
<p>Remote sensing features and their calculation formulas.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Abbreviation</th>
<th valign="middle" align="center">Full Name</th>
<th valign="middle" align="center">Calculation Formulas</th>
<th valign="middle" align="center">Serial Number</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="center">B</td>
<td valign="middle" align="center">Blue band</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im6"><mml:mrow><mml:mstyle mathsize="normal"><mml:mi>B</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mstyle mathsize="normal"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathsize="normal"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:munder><mml:msub><mml:mstyle mathsize="normal"><mml:mi>P</mml:mi></mml:mstyle><mml:mrow><mml:mstyle mathsize="normal"><mml:mi>i</mml:mi></mml:mstyle><mml:mo>,</mml:mo><mml:mstyle mathsize="normal"><mml:mi>j</mml:mi></mml:mstyle></mml:mrow></mml:msub><mml:mo stretchy="false">/</mml:mo><mml:msup><mml:mstyle mathsize="normal"><mml:mi>N</mml:mi></mml:mstyle><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(10)</td>
</tr>
<tr>
<td valign="middle" align="center">H</td>
<td valign="middle" align="center">Hue</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im7"><mml:mrow><mml:mi>H</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:munder><mml:msub><mml:mi>P</mml:mi><mml:mi>H</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(11)</td>
</tr>
<tr>
<td valign="middle" align="center">RGRI</td>
<td valign="middle" align="center">Red-Green Ratio Index</td>
<td valign="middle" align="center">RGRI=R/G</td>
<td valign="middle" align="center">(12)</td>
</tr>
<tr>
<td valign="middle" align="center">WI<sub>std</sub></td>
<td valign="middle" align="center">Width Index Standard Deviation</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im8"><mml:mrow><mml:mi>W</mml:mi><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>s</mml:mi><mml:mi>t</mml:mi><mml:mi>d</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msqrt><mml:mrow><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:msup><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>W</mml:mi><mml:mi>I</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mi>W</mml:mi><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:msqrt></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(13)</td>
</tr>
<tr>
<td valign="middle" align="center">NGRDI</td>
<td valign="middle" align="center">Normalized Green-Red Difference Index</td>
<td valign="middle" align="center">NGRDI=(G-R)/(G+R)</td>
<td valign="middle" align="center">(14)</td>
</tr>
<tr>
<td valign="middle" align="center">VARI</td>
<td valign="middle" align="center">Visible Atmospherically Resistant Index</td>
<td valign="middle" align="center">VARI=(G-R)/(G+R-B)</td>
<td valign="middle" align="center">(15)</td>
</tr>
<tr>
<td valign="middle" align="center">MGVRI</td>
<td valign="middle" align="center">Modified Green-Red Vegetation Index</td>
<td valign="middle" align="center">MGVRI=(G2- R2)/(G2 + R2)</td>
<td valign="middle" align="center">(16)</td>
</tr>
<tr>
<td valign="middle" align="center">GRVI</td>
<td valign="middle" align="center">Green-Red Vegetation Index</td>
<td valign="middle" align="center">GRVI=(G-R)/(G+R)</td>
<td valign="middle" align="center">(17)</td>
</tr>
<tr>
<td valign="middle" align="center">NDGI</td>
<td valign="middle" align="center">Normalized Difference Greenness Index</td>
<td valign="middle" align="center">NDGI=(G-R)/(G+R)</td>
<td valign="middle" align="center">(18)</td>
</tr>
<tr>
<td valign="middle" align="center">GRNDVI</td>
<td valign="middle" align="center">Green-Red Normalized Difference Vegetation Index</td>
<td valign="middle" align="center">GRNDVI=/(NIR/R)*[(NIR-R)/(NIR+R)+1]</td>
<td valign="middle" align="center">(19)</td>
</tr>
<tr>
<td valign="middle" align="center">NDRE</td>
<td valign="middle" align="center">Normalized Difference Red Edge Index</td>
<td valign="middle" align="center">NDRE=(NIR-RE)/(NIR+RE)</td>
<td valign="middle" align="center">(20)</td>
</tr>
<tr>
<td valign="middle" align="center">WI<sub>mean</sub></td>
<td valign="middle" align="center">Width Index Mean</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im9"><mml:mrow><mml:mi>W</mml:mi><mml:msub><mml:mi>I</mml:mi><mml:mrow><mml:mi>m</mml:mi><mml:mi>e</mml:mi><mml:mi>a</mml:mi><mml:mi>n</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mi>N</mml:mi></mml:mfrac><mml:munderover><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mi>N</mml:mi></mml:munderover><mml:mi>W</mml:mi><mml:mi>I</mml:mi><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(21)</td>
</tr>
<tr>
<td valign="middle" align="center">a</td>
<td valign="middle" align="center">color channel a average</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im10"><mml:mrow><mml:mi>a</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi>p</mml:mi></mml:msub></mml:mrow></mml:mfrac><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo><mml:mo>&#x2208;</mml:mo><mml:mi>m</mml:mi><mml:mi>a</mml:mi><mml:mi>s</mml:mi><mml:mi>k</mml:mi></mml:mrow></mml:munder><mml:msub><mml:mi>p</mml:mi><mml:mi>a</mml:mi></mml:msub><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(22)</td>
</tr>
<tr>
<td valign="middle" align="center">ExR</td>
<td valign="middle" align="center">Excess Red Index</td>
<td valign="middle" align="center">EXR=1.4R-G</td>
<td valign="middle" align="center">(23)</td>
</tr>
<tr>
<td valign="middle" align="center">ExGR</td>
<td valign="middle" align="center">Excess Green minus Excess Red Index</td>
<td valign="middle" align="center">ExGR=ExG-1.4R-G</td>
<td valign="middle" align="center">(24)</td>
</tr>
<tr>
<td valign="middle" align="center">GNDVI</td>
<td valign="middle" align="center">Green Normalized Difference Vegetation Index</td>
<td valign="middle" align="center">GNDVI=(NIR&#x2013;G)/(NIR+G)</td>
<td valign="middle" align="center">(25)</td>
</tr>
<tr>
<td valign="middle" align="center">G</td>
<td valign="middle" align="center">Green band</td>
<td valign="middle" align="center"><inline-formula>
<mml:math display="inline" id="im11"><mml:mrow><mml:mi>G</mml:mi><mml:mo>=</mml:mo><mml:munder><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:munder><mml:msub><mml:mi>P</mml:mi><mml:mrow><mml:mi>i</mml:mi><mml:mo>,</mml:mo><mml:mi>j</mml:mi></mml:mrow></mml:msub><mml:mo stretchy="false">/</mml:mo><mml:msup><mml:mi>N</mml:mi><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula></td>
<td valign="middle" align="center">(26)</td>
</tr>
<tr>
<td valign="middle" align="center">COM</td>
<td valign="middle" align="center">Combination Coefficient</td>
<td valign="middle" align="center">COM=0.25ExG+0.3ExGR+0.33CIVE+0.12VEG</td>
<td valign="middle" align="center">(27)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>i denotes the number of rows of pixels, j denotes the number of columns of pixels, P(i, j) denotes the grayscale value at position (i, j) in the image, x and y denote the variances of px(i) and px(j), N denotes the number of rows or columns, and RE, NIR, R, B, G denote the average reflectance of the red edge band, near-infrared band, red band, blue band, and green band respectively.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>To more effectively reduce feature redundancy and enhance the robustness of feature selection, this study further integrated three complementary feature selection methods: LASSO, LARS, and CARS. By comprehensively comparing the features selected by different methods, we extracted common key variables, thereby revealing important feature factors capable of stably and consistently predicting plant canopy coverage.</p>
<p>Comparative results demonstrate that features selected by LARS highly overlap with correlation analysis outcomes, including: RGRI, H, WI<sub>mean</sub>, B, a, ExR, WI<sub>std</sub>, and G, indicating these variables exhibit stable strong linear correlations. Similarly, features chosen by CARS&#x2014;COM, ExR, ExGR, B, a, and NDRE&#x2014;also share substantial overlap with correlation analysis. Moreover, both LARS and CARS jointly selected the four features ExR, B, S, and a, demonstrating their significant robustness across two distinct feature selection methods. Comprehensive analysis indicates that correlation analysis, LARS, and CARS consistently identified three core features: B, a, and ExR. In contrast, the feature subset selected by LASSO was more streamlined, comprising only ExR, B, and H.</p>
</sec>
</sec>
<sec id="s5" sec-type="discussion">
<label>5</label>
<title>Discussion</title>
<sec id="s5_1">
<label>5.1</label>
<title>Comparative performance of semantic segmentation models for Kenaf seedlings</title>
<p>This study systematically compares three representative semantic segmentation models&#x2014;FCN, DeepLabv3+, and U-Net&#x2014;regarding their efficacy in extracting ramie seedlings from high-resolution aerial imagery. Whilst all models successfully separate seedlings from soil backgrounds, performance variations stem from architectural characteristics and feature representation mechanisms.</p>
<p>As a benchmark end-to-end segmentation framework, FCN demonstrates stable performance and high computational efficiency. However, its limited feature fusion capability constrains its handling of fine boundaries and densely distributed seedlings. DeepLabv3+, leveraging dilated convolutions and multi-scale contextual modelling, excels in small object recognition accuracy. Yet, its high computational complexity prolongs inference time, limiting its application in large-scale or real-time processing.</p>
<p>U-Net achieves the optimal balance between segmentation accuracy and efficiency. Its encoder-decoder architecture with skip connections effectively preserves spatial details while integrating high-level semantic information, making it particularly well-suited for ramie seedling segmentation tasks&#x2014;a scenario characterised by fragmented plant morphology, significant size variation, and strong background interference. These results confirm U-Net as a reliable benchmark solution for UAV-based crop seedling segmentation tasks.</p>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Effectiveness of SE attention and optimization strategy enhancement</title>
<p>To further enhance segmentation performance, we introduced a squeeze-excited (SE) channel attention module within the UNet architecture. This mechanism enables the network to amplify information-rich spectral and textural channels while suppressing background noise through adaptive calibration of channel-level feature responses. Improvements in intersection-over-union (IoU) and F1 scores demonstrate that channel attention effectively enhances the ability to distinguish small-scale features.</p>
<p>Furthermore, replacing the SGD optimiser with AdamW further elevates model performance. AdamW decouples weight decay from gradient updates, thereby achieving more stable convergence and stronger generalisation capabilities.</p>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Influence of data augmentation strategies on model robustness</title>
<p>Data augmentation ablation experiments demonstrate that different augmentation strategies contribute disparately to segmentation performance. Among the tested strategies, photometric distortion yields the most pronounced improvement, followed by geometric inversion and scale variation. This phenomenon indicates that variability introduced by changing illumination conditions&#x2014;such as solar angle and shading&#x2014;constitutes a key source of uncertainty affecting segmentation outcomes in drone-captured images of red hemp seedlings.</p>
</sec>
<sec id="s5_4">
<label>5.4</label>
<title>Early salt&#x2013;alkali adaptability and growth structural traits of Kenaf varieties</title>
<p>Under saline-alkali stress, significant variations were observed among different Kenaf varieties in terms of seedling canopy development, stem structure, and yield factors. Varieties exhibiting greater canopy coverage typically demonstrated stronger early growth vigour and higher final dry bark yields, indicating superior early salt tolerance. The contribution of structural traits to yield varied by cultivar: Xiao3 achieved the highest dry bark yield despite possessing medium stem thickness and relatively thin bark; K5 attained high dry bark yield despite lower fresh bark weight, reflecting its outstanding dehydration and biomass conversion efficiency under salt stress. In summary, under the experimental conditions, Xiao3, K5 and Xiao2 demonstrated strong salt tolerance, whereas K3 and K6 exhibited greater sensitivity to salt stress.</p>
</sec>
<sec id="s5_5">
<label>5.5</label>
<title>Implications of multi-method remote sensing feature selection</title>
<p>Building upon semantic segmentation, this study systematically identified the four key radiometric features most strongly correlated with ramie seedling canopy coverage through correlation analysis and multiple feature selection (LASSO, LARS, CARS): B, a, H, and ExR. The robustness demonstrated by these features across different methodologies supports their potential as reliable phenotypic indicators. Future work may integrate precise segmentation results with the selected features to achieve high-precision, non-destructive monitoring of crop growth in saline-alkali environments.</p>
</sec>
<sec id="s5_6">
<label>5.6</label>
<title>Research limitations and future prospects</title>
<p>Although this study systematically evaluated three mainstream semantic segmentation models for Kenaf seedling identification, several limitations should be acknowledged. First, the dataset primarily consisted of UAV imagery from a single growth stage, lacking multi-temporal data needed to comprehensively capture dynamic growth patterns. Second, the limited number of training samples and influences such as varying illumination and shadow conditions may have constrained model generalizability. Future research could advance this work in the following directions:</p>
<list list-type="simple">
<list-item>
<p>(1) Constructing multi-temporal Kenaf growth monitoring datasets to support time-series growth analysis;</p></list-item>
<list-item>
<p>(2) Incorporating emerging network architectures such as Transformer and ConvNeXt to improve recognition performance under complex backgrounds and for small plant objects;</p></list-item>
<list-item>
<p>(3) Integrating remote sensing features with meteorological data and field-measured traits to establish accurate yield prediction models for Kenaf.</p></list-item>
<list-item>
<p>(4) The utilisation of images captured from multiple angles, such as 30&#xb0; and 60&#xb0;, facilitates a more comprehensive documentation of the plant&#x2019;s additional morphological characteristics, thereby potentially enhancing the accuracy of identification.</p></list-item>
<list-item>
<p>(5) Validating Kenaf salt-alkali tolerance by integrating definitive physiological indicators (e.g., ion concentration, photosynthetic efficiency) with image-derived phenotypes to establish a reliable high-throughput phenotyping pipeline.</p></list-item>
</list>
</sec>
</sec>
<sec id="s6" sec-type="conclusions">
<label>6</label>
<title>Conclusion</title>
<p>This study developed a semantic segmentation framework based on drone imagery for the precise identification and quantification of canopy coverage in Kenaf seedlings within saline-alkali fields. Through systematic comparison of classical convolutional neural network architectures, U-Net was found to achieve the optimal balance between segmentation accuracy, boundary preservation, and computational efficiency, rendering it suitable for extracting early-stage seedling canopy coverage from high-resolution RGB imagery.</p>
<p>Building upon the U-Net baseline model, the introduction of a channel attention mechanism (SE module) and the AdamW optimiser further enhanced the model&#x2019;s feature discrimination capabilities under complex lighting conditions and background interference. Data augmentation ablation experiments demonstrated that photometric distortion significantly contributed to improved model robustness, underscoring the necessity of employing light-aware training strategies in UAV agricultural applications.</p>
<p>Based on the optimised segmentation results, this study quantitatively extracted seedling canopy coverage and discovered a significant positive correlation with final dry bark yield. This indicates that early canopy development serves as an effective indicator reflecting plant vigour and production potential. Further feature selection employing correlation analysis, LASSO, LARS, and CARS methods consistently identified a small set of key colour-related features: B, a, H, and ExR. The high consistency across different selection methods enhances the reliability of these features for non-destructive phenotyping.</p>
<p>In summary, this study established an accurate and efficient early-stage phenotyping framework for ramie by integrating semantic segmentation, attention mechanism enhancement, data optimisation strategies, and multi-method feature selection. This framework provides a practical and scalable technical solution for drone-based crop growth monitoring in saline-alkali environments, while also offering robust methodological support for precision cultivation and salt-tolerant variety breeding in ramie.</p>
</sec>
</body>
<back>
<sec id="s7" sec-type="data-availability">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.</p></sec>
<sec id="s8" sec-type="author-contributions">
<title>Author contributions</title>
<p>WW: Writing &#x2013; review &amp; editing, Writing &#x2013; original draft, Conceptualization. KC: Methodology, Writing &#x2013; original draft, Project administration. GL: Writing &#x2013; review &amp; editing, Investigation. RH: Investigation, Writing &#x2013; review &amp; editing. JN: Formal Analysis, Data curation, Writing &#x2013; original draft. JZ: Writing &#x2013; original draft, Formal Analysis. JL: Writing &#x2013; original draft, Visualization. GC: Funding acquisition, Writing &#x2013; review &amp; editing, Supervision. XA: Supervision, Writing &#x2013; review &amp; editing, Resources. WS: Funding acquisition, Writing &#x2013; review &amp; editing, Supervision.</p></sec>
<ack>
<title>Acknowledgments</title>
<p>We sincerely thank Hunan Agricultural University for their invaluable technical assistance throughout this research. We also deeply appreciate the collaborative efforts of all team members.</p>
</ack>
<sec id="s10" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p></sec>
<sec id="s11" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec id="s12" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec>
<ref-list>
<title>References</title>
<ref id="B500">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Alkhudaydi</surname> <given-names>T.</given-names></name>
<name><surname>Reynolds</surname> <given-names>D.</given-names></name>
<name><surname>Griffiths</surname> <given-names>S.</given-names></name>
<name><surname>Zhou</surname> <given-names>J.</given-names></name>
<name><surname>de la Iglesia</surname> <given-names>B.</given-names></name>
</person-group>(<year>2019</year>). 
<article-title>An Exploration of Deep-Learning Based Phenotypic Analysis to Detect Spike Regions in Field Conditions for UK Bread Wheat</article-title>. <source>Plant Phenomics</source> <volume>2019</volume>, <elocation-id>17</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2019/7368761</pub-id>, PMID: <pub-id pub-id-type="pmid">33313535</pub-id>
</mixed-citation>
</ref>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Chen</surname> <given-names>C.</given-names></name>
<name><surname>Cui</surname> <given-names>R.</given-names></name>
<name><surname>An</surname> <given-names>X.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Effects of soil amendments and coverings on the kenaf yield and soil physicochemical properties in saline-alkali land</article-title>. <source>Notulae Botanicae Horti Agrobotanici Cluj-Napoca</source> <volume>52</volume>, <elocation-id>13897</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.15835/nbha52313897</pub-id>
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Chen</surname> <given-names>L. C.</given-names></name>
<name><surname>Zhu</surname> <given-names>Y.</given-names></name>
<name><surname>Papandreou</surname> <given-names>G.</given-names></name>
<name><surname>Schroff</surname> <given-names>F.</given-names></name>
<name><surname>Adam</surname> <given-names>H</given-names></name>
</person-group>. (<year>2018</year>). <source>Encoder-decoder with atrous separable convolution for semantic image segmentation</source> (<publisher-loc>Cham</publisher-loc>: 
<publisher-name>Springer</publisher-name>). doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-030-01234-2_49</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ding</surname> <given-names>Y.</given-names></name>
<name><surname>Yang</surname> <given-names>W.</given-names></name>
<name><surname>Zhang</surname> <given-names>J.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>An improved DeepLabv3+based approach for disease spot segmentation on apple leaves</article-title>. <source>Computers&amp;Electronics Agric.</source> <volume>231</volume>, <fpage>110041</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2025.110041</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Dosovitskiy</surname> <given-names>A.</given-names></name>
<name><surname>Beyer</surname> <given-names>L.</given-names></name>
<name><surname>Kolesnikov</surname> <given-names>A.</given-names></name>
<name><surname>Weissenborn</surname> <given-names>D.</given-names></name>
<name><surname>Zhai</surname> <given-names>X.</given-names></name>
<name><surname>Unterthiner</surname> <given-names>T</given-names></name>
</person-group>. (<year>2020</year>). "
<article-title>An image is worth 16x16 words: transformers for image recognition at scale</article-title>. "<source>in International Conference on Learning Representations (ICLR 2021)</source>, <page-range>611&#x2013;631</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2010.11929</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Feng</surname> <given-names>L.</given-names></name>
<name><surname>Li</surname> <given-names>H.</given-names></name>
<name><surname>Gao</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>Y</given-names></name>
</person-group>. (<year>2020</year>). 
<article-title>A color image segmentation method based on region salient color and fuzzy C-means algorithm</article-title>. <source>Circuits Syst. Signal Process</source> <volume>39</volume>, <fpage>586</fpage>&#x2013;<lpage>610</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s00034-019-01126-w</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Fu</surname> <given-names>H.</given-names></name>
<name><surname>Chen</surname> <given-names>J.</given-names></name>
<name><surname>Lu</surname> <given-names>J.</given-names></name>
<name><surname>Yue</surname> <given-names>Y.</given-names></name>
<name><surname>Xu</surname> <given-names>M.</given-names></name>
<name><surname>Jiao</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>A comparison of different remote sensors for ramie leaf area index estimation</article-title>. <source>Agronomy</source> <volume>13</volume>, <elocation-id>899</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy13030899</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Fu</surname> <given-names>H.</given-names></name>
<name><surname>Lu</surname> <given-names>J.</given-names></name>
<name><surname>Cui</surname> <given-names>G.</given-names></name>
<name><surname>Nie</surname> <given-names>J.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>She</surname> <given-names>W.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Advanced plant phenotyping: unmanned aerial vehicle remote sensing and cimageA software technology for precision crop growth monitoring</article-title>. <source>Agronomy</source> <volume>14</volume>, <elocation-id>2534</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy14112534</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Guo</surname> <given-names>Z.</given-names></name>
<name><surname>Cai</surname> <given-names>D.</given-names></name>
<name><surname>Jin</surname> <given-names>Z.</given-names></name>
<name><surname>Xu</surname> <given-names>T.</given-names></name>
<name><surname>Yu</surname> <given-names>F.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Research on unmanned aerial vehicle(UAV)rice field weed sensing image segmentation method based on CNN-transformer</article-title>. <source>Comput. Electron. Agric.</source> <volume>229</volume>, <elocation-id>109719</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.109719</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hassani</surname> <given-names>A.</given-names></name>
<name><surname>Azapagic</surname> <given-names>A.</given-names></name>
<name><surname>Shokri</surname> <given-names>N.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Global predictions of primary soil salinization under changing climate in the 21st century</article-title>. <source>Nat. Commun.</source> <volume>12</volume>, <fpage>6663</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/S41467-021-26907-3</pub-id>, PMID: <pub-id pub-id-type="pmid">34795219</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hu</surname> <given-names>J.</given-names></name>
<name><surname>Shen</surname> <given-names>L.</given-names></name>
<name><surname>Sun</surname> <given-names>G.</given-names></name>
<name><surname>Albanie</surname> <given-names>S.</given-names></name>
<name><surname>Sun</surname> <given-names>G.</given-names></name>
<name><surname>Wu</surname> <given-names>E.</given-names></name>
<etal/>
</person-group>. (<year>2017</year>). 
<article-title>Squeeze-and-excitation networks</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>99</volume>, <page-range>2011&#x2013;2023</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2019.2913372</pub-id>, PMID: <pub-id pub-id-type="pmid">31034408</pub-id>
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jia</surname> <given-names>X.</given-names></name>
<name><surname>Lei</surname> <given-names>T.</given-names></name>
<name><surname>Du</surname> <given-names>X.</given-names></name>
<name><surname>Liu</surname> <given-names>S.</given-names></name>
<name><surname>Meng</surname> <given-names>H.</given-names></name>
<name><surname>Nandi</surname> <given-names>A.</given-names></name>
<etal/>
</person-group>. (<year>2020</year>). 
<article-title>Robust self-sparse fuzzy clustering for image segmentation</article-title>. <source>IEEE Access</source> <volume>8</volume>, <fpage>146182</fpage>&#x2013;<lpage>146195</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ACCESS.2020.3015270</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jiang</surname> <given-names>K.</given-names></name>
<name><surname>Wenjiang</surname> <given-names>H.</given-names></name>
<name><surname>Ping</surname> <given-names>H.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Dual-task segmentation of oilseed rape and weeds in agricultural fields: A hybrid approach combining enhanced UNet and unsupervised clustering</article-title>. <source>Computers&amp;Electronics Agric.</source> <volume>238</volume>, <fpage>110827</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2025.110827</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jin</surname> <given-names>Z.</given-names></name>
<name><surname>Guo</surname> <given-names>S.</given-names></name>
<name><surname>Li</surname> <given-names>S.</given-names></name>
<name><surname>Yu</surname> <given-names>F.</given-names></name>
<name><surname>Xu</surname> <given-names>T.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Research on the rice fertiliser decision-making method based on uav remote sensing data assimilation</article-title>. <source>Comput. Electron. Agric.</source> <volume>216</volume>, <elocation-id>108508</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/J.COMPAG.2023.108508</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Kattenborn</surname> <given-names>T.</given-names></name>
<name><surname>Leitloff</surname> <given-names>J.</given-names></name>
<name><surname>Schiefer</surname> <given-names>F.</given-names></name>
<name><surname>Hinz</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Review on convolutional neural networks(CNN)in vegetation remote sensing</article-title>. <source>ISPRS J. Photogrammetry Remote Sens.</source> <volume>173</volume>, <page-range>24&#x2013;49</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.isprsjprs.2020.12.010</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lamichhane</surname> <given-names>J. R.</given-names></name>
<name><surname>Debaeke</surname> <given-names>P.</given-names></name>
<name><surname>Steinberg</surname> <given-names>C.</given-names></name>
<name><surname>Ming</surname> <given-names>Y.</given-names></name>
<name><surname>Barbetti</surname> <given-names>M.</given-names></name>
<name><surname>Aubertot</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2018</year>). 
<article-title>Abiotic and biotic factors affecting crop seed germination and seedling emergence: a conceptual framework</article-title>. <source>Plant Soil</source> <volume>432</volume>, <fpage>1</fpage>&#x2013;<lpage>28</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11104-018-3780-9</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lapkovskis</surname> <given-names>A.</given-names></name>
<name><surname>Nefedova</surname> <given-names>N.</given-names></name>
<name><surname>Beikmohammadi</surname> <given-names>A.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Automatic fused multimodal deep learning for plant identification</article-title>. <source>Front. Plant Sci Aug</source> <volume>5.16</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2025.1616020</pub-id>, PMID: <pub-id pub-id-type="pmid">40838081</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Long</surname> <given-names>J.</given-names></name>
<name><surname>Shelhamer</surname> <given-names>E.</given-names></name>
<name><surname>Darrell</surname> <given-names>T.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>Fully convolutional networks for semantic segmentation</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>39</volume>, <fpage>640</fpage>&#x2013;<lpage>651</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2015.7298965</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lucas</surname> <given-names>W.</given-names></name>
<name><surname>Bj&#xf6;rn</surname> <given-names>W.</given-names></name>
<name><surname>Thomas</surname> <given-names>J</given-names></name>
</person-group>. (<year>2025</year>). 
<article-title>Multi-Modal Vision Transformer for high-resolution soil texture prediction of German agricultural soils using remote sensing imagery</article-title>. <source>Remote Sens. Environ.</source> <volume>331</volume>, <elocation-id>114985</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.rse.2025.114985</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ma</surname> <given-names>X.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Pun</surname> <given-names>M.</given-names></name>
<name><surname>Liu</surname> <given-names>M.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>A multilevel multimodal fusion transformer for remote sensing semantic segmentation</article-title>. <source>IEEE Trans. Geosci. Remote Sens.</source> <volume>62</volume>, <fpage>1</fpage>&#x2013;<lpage>15</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TGRS.2024.3373033</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Rahman</surname> <given-names>T. A. E.</given-names></name>
<name><surname>Oirdi</surname> <given-names>M. E.</given-names></name>
<name><surname>Gonzalez-Lamothe</surname> <given-names>R.</given-names></name>
<name><surname>Bouarab</surname> <given-names>K</given-names></name>
</person-group>. (<year>2002</year>). 
<article-title>Necrotrophic pathogens use the salicylic acid signaling pathway to promote disease development in tomato</article-title>. <source>Mol. Plant Microbe Interact.</source> <volume>25</volume>, <fpage>1584</fpage>&#x2013;<lpage>1593</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1094/MPMI-07-12-0187-R</pub-id>, PMID: <pub-id pub-id-type="pmid">22950753</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Riehle</surname> <given-names>D.</given-names></name>
<name><surname>Reiser</surname> <given-names>D.</given-names></name>
<name><surname>Griepentrog</surname> <given-names>H.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Robust index-based semantic plant/background segmentation for rgb-images</article-title>. <source>Comput. Electron. Agric.</source> <volume>169</volume>, <fpage>12</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2019.105201</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ronneberger</surname> <given-names>O.</given-names></name>
<name><surname>Fischer</surname> <given-names>P.</given-names></name>
<name><surname>Brox</surname> <given-names>T.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>U-net: convolutional networks for biomedical image segmentation</article-title>. <source>Springer Int. Publishing</source> <volume>9351</volume>, <page-range>234&#x2013;241</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sadashivan</surname> <given-names>S.</given-names></name>
<name><surname>Bhattacherjee</surname> <given-names>S.</given-names></name>
<name><surname>Priyanka</surname> <given-names>G.</given-names></name>
<name><surname>Pochamuthu</surname> <given-names>R.</given-names></name>
<name><surname>Kholova</surname> <given-names>J</given-names></name>
</person-group>. (<year>2021</year>). 
<article-title>Fully automated region of interest segmentation pipeline for UAV based RGB images</article-title>. <source>Biosyst. Eng.</source> <volume>211</volume>:<page-range>192&#x2013;204</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/J.BIOSYSTEMSENG.2021.08.032</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Li</surname> <given-names>Y.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<name><surname>Wang</surname> <given-names>A.</given-names></name>
<name><surname>Ma</surname> <given-names>C.</given-names></name>
<name><surname>Song</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Crop disease recognition method based on improved vision transformer network</article-title>. <source>J. Chin. Comput. Syst.</source> <volume>45</volume>, <fpage>887</fpage>&#x2013;<lpage>893</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.20009/j.cnki.21-1106/TP.2022-0605</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Qin</surname> <given-names>F.</given-names></name>
</person-group> (<year>2022</year>a). 
<article-title>Evaluation of a deep-learning model for multispectral remote sensing of land use and crop classification</article-title>. <source>Crop J.</source> <volume>10</volume>, <fpage>1435</fpage>&#x2013;<lpage>1451</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.cj.2022.01.009</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Qin</surname> <given-names>F.</given-names></name>
</person-group> (<year>2022</year>b). 
<article-title>Deep segmentation and classification of complex crops using multi-feature satellite imagery</article-title>. <source>Comput. Electron. Agric.</source> <volume>200</volume>, <elocation-id>107249</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2022.107249</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Webber</surname> <given-names>C. L.</given-names></name>
<name><surname>Bledsoe</surname> <given-names>V. K.</given-names></name>
<name><surname>Bledsoe</surname> <given-names>R. E.</given-names></name>
</person-group> (<year>2002</year>). &#x201c;
<article-title>Kenaf harvesting and processing</article-title>,&#x201d; in <source>In trends in new crops and new uses</source>. Eds. 
<person-group person-group-type="editor">
<name><surname>Janick</surname> <given-names>J.</given-names></name>
<name><surname>Whipkey</surname> <given-names>A.</given-names></name>
</person-group> Dordrecht: (
<publisher-name>ASHS Press</publisher-name>, <publisher-loc>Alexandria, VA, USA</publisher-loc>), <fpage>340</fpage>&#x2013;<lpage>347</lpage>.
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yao</surname> <given-names>J.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>Fu</surname> <given-names>H.</given-names></name>
<name><surname>Deng</surname> <given-names>Z.</given-names></name>
<name><surname>Cui</surname> <given-names>G.</given-names></name>
<name><surname>Wang</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Automated measurement of field crop phenotypic traits using UAV 3D point clouds and an improved PointNet+</article-title>. <source>Front. Plant Sci.</source> <volume>16</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2025.1654232</pub-id>, PMID: <pub-id pub-id-type="pmid">41019733</pub-id>
</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zeng</surname> <given-names>F.</given-names></name>
<name><surname>Wang</surname> <given-names>R.</given-names></name>
<name><surname>Jiang</surname> <given-names>Y.</given-names></name>
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Ding</surname> <given-names>Y.</given-names></name>
<name><surname>Dong</surname> <given-names>W.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Growth monitoring of rapeseed seedlings in multiple growth stages based on low-altitude remote sensing and semantic segmentation</article-title>. <source>Comput. Electron. Agric.</source> <volume>232</volume>, <elocation-id>110135</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/J.COMPAG.2025.110135</pub-id>
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>W.</given-names></name>
<name><surname>Xia</surname> <given-names>X.</given-names></name>
<name><surname>Zhou</surname> <given-names>G.</given-names></name>
<name><surname>Du</surname> <given-names>J.</given-names></name>
<name><surname>Chen</surname> <given-names>T.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<etal/>
</person-group>. (<year>2022</year>). 
<article-title>Research on the identification and detection of field pests in the complex background based on the rotation detection algorithm</article-title>. <source>Front. Plant Sci.</source> <volume>13</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2022.1011499</pub-id>, PMID: <pub-id pub-id-type="pmid">36582640</pub-id>
</mixed-citation>
</ref>
<ref id="B501">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>Y.</given-names></name>
<name><surname>Lv</surname> <given-names>C.</given-names></name>
</person-group>(<year>2024</year>). 
<article-title>TinySegformer: A lightweight visual segmentation model for real-time agricultural pest detection</article-title>. <source>Comput. Electron. Agric.</source> <volume>218</volume>, <fpage>108740</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2024.108740</pub-id>
</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/172404">Alejandro Isabel Luna-Maldonado</ext-link>, Autonomous University of Nuevo Le&#xf3;n, Mexico</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/277388">Jiangwei Qiao</ext-link>, Chinese Academy of Agricultural Sciences, China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1584721">Zhenguo Zhang</ext-link>, Xinjiang Agricultural University, China</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2904301">Qiyuan Wang</ext-link>, Jiangsu Ocean University, China</p></fn>
</fn-group>
</back>
</article>