<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="review-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Plant Sci.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Plant Science</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Plant Sci.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">1664-462X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpls.2025.1732979</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Artificial intelligence in plant science: from image-based phenotyping to yield and trait prediction</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes" equal-contrib="yes">
<name><surname>Wang</surname><given-names>Tong</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>*</sup></xref>
<xref ref-type="author-notes" rid="fn003"><sup>&#x2020;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/816887/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Tong</surname><given-names>Ran</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn003"><sup>&#x2020;</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/2931252/overview"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Xu</surname><given-names>Ting</given-names></name>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Li</surname><given-names>Yue</given-names></name>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Chen</surname><given-names>Yonghao</given-names></name>
<xref ref-type="aff" rid="aff5"><sup>5</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &amp; editing</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Plant Science and Landscape Architecture, University of Connecticut</institution>, <city>Storrs</city>, <state>CT</state>,&#xa0;<country country="us">United States</country></aff>
<aff id="aff2"><label>2</label><institution>Mathematics and Statistics Department, University of Texas at Dallas</institution>, <city>Richardson</city>, <state>TX</state>,&#xa0;<country country="us">United States</country></aff>
<aff id="aff3"><label>3</label><institution>Department of Computer Science, University of Massachusetts Boston</institution>, <city>Boston</city>, <state>MA</state>,&#xa0;<country country="us">United States</country></aff>
<aff id="aff4"><label>4</label><institution>Department of Computer Science, Purdue University</institution>, <city>West Lafayette</city>, <state>IN</state>,&#xa0;<country country="us">United States</country></aff>
<aff id="aff5"><label>5</label><institution>San Diego Health, University of San Diego</institution>, <city>San Diego</city>, <state>CA</state>,&#xa0;<country country="us">United States</country></aff>
<author-notes>
<corresp id="c001"><label>*</label>Correspondence: Tong Wang, <email xlink:href="mailto:wangtongnelly@gmail.com">wangtongnelly@gmail.com</email></corresp>
<fn fn-type="equal" id="fn003">
<label>&#x2020;</label>
<p>These authors have contributed equally to this work</p></fn>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-01-29">
<day>29</day>
<month>01</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>16</volume>
<elocation-id>1732979</elocation-id>
<history>
<date date-type="received">
<day>27</day>
<month>10</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>26</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>06</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Wang, Tong, Xu, Li and Chen.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Wang, Tong, Xu, Li and Chen</copyright-holder>
<license>
<ali:license_ref start_date="2026-01-29">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>With the development of artificial intelligence (AI) in complicated imaging and remote sensing technologies, plant research is transitioning from manual measurements to automated data collecting. High-throughput image-based phenotyping enables the precise and automated acquisition of traits across various spatial and temporal scales, ranging from controlled laboratory settings to intricate field. Furthermore, AI facilitates the combination of satellite observations, unmanned aerial vehicle (UAV) imaging, soil and climate data, and spatiotemporal information to enhance the precision of trait monitoring and yield prediction. These advances enhance the ability to evaluate and predict crop performance under variable environmental conditions. This paper offers a cross-disciplinary paradigm for accurate and sustainable modern agriculture by merging AI methodologies with plant phenotyping and yield forecasting.</p>
</abstract>
<kwd-group>
<kwd>AI</kwd>
<kwd>image analysis</kwd>
<kwd>phenotyping</kwd>
<kwd>precision agriculture</kwd>
<kwd>remote sensing</kwd>
<kwd>yield prediction</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="6"/>
<table-count count="3"/>
<equation-count count="0"/>
<ref-count count="162"/>
<page-count count="23"/>
<word-count count="12653"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Technical Advances in Plant Science</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Global food systems are increasingly challenged by rapid population growth, climate variability, and limited natural resources, creating an urgent need for innovative strategies to enhance crop productivity and ensure sustainable agricultural production (<xref ref-type="bibr" rid="B138">van Dijk et&#xa0;al., 2021</xref>). Conventional methods for phenotyping and yield forecasting are laborious and unable to capture complex interactions.</p>
<p>Artificial intelligence (AI) is now widely applied across scientific disciplines (<xref ref-type="bibr" rid="B60">Jumper et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B109">Pyzer-Knapp et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B112">Rajpurkar et&#xa0;al., 2022</xref>; <xref ref-type="bibr" rid="B7">Advancing geoscience with AI, 2024</xref>; <xref ref-type="bibr" rid="B92">Murphy et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B121">Serrano et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B14">Bodnar et&#xa0;al., 2025</xref>). In recent years, the utilization of machine learning (ML) and deep learning (DL) in agriculture has expanded swiftly, propelled by substantial methodological advancements in algorithm creation and the rising accessibility of extensive crop datasets (<xref ref-type="bibr" rid="B59">Joshi et&#xa0;al., 2023</xref>; <xref ref-type="bibr" rid="B148">Waqas et&#xa0;al., 2025</xref>). These approaches remedy long-standing weaknesses in lab and field by taking over tasks that are tedious or require expert handling (<xref ref-type="bibr" rid="B34">Fahlgren et&#xa0;al., 2015</xref>). Plant scientists are beginning to use more computer-driven procedures to automatically quantify image-based traits and predict crop yield based on data collected across the different locations. Since global farming is impacted by increasing food demands and shifting weather patterns, this change occurs at a pivotal time in world agriculture. Secure and high productivity of crops amid the environmental changes will demand agile and high yielding agricultural systems, which, are very difficult to achieve using traditional phenotyping and breeding strategies. The use of computational techniques minimizes time intervals between crops and allows for improved accuracy and lower down the repeat measurement of trait assessments (<xref ref-type="bibr" rid="B140">van Klompenburg et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B24">Crossa et&#xa0;al., 2025a</xref>).</p>
<p>In this review we present a two-part summary on how AI has started to revolutionize plant science in: (i) Image-Based Phenotyping, where algorithms analyze plant images at varying scales from single organs to field canopies. Another application includes (ii) Yield and Trait Prediction through remote sensing, where a combination of varied environment factors like rainfall and temperature data is used to estimate crop yield. Increased adoption and usage of plant imaging provide a switch across scientific disciplines. With the ability of applying reliable and phenotypic information and integrating smart modeling tools, plant biologists can facilitate and improve development of crops; with the integration of high-throughput phenotypic data and advanced predictive algorithms with complicated environment data, the crop improvement cycle can be both fast tracked and enhanced in precision and efficacy (<xref ref-type="fig" rid="f1"><bold>Figure&#xa0;1</bold></xref>).</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>Conceptual framework of AI-enabled plant phenotyping and yield forecasting. The core structure and conceptual framework of the study. (1) Image-based phenotyping (2) Multi-Source remote sensing.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g001.tif">
<alt-text content-type="machine-generated">Illustration depicting a framework for agricultural research using AI. A brain with AI elements connects lab research and field applications, involving machine learning and deep learning. On the left, a corn plant under analysis with a computer and drone. On the right, a farmer holds a tablet, symbolizing remote sensing for yield prediction. The lower section highlights applications such as precision management, early warning, breeding, and policy for food security. Concepts like robustness, interpretability, open science, and edge-cloud co-design are emphasized.</alt-text>
</graphic></fig>
<p>The core structure and conceptual framework of the study (<xref ref-type="bibr" rid="B138">van Dijk et&#xa0;al., 2021</xref>). Image-based phenotyping (<xref ref-type="bibr" rid="B112">Rajpurkar et&#xa0;al., 2022</xref>) Multi-Source remote sensing.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Unified AI framework for plant phenotyping and trait analysis</title>
<p>Modern AI has moved from engineered features to end-to-end DL with automatic representation learning (<xref ref-type="fig" rid="f2"><bold>Figure&#xa0;2</bold></xref>).</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>Framework of AI modeling in plant phenotyping and remote sensing. Traditional ML methods, such as SVM, RF, and GBLUP, for early phenotypic prediction tasks. With the advent of DL, advanced models including CNN, RNN, Transformer, GNN, multimodal learning, and FL for efficient and scalable data analysis for phenotyping and sensing.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g002.tif">
<alt-text content-type="machine-generated">Flowchart illustrating the progression from traditional machine learning (using SVM, RF, GBLUP) to deep learning techniques (CNN, RNN, Transformers, GNN). It shows applications in image-based phenotyping and remote sensing for yield and trait prediction.</alt-text>
</graphic></fig>
<sec id="s2_1">
<label>2.1</label>
<title>Traditional machine learning</title>
<p>Random forest (RF) is suitable for nonlinear, high-dimensional data and used well in trait prediction and remote sensing (<xref ref-type="bibr" rid="B12">Belgiu and Dr&#x103;gu&#x163;, 2016</xref>), Support Vector Machine (SVM) improve generalization by maximizing class margins (<xref ref-type="bibr" rid="B22">Cortes and Vapnik, 1995</xref>), and bagging strengthens the robustness of RF (<xref ref-type="bibr" rid="B16">Breiman, 2001</xref>). Genomic Best Linear Unbiased Prediction (GBLUP) is a linear mixed model in genomic selection (<xref ref-type="bibr" rid="B21">Clark and van der Werf, 2013</xref>).</p>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>Deep feature learning</title>
<sec id="s2_2_1">
<label>2.2.1</label>
<title>Convolutional neural networks</title>
<p>CNNs, designed to capture 2D structural variation, perform well in handwritten character recognition. Graph transformer networks (GTNs) enable joint training of multiple modules and enhance performance. Early work such as LeNet illustrated the effectiveness of CNNs in learning hierarchical features from raw images (<xref ref-type="bibr" rid="B71">Lecun et&#xa0;al., 1998</xref>). Building on this foundation, AlexNet marked a major advance in large-scale image classification by leveraging graphics processing units (GPUs) and the rectified linear unit (ReLU) activation function (<xref ref-type="bibr" rid="B67">Krizhevsky et&#xa0;al., 2017</xref>). Subsequent CNN architectures further pushed performance: VGG deepened networks using small convolutional filters (<xref ref-type="bibr" rid="B126">Simonyan and Zisserman, 2015</xref>), and ResNet introduced residual connections that enabled very deep models (<xref ref-type="bibr" rid="B50">He et&#xa0;al., 2016</xref>). U-Net and related encoder-decoder models are effective for dense prediction tasks by combining multi-scale features (<xref ref-type="bibr" rid="B116">Ronneberger et&#xa0;al., 2015</xref>).</p>
</sec>
<sec id="s2_2_2">
<label>2.2.2</label>
<title>Recurrent neural networks and sequence modeling</title>
<p>Recurrent Neural Networks (RNNs) incorporate feedback connections that create an internal memory of sequential data. Basic RNNs suffer from vanishing or exploding gradients when learning very long-term dependencies. Two gating architectures were introduced to mitigate this: the Long Short-Term Memory (LSTM) (<xref ref-type="bibr" rid="B51">Hochreiter and Schmidhuber, 1997</xref>). and the Gated Recurrent Unit (GRU) (<xref ref-type="bibr" rid="B20">Cho et&#xa0;al., 2014</xref>). Another pivotal innovation was the Attention mechanism, first applied to neural machine translation by Jointly Learning to Align and Translate (<xref ref-type="bibr" rid="B8">Bahdanau et&#xa0;al., 2014</xref>).</p>
</sec>
<sec id="s2_2_3">
<label>2.2.3</label>
<title>Transformers and self-attention</title>
<p>The limitations of RNNs on long sequences led to the emergence of Transformers, which rely on self-attention to process entire sequences in parallel by relating each element to every other (<xref ref-type="bibr" rid="B142">Vaswani et&#xa0;al., 2017</xref>). BERT use bidirectional encoders to generate rich representations for downstream tasks (<xref ref-type="bibr" rid="B31">Devlin et&#xa0;al., 2019</xref>). In computer vision, Vision Transformers (ViT) demonstrated that one can treat an image as a sequence of patches and achieve competitive results using Transformer architectures (<xref ref-type="bibr" rid="B33">Dosovitskiy et&#xa0;al., 2021</xref>). Hierarchical variants such as the Video Shifted Window (Swin) Transformer introduce localized attention windows and multiscale processing to improve efficiency on high-resolution images (<xref ref-type="bibr" rid="B78">Liu et&#xa0;al., 2021</xref>).</p>
</sec>
<sec id="s2_2_4">
<label>2.2.4</label>
<title>Graph neural networks</title>
<p>Graph Neural Networks (GNNs) extend deep learning to graph-structured data. One influential GNN variant is the Graph Convolutional Network (GCN) (<xref ref-type="bibr" rid="B65">Kipf and Welling, 2017</xref>), which generalizes the notion of convolution to graph neighborhoods. More broadly, Message Passing Neural Networks (MPNNs) provide a framework where messages sent along edges can be learned functions of both neighbor and edge features (<xref ref-type="bibr" rid="B42">Gilmer et&#xa0;al., 2017</xref>). The Graph Attention Network (GAT) introduces attention weights on edges (<xref ref-type="bibr" rid="B143">Velickovic et&#xa0;al., [[NoYear]]</xref>).</p>
</sec>
<sec id="s2_2_5">
<label>2.2.5</label>
<title>Multimodal learning</title>
<p>Baltrusaitis et&#xa0;al. provides a comprehensive survey of multimodal Learning, tracing the progression from such independent processing to modern approaches that learn joint representations (<xref ref-type="bibr" rid="B9">Baltrusaitis et&#xa0;al., 2019</xref>). Multimodal deep learning integrates heterogeneous data types within a single framework and uses fusion and alignment strategies to learn shared representations, improving performance across diverse tasks (<xref ref-type="bibr" rid="B122">Shaban and Yousefi, 2024</xref>). Vision-and-language models like ViLBERT (<xref ref-type="bibr" rid="B80">Lu et&#xa0;al., 2019</xref>) and LXMERT (<xref ref-type="bibr" rid="B131">Tan and Bansal, 2019</xref>) are examples of vision-and-language models that have co-attentional transformer layers. These layers facilitate interaction and mutual influence between the picture areas and the language tokens. CLIP model exemplifies this methodology by training on image-text pairs to provide analogous embeddings for corresponding pairs and unique embeddings for non-corresponding pairs, allowing zero-shot transfer by enabling the model to learn a universal correspondence between images and natural language descriptions (<xref ref-type="bibr" rid="B111">Radford et&#xa0;al., 2021</xref>).</p>
</sec>
<sec id="s2_2_6">
<label>2.2.6</label>
<title>Federated learning</title>
<p>Modern AI systems are being used in situations such as user smartphones or different universities and institutions, and can&#x2019;t be combined directly because of privacy concerns. FL solves this problem by letting models be trained together without putting all the raw data in one place (<xref ref-type="bibr" rid="B86">McMahan et&#xa0;al., 2017</xref>). Federated Averaging (FedAvg) averages the model parameters from each client, making communication in federated networks more efficient (<xref ref-type="bibr" rid="B61">Kairouz et&#xa0;al., 2021</xref>).</p>
<p>Traditional ML methods, such as SVM, RF, and GBLUP, for early phenotypic prediction tasks. With the advent of DL, advanced models including CNN, RNN, Transformer, GNN, multimodal learning, and FL for efficient and scalable data analysis for phenotyping and sensing.</p>
</sec>
</sec>
<sec id="s2_3">
<label>2.3</label>
<title>Methodological foundation of AI in plant phenotyping and remote sensing</title>
<p>There are trade-offs between convergence and accuracy (<xref ref-type="table" rid="T1"><bold>Table&#xa0;1</bold></xref>; <xref ref-type="fig" rid="f2"><bold>Figure&#xa0;2</bold></xref>). This trend show that model design is moving in a more general, flexible, and real-world-aligned direction.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Comparative summary of AI methods for plant phenotyping applications.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="center">Method category</th>
<th valign="middle" align="center">Representative model</th>
<th valign="middle" align="center">Core principle</th>
<th valign="middle" align="left">Advantages</th>
<th valign="middle" align="center">Limitations</th>
<th valign="middle" align="center">Representative literature</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">Machine Learning</td>
<td valign="middle" align="left">SVM, RF, GBLUP</td>
<td valign="middle" align="left">Tree-Based Ensemble Models and Linear Mixed Models</td>
<td valign="middle" align="left">Good interpretability and strong robustness</td>
<td valign="middle" align="left">Difficult to capture complex nonlinearity</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B22">Cortes and Vapnik, 1995</xref>; <xref ref-type="bibr" rid="B16">Breiman, 2001</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">CNN</td>
<td valign="middle" align="left">Local Feature Extraction</td>
<td valign="middle" align="left">High accuracy and easy deployment</td>
<td valign="middle" align="left">Insufficient global modeling capability</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B67">Krizhevsky, A. et&#xa0;al. 2012</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">RNN</td>
<td valign="middle" align="left">Temporal Dependency Modeling</td>
<td valign="middle" align="left">Suitable for time-series data</td>
<td valign="middle" align="left">High training cost and poor parallelizability</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B20">Cho et&#xa0;al., 2014</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">Transformers and Self-Attention</td>
<td valign="middle" align="left">Global Dependency and Efficient Representation Learning</td>
<td valign="middle" align="left">Strong global modeling capability</td>
<td valign="middle" align="left">High computational cost and memory footprint</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B142">Vaswani et&#xa0;al., 2017</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">GNN</td>
<td valign="middle" align="left">Graph-Structured Relationship Modeling</td>
<td valign="middle" align="left">Suitable for complex relationship modeling</td>
<td valign="middle" align="left">Performance sensitive to graph construction; limited scalability on large, dense graphs</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B65">Kipf and Welling, 2017</xref>; Velickovic et al., 2017</td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">Multimodal Learning</td>
<td valign="middle" align="left">Multi-source Feature Joint/Joint Embedding</td>
<td valign="middle" align="left">Strong cross-modal representation learning and generalization</td>
<td valign="middle" align="left">Modality alignment and missing-modality handling remain challenging</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B98">Ngiam et al., 2011</xref>; <xref ref-type="bibr" rid="B80">Lu et al., 2019</xref> (ViLBERT) <xref ref-type="bibr" rid="B111">Radford et al., 2021</xref> (CLIP) Cem <xref ref-type="bibr" rid="B1">Akkus et al., 2023</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Deep Learning</td>
<td valign="middle" align="left">FL</td>
<td valign="middle" align="left">Distributed Collaborative Learning with Data Locality</td>
<td valign="middle" align="left">Non-IID data, communication overhead, and possible convergence degradation</td>
<td valign="middle" align="left">Slow convergence and Accuracy Loss</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B86">McMahan et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B61">Kairouz et&#xa0;al., 2021</xref></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Image-based phenotyping: from lab research to field applications</title>
<p>Traditional plant phenotyping, which relies on manual measurements of plant height and leaf size, is slow, small in scale, and subjective. Modern image-based approaches enable nondestructive, large-scale trait acquisition through automated analysis. An early breakthrough in image-based plant phenotyping was achieved through automated segmentation of time-series plant images under variable backgrounds (<xref ref-type="bibr" rid="B88">Minervini et&#xa0;al., 2014</xref>). Advances in imaging hardware (high-resolution cameras, multi/hyperspectral sensors, depth sensors) and robotics now allow plants to be imaged from many angles and modalities, from controlled greenhouse to over field. The improvement of computer image analysis has made it much more accurate, allowing for the consistent extraction of morphological and physiological information across a wide range of temporal and spatial scale (<xref ref-type="bibr" rid="B133">Tardieu et&#xa0;al., 2017</xref>).</p>
<p>Modern image-based phenomics allows for the large-scale, non-invasive assessment of essential plant characteristics. By reviewing more than 200 studies on crop diseases, animal health, and aquaculture, researchers identified several common challenges in data and practical implementation. They emphasize the need for flexible, efficient, and scalable AI models to address these issues (<xref ref-type="bibr" rid="B95">Nawaz et&#xa0;al., 2025</xref>). Recent advancements in computer techniques have considerably improved the accuracy, efficiency, and scope of trait analysis. This change helps solve problems that have been around for a long time, such as manual scoring, changes in the environment, and low temporal resolution. More accurate trait characterization makes it easier to combine phenotypic data, which speeds up the identification of new traits and the development of crops (<xref ref-type="bibr" rid="B133">Tardieu et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B57">Jiang and Li, 2020</xref>; <xref ref-type="bibr" rid="B93">Nabwire et&#xa0;al., 2021</xref>).</p>
<p>This section reviews how imaging technologies and AI algorithms for plant phenotyping have progressed, and how they are being deployed from controlled lab settings to real-world field applications (<xref ref-type="fig" rid="f3"><bold>Figure&#xa0;3</bold></xref>, <xref ref-type="table" rid="T2"><bold>Table&#xa0;2</bold></xref>).</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Overview of image-based phenotyping from lab to field. The controlled-environment phenotyping (left, e.g. greenhouse imaging systems with cameras on rails or conveyor belts) to field phenotyping (right, e.g. UAV drone surveys over crop plots).</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g003.tif">
<alt-text content-type="machine-generated">Illustration showing image-based phenotyping from lab to field. On the left, plants are monitored in a greenhouse using cameras connected to a computer. An arrow labeled “Advanced AI models” points to the right, where a drone surveys crops in an open field.</alt-text>
</graphic></fig>
<table-wrap id="T2" position="float">
<label>Table&#xa0;2</label>
<caption>
<p>Representative AI-based image phenotyping applications.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="left">Crop</th>
<th valign="middle" align="left">Region</th>
<th valign="middle" align="left">AI model</th>
<th valign="middle" align="left">Data type</th>
<th valign="middle" align="left">Performance metrics</th>
<th valign="middle" align="left">Key findings</th>
<th valign="middle" align="left">Reference</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">Multiple crops</td>
<td valign="middle" align="left">Lab<break/>(global images)</td>
<td valign="middle" align="left">CNN (AlexNet/GoogLeNet trained on PlantVillage)</td>
<td valign="middle" align="left">RGB images of leaves</td>
<td valign="middle" align="left">99.35% overall accuracy, F1 = 0.9934</td>
<td valign="middle" align="left">99% accuracy in identifying 38 crop-disease classes from images, establishing a new automated plant disease diagnosis.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B89">Mohanty et&#xa0;al., 2016</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Multiple species<break/>(leaves)</td>
<td valign="middle" align="left">Lab (leaf images database)</td>
<td valign="middle" align="left">CNN (LeafNet)</td>
<td valign="middle" align="left">RGB images of isolated leaves</td>
<td valign="middle" align="left" style="">86.3%, 95.8%, and 97.9% Top-1 accuracy on the LeafSnap, Foliage, and Flavia datasets, respectively.</td>
<td valign="middle" align="left">Automated plant identification based on scans and smartphone pictures to learn leaf shape/texture features.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B10">Barr&#xe9; et&#xa0;al., 2017</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Arabidopsis</td>
<td valign="middle" align="left">Lab (controlled environment)</td>
<td valign="middle" align="left">CNN</td>
<td valign="middle" align="left">Ara2012 and Ara2013-CanonRGB images</td>
<td valign="middle" align="left">R2 = 0.42 (improved count accuracy <italic>vs</italic> prior methods)</td>
<td valign="middle" align="left">Enhancing leaf counting accuracy in Arabidopsis seedlings and enabling reliable high-throughput phenotyping of leaf number.</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B136">Ubbens et&#xa0;al., 2018</xref> (<xref ref-type="bibr" rid="B137">Ubbens and Stavness, 2017</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Clover</td>
<td valign="middle" align="left">Greenhouse (controlled environment)</td>
<td valign="middle" align="left">Faster R-CNN + U-Net (Greenotyper pipeline)</td>
<td valign="middle" align="left">Time-lapse RGB images</td>
<td valign="middle" align="left">97.98% bounding-box AP; 0.95pixel accuracy (IoU = 0.84)</td>
<td valign="middle" align="left">Demonstrated a scalable, affordable high-throughput phenotyping system for plant growth traits.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B134">Tausen et&#xa0;al., 2020</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Rice<break/>(paddy weeds)</td>
<td valign="middle" align="left">Hunan, China (rice fields)</td>
<td valign="middle" align="left">MobileNetV2-UNet &amp; FFB-BiSeNetV2 (lightweight CNN semantic segmentation)</td>
<td valign="middle" align="left">UAV low-altitude RGB imagery</td>
<td valign="middle" align="left">PA = 93.09%; weed IoU = 80.28%; mIoU = 64.38%</td>
<td valign="middle" align="left">Real-time weed detection model via UAV in rice paddies, enabling timely weed mapping.</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B70">Lan et&#xa0;al., 2021</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Fruits</td>
<td valign="middle" align="left">Cross-dataset (orchards, greenhouse)</td>
<td valign="middle" align="left">CNN-based detector + CNN-based GAN (CycleGAN + Improved-YOLOv3)</td>
<td valign="middle" align="left">RGB images of fruit on trees (unlabeled target domain)</td>
<td valign="middle" align="left">mAP = 87.5% (apple), 76.9% (tomato) after adaptation</td>
<td valign="middle" align="left">Trained on orange images generalized well to apples and tomatoes, greatly bridging the species gap in fruit detection.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B158">Zhang et&#xa0;al., 2021</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left" style="">Wheat</td>
<td valign="middle" align="left" style="">Outdoor field</td>
<td valign="middle" align="left" style="">DeepLabV3+ (CNN semantic segmentation)</td>
<td valign="middle" align="left" style="">UAV and ground RGB images (various wheat genotypes)</td>
<td valign="middle" align="left" style="">IoU of 0.77 and 0.90 for plants and soil, respectively (robust segmentation across genotypes)</td>
<td valign="middle" align="left" style="">Achieved robust pixel level plant-soil segmentation in field-grown wheat canopies across years and genotypes.</td>
<td valign="middle" align="left" style="">(<xref ref-type="bibr" rid="B156">Zenkl et&#xa0;al., 2021</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left" style="">Various crops</td>
<td valign="middle" align="left" style="">Controlled environments</td>
<td valign="middle" align="left" style="">PSegNet (3D CNN&#x2013;based point cloud segmentation network)</td>
<td valign="middle" align="left" style="">3D point clouds of plants (lidar scans)</td>
<td valign="middle" align="left" style="">mIoU = 89.9% mRec 90.15%, mCov 83.38%, mWCov 89.54%</td>
<td valign="middle" align="left" style="">For simultaneous segmentation of plant point clouds, improving the precision of 3D organ-level trait extraction.</td>
<td valign="middle" align="left" style=""><xref ref-type="bibr" rid="B72">Li et&#xa0;al., 2022</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Multiple crops (diseases)</td>
<td valign="middle" align="left">Field conditions</td>
<td valign="middle" align="left">MobileViT (PMVT &#x2013; lightweight Vision Transformer)</td>
<td valign="middle" align="left">Field RGB images of plant leaves</td>
<td valign="middle" align="left">Accuracy: 93-94% (wheat, rice datasets); 85.4% (coffee)</td>
<td valign="middle" align="left">Mobile-friendly transformer for on-field disease classification, enabling real-time mobile deployment.</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B74">Li et&#xa0;al., 2023</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Fruits</td>
<td valign="middle" align="left">orchards (multi-fruit)</td>
<td valign="middle" align="left">CNN (YOLOv5) + GAN + Mean Teacher + Attention</td>
<td valign="middle" align="left">RGB fruit images (different crops)</td>
<td valign="middle" align="left">Accuracy: 87.2% (tomato), 89.9% (apple); +10.3% and +2.4% improvements <italic>vs</italic> existing methods</td>
<td valign="middle" align="left">Significantly improved cross-fruit detection performance, effective transferacross fruit datasets.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B46">Guo et&#xa0;al., 2023</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left" style="">Wheat</td>
<td valign="middle" align="left" style="">Outdoor field</td>
<td valign="middle" align="left" style="">Semi-supervised CNN Based Encoder&#x2013;Decoder Domain adaptation</td>
<td valign="middle" align="left" style="">RGB images, video frames (wheat heads)</td>
<td valign="middle" align="left" style="">Strong generalization, scoring 80.7% Dice internally and 64.8% on external multi-country data across 18 domains.</td>
<td valign="middle" align="left" style="">Achieved robust segmentation performance across diverse field conditions with very limited labeling effort.</td>
<td valign="middle" align="left" style="">(<xref ref-type="bibr" rid="B40">Ghanbari et&#xa0;al., 2024</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Maize</td>
<td valign="middle" align="left">Field-based HTPP maize plots</td>
<td valign="middle" align="left" style="">CNN segmentation model with bagging ensemble</td>
<td valign="middle" align="left">High-resolution RGB images of maize plots)</td>
<td valign="middle" align="left">mIoU = 0.70 (IoU<sub>0</sub> = 0.67, IoU<sub>1</sub> = 0.72; +7.7% <italic>vs</italic> single Deeplabv3+)</td>
<td valign="middle" align="left">Improve segmentation accuracy under limited labeled data in field-based phenotyping.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B159">Zhang L. et&#xa0;al., 2024</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Black spruce</td>
<td valign="middle" align="left">Field (Qu&#xe9;bec, Canada)</td>
<td valign="middle" align="left">CNN, K-Means</td>
<td valign="middle" align="left">Time-series PhenoCam images (RGB), Greenness Indices</td>
<td valign="middle" align="left">IoU=0.9121 and an F1-score= 0.9235, (high accuracy)</td>
<td valign="middle" align="left">PhenoAI automated phenology detection from near-surface cameras and multi-index time-series analysis.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B68">Kumar et&#xa0;al., 2025</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Wheat (spike detection)</td>
<td valign="middle" align="left">GWHD2021 + Xinjiang field dataset (multi-country and site hybrid)</td>
<td valign="middle" align="left">FEWheat-YOLO (CNN-based object detector (YOLOv11n))</td>
<td valign="middle" align="left">UAV &amp; smartphone field RGB images</td>
<td valign="middle" align="left">R&#xb2; = 0.941, MAE = 3.46, RMSE = 6.25</td>
<td valign="middle" align="left">Model simplifications enable efficient, high-throughput wheat spike detection.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B149">Wu et&#xa0;al., 2025</xref>)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The controlled-environment phenotyping (left, e.g. greenhouse imaging systems with cameras on rails or conveyor belts) to field phenotyping (right, e.g. UAV drone surveys over crop plots).</p>
<sec id="s3_1">
<label>3.1</label>
<title>Advances in deep learning for plant image analysis</title>
<sec id="s3_1_1">
<label>3.1.1</label>
<title>Deep learning reshaping plant image analysis</title>
<p>Over the last five years, DL has become the dominant approach for analyzing large plant image datasets, largely replacing earlier hand-crafted image analysis techniques. DL models can learn complex visual patterns of plant architecture, health, and development directly from data, reducing the need for human-designed features or heuristics (<xref ref-type="bibr" rid="B92">Murphy et&#xa0;al., 2024</xref>).</p>
<p>PhenoAI is an open-source Python framework for processing PhenoCam time-series images. It integrates quality control, vegetation segmentation, and phenological metrics extraction. In a black spruce case study in Quebec, it accurately detected key phenological stages, reducing manual work and improving monitoring efficiency (<xref ref-type="bibr" rid="B68">Kumar et&#xa0;al., 2025</xref>).</p>
</sec>
<sec id="s3_1_2">
<label>3.1.2</label>
<title>CNNs: the foundation of modern plant image analysis</title>
<p>CNNs were one of the earliest DL architectures widely applied in plant phenotyping. Using large image datasets, researchers have trained CNNs to identify subtle phenotypic differences and to recognize plant structure (<xref ref-type="bibr" rid="B136">Ubbens et&#xa0;al., 2018</xref>). For example, CNN-based semantic segmentation models have been applied to partition images into meaningful classes, which distinguishing plant pixels from background or even segmenting individual leaves and organs in rosette plants and cereal canopies, thereby enabling precise measurement of traits such as leaf area, shape, and number; CNN models LeafNet have also been used for stress and disease phenotyping; by pairing classification networks with saliency maps, researchers can highlight which regions of a leaf or plant (e.g. edges, color patches) most contributed to a disease or drought stress prediction, aligning these regions with known symptoms and improving biological interpretability (<xref ref-type="bibr" rid="B10">Barr&#xe9; et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B11">Beikmohammadi et&#xa0;al., 2022</xref>).</p>
<p>In one early demonstration of DL&#x2019;s potential, Pound et&#xa0;al. achieved over 97% accuracy in detecting and localizing plant organs from imagines Their deep CNN-based system could automatically find root tips and other root/shoot structures with great accuracy. The program can detect 12 of 14 QTLs that human experts had already found (<xref ref-type="bibr" rid="B108">Pound et&#xa0;al., 2017</xref>). This discovery showed that automated photo processing makes phenotyping easier, showing that phenotyping is moving toward data-driven era.</p>
</sec>
<sec id="s3_1_3">
<label>3.1.3</label>
<title>Vision transformers: broadening the analytical scope</title>
<p>Vision Transformers (ViTs) have developed innovative techniques for processing plant images that exceed conventional models. Transformers show long-range interdependence and global spatial linkages in visuals by using self-attention methods. This feature makes it easier to see the general shape and context of a plant, while CNNs only look at small sections. ViTs are being used more and more to find plant diseases and for precision farming. They are good at working with different datasets and can even take good performance in hybrid models. Current research on transfer learning, model compression, and attention visualization is aiding in the reduction of data and computing requirements, underscoring their applicability for field utilization in crop monitoring and management (<xref ref-type="bibr" rid="B87">Mehdipour et&#xa0;al., 2025</xref>). ViT-based models can do plant phenotyping tasks with high accuracy and efficiency that is competitive. For example, a lightweight ViT model (PMVT) made for mobile devices was able to accurately classify plant diseases in the field (<xref ref-type="bibr" rid="B74">Li et&#xa0;al., 2023</xref>). Another method used convolutional features and transformer-based context integration (ST-CFI) to improve the ability to find diseases in leaves (<xref ref-type="bibr" rid="B154">Yu et&#xa0;al., 2025</xref>). Combined two ViT modules for extracting image features with a temporal transformer to model time-series data, then added seed information to estimate soybean production in Canadian farms. This method reduced the prediction error by more than 40% compared to the baseline and showed that seed traits play a key role, especially in identifying low-yield plots (<xref ref-type="bibr" rid="B13">Bi et&#xa0;al., 2023</xref>). Research indicates that Vision Transformers mitigate human and structural biases in picture analysis. This lets researchers examine at pictures more closely and get better features about plants with complicated traits.</p>
</sec>
<sec id="s3_1_4">
<label>3.1.4</label>
<title>Overcoming data limitations through model fusion</title>
<p>A research study shown that employing a deep CNN segmentation network alongside an ensemble bagging technique significantly enhanced segmentation accuracy on a limited set of labeled crop images (<xref ref-type="bibr" rid="B157">Zhan et&#xa0;al., 2024</xref>). Deep 1D/2D CNNs were used to forecast yields in food-insecure areas based on normalized difference vegetation index (NDVI) and climate data. While the method transferred to Algeria (2002&#x2013;2018), it performed worse than simple ML models and NDVI baselines, highlighting the limitations of small datasets for yield prediction (<xref ref-type="bibr" rid="B117">Sabo et&#xa0;al., 2023</xref>). As methods get better, they make outputs easier to understand, and help us understand phenotypic variance better.</p>
</sec>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Core computer vision tasks: segmentation, detection, and trait extraction</title>
<p>In controlled lab, acquired images are processed using semantic segmentation to identify specific plant organs (e.g., flower, leaf) and disease symptoms. Object detection algorithms further classify traits. UAV-based imaging enables large-scale trait extraction, and 3D reconstruction provides structural and morphological measurements.</p>
<p>For quantitative trait analysis, it based on basic computer vision tasks like segmentation, regression, and classification (<xref ref-type="bibr" rid="B91">Mostafa et&#xa0;al., 2023</xref>). Key vision tasks in plant phenotyping include semantic segmentation, object detection, and trait extraction through 3D reconstruction (<xref ref-type="fig" rid="f4"><bold>Figure&#xa0;4</bold></xref>). AI has sped up progress in these areas.</p>
<fig id="f4" position="float">
<label>Figure&#xa0;4</label>
<caption>
<p>Controlled-environment Lab phenotyping. In controlled lab, acquired images are processed using semantic segmentation to identify specific plant organs (e.g., flower, leaf) and disease symptoms. Object detection algorithms further classify traits. UAV-based imaging enables large-scale trait extraction, and 3D reconstruction provides structural and morphological measurements.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g004.tif">
<alt-text content-type="machine-generated">Diagram illustrating LAB Research in plant analysis. A greenhouse setup shows plants monitored by a camera. Processes include semantic segmentation dividing plant parts, object detection identifying health and disease traits, and UAV-based trait extraction with 3D reconstruction of plants.</alt-text>
</graphic></fig>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Semantic segmentation</title>
<p>Semantic Segmentation categorizes each pixel individually rather than considering the image as a whole (<xref ref-type="bibr" rid="B105">Osco et&#xa0;al., 2021b</xref>). Detailed pixel-level clarity lets you see the structure of plants in great detail, making it possible to precisely measure things like leaf area, canopy coverage, and organ morphology. Zenkl et&#xa0;al. employed a DeepLabV3+ model, trained on a variety of wheat photos, to effectively distinguish plants from the background across different genotypes (<xref ref-type="bibr" rid="B156">Zenkl et&#xa0;al., 2021</xref>). A UAV-enabled framework for cotton yield prediction based on cotton semantic segmentation using YOLO and SAM models (<xref ref-type="bibr" rid="B113">Reddy et&#xa0;al., 2024</xref>). Li et&#xa0;al. developed PSegNet, a framework for instance segmentation of 3D plant point clouds, enabling accurate organ identification combines voxel-based downsampling with multi-scale segmentation to provide better results across different plant species (<xref ref-type="bibr" rid="B72">Li et&#xa0;al., 2022</xref>). These instances demonstrate a shift towards sophisticated segmentation techniques that bring precision with AI.</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Object detection</title>
<p>Modern object recognition like Faster R-CNN and YOLO (<xref ref-type="bibr" rid="B114">Ren et&#xa0;al., 2017</xref>; <xref ref-type="bibr" rid="B83">Luo et&#xa0;al., 2024</xref>). have made it easier to do things like count fruits, find cereal spikes, and spot leaf defects. One of the biggest problems is generalizing to different species or varieties, since models that work well on one type of fruit may not work well on others that look different. To fix this problem, more and more people are using domain adaptation strategies.</p>
<p>Zhang et&#xa0;al. and Guo et&#xa0;al. studied fruit detection models to &#x201c;fill the species gap&#x201d;. Domain adaptation enhances the transferability of fruit identification models, facilitating cross-species application without the necessity for human labeling. When CycleGAN-based visual translation and pseudo-labeling were utilized, models trained on oranges did very well with apples and tomatoes (<xref ref-type="bibr" rid="B158">Zhang et&#xa0;al., 2021</xref>) Transfer performance has improved a lot because to cross-domain detection frameworks that leverage GANs and features to match images through knowledge distillation. To make sure that instance- and image-level representations are the same across domains, SSDA-YOLO uses YOLOv5, Mean Teacher distillation, and style transfer (<xref ref-type="bibr" rid="B162">Zhou et&#xa0;al., 2023</xref>). Adding contextual aggregation to improve global feature learning made it easier to transition models from datasets of oranges to datasets of tomatoes and apples. This showed that fruit recognition is more generalizable and efficient (<xref ref-type="bibr" rid="B46">Guo et&#xa0;al., 2023</xref>). FEWheat-YOLO (YOLOv11n) is a lightweight and efficient model achieved R&#xb2; = 0.941 for Wheat Spike Detection (<xref ref-type="bibr" rid="B149">Wu et&#xa0;al., 2025</xref>). DomAda-FruitDet is a domain-adaptive model that doesn&#x2019;t use anchors. It fills in the gaps between the foreground and background by using multi-scale prediction and adaptive sampling. This makes auto-labeling more precise and speeds up, thereby makes fruit recognition more general (<xref ref-type="bibr" rid="B160">Zhang W. et&#xa0;al., 2024</xref>).</p>
<p>These tactics, along with self-training, enabled the precise identification of fruits without manual labels. This form of domain adaptation is crucial for applying models to crops and in regions with limited data.</p>
</sec>
<sec id="s3_2_3">
<label>3.2.3</label>
<title>Trait extraction and 3D reconstruction</title>
<p>Researchers that study plants have built a system that uses AI to look at drone pictures of crops from different angles. The technique automatically measures plant height and panicle length and creates realistic 3D crop canopy models. This method delivers images to the best ViT model based on clarity, noise, and blurriness. Segmentation accuracy and computing efficiency are higher than typical CNN approaches. Transformer designs can handle a variety of precision agricultural imaging situations, as shown by this modular design (<xref ref-type="bibr" rid="B43">Gopalan et&#xa0;al., 2025</xref>). To sum up the architectures used in agricultural research, we focus on two types of Transformer-based models: pure and hybrid (<xref ref-type="bibr" rid="B151">Xie et&#xa0;al., 2024</xref>). When put CNN and Transformer parts together, Transformers work from several viewpoints to display the overall shape of the plant, while CNNs gather small details like the margins of leaves and the textures of panicles. In 2023, Hu et&#xa0;al. introduced FOTCA, a hybrid plant disease detection model using adaptive Fourier neural operators and CNN-Transformer fusion. This design improves generalization and recognition by collecting local and global information (<xref ref-type="bibr" rid="B53">Hu et&#xa0;al., 2023</xref>).</p>
<p>In addition to 2D picture analysis, computational techniques are being utilized to rebuild plant morphology in three dimensions and to extract structural characteristics. Multi-view imagery and depth sensing make it possible to make 3D point clouds or volumetric models. Transformers are built into CNN-based stereo networks for plant research. UAV-based RS offers high-resolution monitoring that works well with yield data. DL models (CNN, LSTM, CNN-LSTM, ConvLSTM, 3D-CNN) were used to look at UAV RGB time series and meteorological data in Finland to estimate yield. The 3D-CNN achieved the highest accuracy, with an MAE of 219 kg/ha (5.5% MAPE) over a 15-week sequence and 293 kg/ha (7.2% MAPE) using early-season data (<xref ref-type="bibr" rid="B96">Nevavuori et&#xa0;al., 2020</xref>). Furthermore, EdgeMVSFormer, a transformer-based multi-view stereo method that reconstructs detailed 3D plant models from UAV images (<xref ref-type="bibr" rid="B18">Cheng et&#xa0;al., 2025</xref>). SegVoteNet represents a novel framework for 3D sorghum canopy phenotyping, by integrating UAV-based data acquisition, NeRF-based reconstruction, and DL-driven analysis. By integrating VoteNet and PointNet++ in a unified backbone, the model performs semantic segmentation and panicle detection directly from point clouds. This framework offers a scalable approach for canopy and panicle trait characterization in sorghum (<xref ref-type="bibr" rid="B55">James et&#xa0;al., 2025</xref>). Ninomiya (2022) analyzed high-throughput field phenotyping technologies, concentrating on diverse crop canopy characteristics, including plant height, coverage, biomass, stress indicators, and organ identification and enumeration. Improvements in photography, 3D reconstruction, sensor technologies, UAVs, and computational methods have made these observations possible (<xref ref-type="bibr" rid="B100">Ninomiya, 2022</xref>). Gaillard et&#xa0;al. devised a high-throughput voxel carving technique to recreate 3D models of sorghum plants from several RGB photos, facilitating the assessment of canopy features pertinent to light interception and genetic research (<xref ref-type="bibr" rid="B37">Gaillard et&#xa0;al., 2020</xref>). These 3D reconstructions make it possible to assess properties that are hard to see in 2D photos, such as plant structure, canopy volume, and branching patterns.</p>
<p>In short, DL-based segmentation, detection, and trait extraction dramatically increase the number of traits that can be measured from photos especially for varies situations.</p>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Extending image-based phenotyping to real-world field conditions</title>
<p>One of the problems with image-based phenotyping is that it is hard to transfer approaches from controlled greenhouse to field settings, where changes in lighting, backdrop, weather, and plant structure can make models less reliable (<xref ref-type="fig" rid="f5"><bold>Figure&#xa0;5</bold></xref>). Three primary research trajectories are tackling this deficiency.</p>
<fig id="f5" position="float">
<label>Figure&#xa0;5</label>
<caption>
<p>Image-based phenotyping to real-world field conditions. Domain adaptation allows models to generalize across crops and environments, while edge deployment supports real-time trait analysis. High-throughput field phenotyping allows rapid trait measurement in real-world agricultural environments.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g005.tif">
<alt-text content-type="machine-generated">Illustration showing the application of drones in agriculture. On the left, a drone hovers over crop fields. On the right, three panels depict the process: robust model adaptation for different crops, edge deployment on UAVs and mobile devices by a farmer, and high-throughput field phenotyping with a drone collecting data.</alt-text>
</graphic></fig>
<p>Domain adaptation allows models to generalize across crops and environments, while edge deployment supports real-time trait analysis. High-throughput field phenotyping allows rapid trait measurement in real-world agricultural environments.</p>
<sec id="s3_3_1">
<label>3.3.1</label>
<title>Robust model adaptation</title>
<p>Models that work well in one situation (like greenhouse photos of plants that all develop the same way) frequently don&#x2019;t work well in another (like field plots with changing conditions). Recent research has investigated semi-self-supervised domain adaptation, utilizing little manual annotation alongside video-derived data and neural encoder-decoder architectures to produce substantial training datasets. These methods have shown good results in a variety of fields, showing that they could make DL applications in agriculture more stable and scalable (<xref ref-type="bibr" rid="B40">Ghanbari et&#xa0;al., 2024</xref>). As noted in the last section, researchers are now using domain adaptation and intense data augmentation to make models more robust.</p>
<p>For example, large annotated field image datasets such as the Global Wheat Head Detection (GWHD) dataset (<xref ref-type="bibr" rid="B27">David et&#xa0;al., 2020</xref>) have been established to facilitate training of wheat spike detection models across diverse genotypes and locations. Exposure to such diverse training data helps models learn more generalized representations. Techniques such as adversarial domain adaptation and synthetic data generation are also applied. One study proposed a semi-self-supervised CNN with probabilistic diffusion, requiring minimal manual labels and achieving robust crop image segmentation across variable field conditions (weather, lighting changes) (<xref ref-type="bibr" rid="B40">Ghanbari et&#xa0;al., 2024</xref>). The focus is on careful dataset design and curation. Training that covers a wide range of environmental circumstances makes models more stable when they are used in the real world. Transfer learning through fine-tuning on small datasets is commonly used to quickly adapt models to new crops or sites without having to relabel a lot of data.</p>
<p>Recent developments have investigated the application of Generative Adversarial Networks (GANs) to reduce dependence on extensive annotated datasets. Varela et&#xa0;al. introduced the Efficiently Supervised GAN (ESGAN) framework, which integrates extensive amounts of unlabeled UAV footage with a limited collection of annotated samples to precisely categorize heading phases in Miscanthus breeding populations (<xref ref-type="bibr" rid="B141">Varela et&#xa0;al., 2025</xref>). ESGAN was able to equal the accuracy of fully supervised models using only 1% of the labeled data, which made labeling much less of a chore. Its generator-discriminator design allows it to work well under changing field circumstances, making it a scalable solution for complicated agricultural situations (<xref ref-type="bibr" rid="B29">Depaepe, 2025</xref>).</p>
</sec>
<sec id="s3_3_2">
<label>3.3.2</label>
<title>Edge deployment on mobile devices and UAVs</title>
<p>To reach this goal, models need to be quickly put on edge devices like drones, robots, and mobile phones without using the cloud. Lightweight architectures and model compression methods like pruning, quantization, knowledge distillation, and deep compression have made it possible to run on less powerful hardware by using less memory and computing resources. This method keeps accuracy while cutting the network&#x2019;s storage needs by more than 30 times. Tests on AlexNet and VGG-16 demonstrated that compressed models can fit in memory on the chip, which speeds things up and uses less power. This enables you employ complicated networks on mobile and embedded devices (<xref ref-type="bibr" rid="B49">Han et&#xa0;al., 2016</xref>). MobileNetV3 is a new and efficient CNN design that uses hardware-aware neural architecture search, the NetAdapt algorithm, and other advancements to the architecture. This leads to MobileNetV3-Large for apps that need a lot of resources and MobileNetV3-Small for apps that don&#x2019;t. These models make it faster and more accurate to classify, detect, and segment images semantically. MobileNetV3-Large and the LR-ASPP decoder are superior than MobileNetV2 and have shorter lag time (<xref ref-type="bibr" rid="B52">Howard et&#xa0;al., 2019</xref>).</p>
<p>DL and smartphones work together to make it feasible to find crop illnesses on a big scale. A revolutionary study trained a CNN on 54,306 leaf pictures from 14 species and 26 diseases, achieving 99.35% accuracy and demonstrating the potential for large-scale smartphone-based detection (<xref ref-type="bibr" rid="B89">Mohanty et&#xa0;al., 2016</xref>).</p>
<p>Recent UAV phenotyping research has focused on lightweight hardware suitable for field deployment. Improvements have made it possible to process data in real time, which means that data doesn&#x2019;t have to be moved around as often. A number of studies show that optimized embedded models can be just as accurate as bigger cloud-based systems. This means that they may be able to keep working well in the field.</p>
<p>Measuring features like leaf color, leaf area index (LAI), chlorophyll content, biomass, and yield by hand is time-consuming and not very effective. UAV remote sensing platforms (UAV-RSPs) with different sensors, on the other hand, now offer a flexible, fast, and non-destructive way to do large-scale phenotyping (<xref ref-type="bibr" rid="B153">Yang et&#xa0;al., 2017</xref>). Later, Osco and his team did 232 studies utilizing pictures taken by UAVs. They looked into a lot of different application areas, sensor integration, and ways to classify and regress. Their experiment showed how well-made models can help in monitoring crops and phenotyping in the field (<xref ref-type="bibr" rid="B104">Osco et&#xa0;al., 2021a</xref>). Guo et&#xa0;al. examined recent developments in UAS-based plant phenotyping, emphasizing its cost-effectiveness, versatility, and potential to integrate many sensors in settings (<xref ref-type="bibr" rid="B45">Guo et&#xa0;al., 2021</xref>).</p>
<p>To automatically look at canopy images, CimageA is a software system that can grow and uses machine vision and ML. It helps quickly and accurately determine phenotypes such as LAI, canopy coverage, and plant height (<xref ref-type="bibr" rid="B36">Fu et&#xa0;al., 2024</xref>). To strike a balance between speed and precision, there are two optimized models: MobileNetV2-UNet and FFB-BiSeNetV2. TensorRT was used to optimization to execute them on Jetson AGX Xavier. MobileNetV2-UNet reduced the number of parameters and calculations while speeding up inference. FFB-BiSeNetV2 had a mean Intersection over Union (IoU) of 80.28%. Both models worked in real time (more than 40 frames per second), showing that UAV-based image analysis is a good way to manage weeds and safeguard plants (<xref ref-type="bibr" rid="B70">Lan et&#xa0;al., 2021</xref>).</p>
<p>The lightweight model design, embedded hardware capability, and UAV sensing platforms are making the next generation of scalable solutions for plant phenotyping.</p>
</sec>
<sec id="s3_3_3">
<label>3.3.3</label>
<title>High-throughput field phenotyping</title>
<p>Drones with RGB or multispectral sensors quickly take pictures of field plots and get traits faster in the sectors of precision agriculture and breeding. Future breeding advancements necessitate the development of cost-effective, field-deployed high-throughput phenotyping devices (HTPPs). They need to combine large-scale, non-invasive sensing with automated environmental monitoring and fast ways to collect, score, and analyze data (<xref ref-type="bibr" rid="B4">Araus and Cairns, 2014</xref>).</p>
<p>A detailed study looked at how Unmanned aerial system (UAS)-based high-throughput phenotyping can be used in breeding programs. It covered tools, methods, direct trait measurement, predictive breeding, QTL finding, and the ability to speed up genetic gain (<xref ref-type="bibr" rid="B64">Khuimphukhieo and da Silva, 2025</xref>). Moreover, DL has enhanced both the throughput and accuracy of these UAV phenotyping pipelines. Yang et&#xa0;al. developed a UAV image analysis method that accurately extracts plot boundaries from field trial images, greatly speeding up trait measurements (<xref ref-type="bibr" rid="B152">Yang et&#xa0;al., 2021</xref>). Tang et&#xa0;al. demonstrated automated plot delineation and biomass estimation in alfalfa field trials. Using canopy area and plant height features extracted from images, the model achieved an R&#xb2; of about 0.6 for biomass prediction without manual input. Multitask DL frameworks have also been applied to predict multiple traits from the same images (<xref ref-type="bibr" rid="B132">Tang et&#xa0;al., 2021</xref>). UAV-based multisensory data fusion and GeoAI (a combination of geospatial and artificial intelligence research) was designed for high-throughput maize phenotyping. UAVs equipped with hyperspectral, thermal, and LiDAR sensors were used to predict eight yield- and nitrogen-related traits. Extended NDSI analysis, classical ML models (SVM, RF), and a multitask CNN were compared. Integrating hyperspectral and LiDAR data improved prediction accuracy, and the multitask CNN achieved performance comparable to or better than single-task models, demonstrating the potential of GeoAI and sensor fusion for field-scale trait prediction (<xref ref-type="bibr" rid="B99">Nguyen et&#xa0;al., 2023</xref>).</p>
<p>Tausen et&#xa0;al. developed Greenotyper, a low-cost, distributed HTP platform using 180 Raspberry Pi cameras to monitor 1,800 clover plants and generate over 355,000 images in one experiment. A U-Net based pipeline achieved ~98% plant localization accuracy and 95% segmentation accuracy. All images were processed within one day with 96% system uptime, demonstrating an efficient and scalable alternative to traditional phenotyping platforms (<xref ref-type="bibr" rid="B134">Tausen et&#xa0;al., 2020</xref>). MtCro is a multi-task framework that models associated phenotypes together in a shared parameter space. It did better than DNNGP and SoyDNGP on big datasets of wheat and maize, making predictions up to 9% more accurate and making predictions for several phenotypes better (<xref ref-type="bibr" rid="B17">Chao et&#xa0;al., 2025</xref>).</p>
<p>In the end, the outcome is a phenotyping pipeline that can be scaled up and can fly a drone over a field and automatically extract yield-relevant features from each plot within minutes of data acquisition. This gives plant breeders data-driven insights that are almost real-time for making decisions about which plants to breed. It let crop researchers run tests on thousands of genotypes in different settings, helping farmers on their fields manage with accuracy.</p>
</sec>
</sec>
<sec id="s3_4">
<label>3.4</label>
<title>Open Science platform in plant phenotyping</title>
<p>As image-based phenotyping grows, data needs to be shared and made more open. Open platforms let users add new features or functions, which encourages collaboration in community. Tools like PlantCV (<xref ref-type="bibr" rid="B39">Gehan et&#xa0;al., 2017</xref>) and Image Harvest (<xref ref-type="bibr" rid="B66">Knecht et&#xa0;al., 2016</xref>) offer researchers standardized image analysis methods that are accessible at no cost. Deep Plant Phenomics offers pre-trained neural networks for plant phenotyping and enables researchers to tailor the models to their particular requirements. In Arabidopsis thaliana, it has showed promising mutant classification and age regression with high counting accuracy (<xref ref-type="bibr" rid="B137">Ubbens and Stavness, 2017</xref>). Furthermore, the publication of the Global Wheat Head Detection Dataset (GWHD) was a collaborative initiative aimed at establishing a substantial and varied standard for head detection algorithms (<xref ref-type="bibr" rid="B27">David et&#xa0;al., 2020</xref>).</p>
<p>Alongside software, there is an initiative for enhanced data governance and adherence to Findable, Accessible, Interoperable, and Reusable (FAIR) standards (<xref ref-type="bibr" rid="B106">Papoutsoglou et&#xa0;al., 2023</xref>).</p>
<p>Common benchmark datasets improve algorithm development and allow research organizations to compare results. Researchers are standardizing data to make collaboration easier. Researchers increasingly share code and models, making results easier to replicate and strengthening trust in open science.</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Remote sensing for yield and trait prediction</title>
<p>Beyond image-based phenotyping, recent analytical methods highlight data from several sources to better describe environmental aspects.</p>
<p>Forecasting agricultural yield is a plant area priority. Traditional agronomic methods use experience criteria and provide only basic biological process representations, making it difficult to understand agricultural systems. Many methods fail to capture the complex genotype&#xd7;environment (G&#xd7;E) interactions that impact agricultural outcomes. In recent years, AI has shown a remarkable ability to find patterns in large, heterogeneous datasets, complementing traditional models. Recent progress in HTP has enabled multi-source frameworks that integrate different data for yield and trait prediction. These may include time-series satellite imagery, weather data, soil maps, management records, and even genomic information about crop varieties. The data are diverse in nature (spatial, temporal), and accordingly, flexible AI architectures are employed to fuse them. At the same time, challenges remain in terms of model generalizability across environments and seasons when predictions are used to inform agricultural decisions or policy.</p>
<p>In this section, we will outline how multi-source data are integrated for crop prediction, the algorithmic advances for spatial-temporal modeling, and key applications from farm to global scales (<xref ref-type="fig" rid="f6"><bold>Figure&#xa0;6</bold></xref>; <xref ref-type="table" rid="T3"><bold>Table&#xa0;3</bold></xref>).</p>
<fig id="f6" position="float">
<label>Figure&#xa0;6</label>
<caption>
<p>AI-enabled workflow for crop monitoring and decision support. Multisource environmental, phenotypic, and management data are processed using machine learning and deep learning models. Model fusion and calibration produce spatial yield and trait maps, enabling real-time, data-driven agricultural management.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpls-16-1732979-g006.tif">
<alt-text content-type="machine-generated">Flowchart showing an agricultural decision support system. Data sources include satellite, UAV, weather forecasts, soil, measurements, temporal-spatio dynamics, and management. This data undergoes preprocessing using machine learning (SVM, RF, LASSO, GBLUP) and deep learning (CNN, RNN, GNN) techniques. Fusion modeling, calibration, and validation are performed. Outputs include a yield map, trait map, and stress risk shown on a mobile device, leading to field irrigation decisions by a farmer.</alt-text>
</graphic></fig>
<table-wrap id="T3" position="float">
<label>Table&#xa0;3</label>
<caption>
<p>Representative AI-based models for crop yield and trait forecasting.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="middle" align="left">Crop</th>
<th valign="middle" align="left">Region</th>
<th valign="middle" align="left">AI model</th>
<th valign="middle" align="left">Data type</th>
<th valign="middle" align="left">Performance metrics</th>
<th valign="middle" align="left">Key findings</th>
<th valign="middle" align="left">Reference</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="middle" align="left">Wheat (winter)</td>
<td valign="middle" align="left">South Asia (district scale)</td>
<td valign="middle" align="left">LSTM-based model (with integrated climate &amp; NDVI)</td>
<td valign="middle" align="left">MODIS NDVI time-series + climate variables</td>
<td valign="middle" align="left">R&#xb2; = 0.78&#x2013;0.82(depending on district/year)</td>
<td valign="middle" align="left">Enabling accurate district-level yield predictions in South Asia without field surveys using only satellite NDVI sequences and climate data.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B6">Ashfaq et&#xa0;al., 2025a</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Wheat (winter)</td>
<td valign="middle" align="left">Southern Pakistan (district scale)</td>
<td valign="middle" align="left">DeepAgroNet (multi-branch CNN-RNN-ANN ensemble)</td>
<td valign="middle" align="left">Satellite imagery + weather time-series + soil data</td>
<td valign="middle" align="left">R&#xb2; = 0.77; 98% accuracy one month before harvest; &lt;10% error</td>
<td valign="middle" align="left">Dramatically improving early forecasting for Pakistan&#x2019;s wheat belt.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B5">Ashfaq et&#xa0;al., 2025b</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Wheat (winter)</td>
<td valign="middle" align="left">Henan, China (provincial)</td>
<td valign="middle" align="left">CNN + BiLSTM (Bayesian optimized; &#x201c;BCBL&#x201d; model)</td>
<td valign="middle" align="left">Satellite indices (EVI, LAI, SIF) + climate</td>
<td valign="middle" align="left">R&#xb2; = 0.81; RMSE = 617 kg/ha; MRE = 7.14%</td>
<td valign="middle" align="left">Novel satellite data with traditional indices and weather data for yield prediction. Stable yield estimates 25 days before harvest.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B160">Zhang et&#xa0;al., 2024</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Maize (field scale)</td>
<td valign="middle" align="left">U.S. Midwest (Corn Belt)</td>
<td valign="middle" align="left">DeepCropNet (Attention LSTM + multitask learning)</td>
<td valign="middle" align="left">Historical yield + climate data (1981&#x2013;2016)</td>
<td valign="middle" align="left">RMSE = 0.82 Mg/ha (<italic>vs</italic> 1.14 LASSO and 1.15 RF)</td>
<td valign="middle" align="left">Effectively captured temporal patterns like weather trends, boosting accuracy.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B77">Lin et&#xa0;al., 2020</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Crops (mixed)</td>
<td valign="middle" align="left">Indian</td>
<td valign="middle" align="left">FedGAvg (FL-AGRN (Federated Attention GNN + RNN))</td>
<td valign="middle" align="left">Agricultural Crop Yield in Indian States Dataset (19,524 data points)</td>
<td valign="middle" align="left">R&#xb2; = 0.99; MAE = 1.23; RMSE = 1.45</td>
<td valign="middle" align="left">This demonstrated scalable, confidential yield forecasting across regions.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B101">Nirosha and Vennila, 2025a</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Muitiple Crops</td>
<td valign="middle" align="left">Multi-region (trial plots) farms located in Ames, Iowa</td>
<td valign="middle" align="left">Agri-GNN (GraphSAGE-based neural network)</td>
<td valign="middle" align="left">Vegetation indices + genotype + time + location</td>
<td valign="middle" align="left">RMSE=4.565, MAE = 3.590, R<break/>2 = 0.876</td>
<td valign="middle" align="left">More accurate plot-level yield estimates than KNN methods and underscoring GNNs in breeding.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B47">Gupta and Singh, 2023</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Corn, cotton, soybean, winter wheat</td>
<td valign="middle" align="left">Mississippi, Louisiana, Iowa, Illinois</td>
<td valign="middle" align="left">MMST-ViT (Transformer)</td>
<td valign="middle" align="left">Sentinel-2 imagery, HRRR meteorological data, USDA crop data</td>
<td valign="middle" align="left">Corn RMSE = 10.5, Corr=0.918; Soybean RMSE = 3.9, Corr=0.900;</td>
<td valign="middle" align="left">The model also demonstrates robust behavior by enabling reliable one-year-ahead yield prediction</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B75">Lin et&#xa0;al., 2023</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Soybean</td>
<td valign="middle" align="left">USA</td>
<td valign="middle" align="left">CNN + LSTM (hybrid model)</td>
<td valign="middle" align="left">MODIS NDVI imagery + climate, via Google Earth Engine</td>
<td valign="middle" align="left">End-of-season RMSE = 329.53</td>
<td valign="middle" align="left">Improved both in-season and end-of-season county yield predictions.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B130">Sun et&#xa0;al., 2019</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Mango</td>
<td valign="middle" align="left">Australia (orchard blocks)</td>
<td valign="middle" align="left">Machine Learning (time-series regression)</td>
<td valign="middle" align="left">8-year yield history + satellite VIs + climate</td>
<td valign="middle" align="left">Block-level 2021 (MAE = 2.9 t/ha) and RF model (farm = 3.7%, block (NMAE) = 33.6% for 2021)</td>
<td valign="middle" align="left">Maintaining reliable predictions for orchard yields without fruit counting, greatly cut down on labor.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B135">Torgbor et&#xa0;al., 2023</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Maize</td>
<td valign="middle" align="left">Missouri, USA (farm trial)</td>
<td valign="middle" align="left">Machine Learning (for best SVR/KNN models))</td>
<td valign="middle" align="left">UAV multispectral imagery (V6 &amp; R5) &#x2192; spectral bands + 26 Vis</td>
<td valign="middle" align="left">Fallow(SVR)<break/>V6:R&#xb2; = 0.84, RMSE = 0.69 Mg/ha<break/>R5:R&#xb2; = 0.83, RMSE = 1.05 Mg/ha</td>
<td valign="middle" align="left">Demonstrated accurate yield estimation with minimal training data of Simpler ML models (SVR, k-NN) slightly outperformed deep networks.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B68">Kumar et&#xa0;al., 2025</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Cotton</td>
<td valign="middle" align="left">Texas, USA (field rows)</td>
<td valign="middle" align="left">YOLOv8 + SAM (object detection &amp; segmentation)</td>
<td valign="middle" align="left">UAV RGB images (open-canopy cotton bolls)</td>
<td valign="middle" align="left">R&#xb2; = 0.913, Low MAE</td>
<td valign="middle" align="left">Provides a robust, rapid yield estimation linking image segmentation to cotton yield with high confidence.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B113">Reddy et&#xa0;al., 2024</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Multi-crop (10)</td>
<td valign="middle" align="left">Global (cross-continent)</td>
<td valign="middle" align="left">Machine Learning</td>
<td valign="middle" align="left">Historical yields + climate/soil/management (global dataset)</td>
<td valign="middle" align="left">R&#xb2; &gt; 0.8 prediction accuracy</td>
<td valign="middle" align="left">Machine learning can capture broad G&#xd7;E patterns at scale and improve global yield projections.</td>
<td valign="middle" align="left"><xref ref-type="bibr" rid="B97">Neumann and Furbank, 2021</xref></td>
</tr>
<tr>
<td valign="middle" align="left">Maize (multimodal HTP)</td>
<td valign="middle" align="left">Illinois, USA (experimental corn field)</td>
<td valign="middle" align="left">Multitask CNN (GeoAI sensor fusion framework)</td>
<td valign="middle" align="left">UAV hyperspectral + thermal + LiDAR data</td>
<td valign="middle" align="left">R&#xb2; up to ~0.85 for N traits; ~0.5&#x2013;0.6 for biomass traits; ~0.34 for grain density (multi-task CNN comparable or slightly better than mono-task and RF/SVR).</td>
<td valign="middle" align="left">The sensor fusion improved prediction robustness, highlighting the value of multimodal data for field-scale yield trait estimation.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B99">Nguyen et&#xa0;al., 2023</xref>)</td>
</tr>
<tr>
<td valign="middle" align="left">Wheat (genomic G&#xd7;E)</td>
<td valign="middle" align="left">Multi-country (breeding trials)</td>
<td valign="middle" align="left">DNNGP</td>
<td valign="middle" align="left">Genomic markers + transcriptomics (for maize) + multi-environment phenotypes</td>
<td valign="middle" align="left">+9% accuracy <italic>vs</italic> prior GP methods</td>
<td valign="middle" align="left">Accelerating selection of high-yield lines under varying conditions.</td>
<td valign="middle" align="left">(<xref ref-type="bibr" rid="B144">Wang et&#xa0;al., 2023</xref>)</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Multisource environmental, phenotypic, and management data are processed using machine learning and deep learning models. Model fusion and calibration produce spatial yield and trait maps, enabling real-time, data-driven agricultural management.</p>
<sec id="s4_1">
<label>4.1</label>
<title>Diverse data sources and multi-source inputs</title>
<p>Crop yields are influenced by multiple factors: weather throughout the season (rainfall, temperature extremes, solar radiation), soil properties (nutrient levels, texture, water-holding capacity, PH), and management (planting date, fertilization, irrigation). A typical multi-source yield prediction pipeline might take as input a sequence of satellite or UAV images (capturing the crop&#x2019;s canopy development via vegetation indices), a time-series of weather variables, static soil and topography maps, and categorical variables like crop type or cultivar. AI models extract these disparate inputs that might be hard to specify in explicit models then learns to associate patterns in these various inputs with final yield.</p>
<sec id="s4_1_1">
<label>4.1.1</label>
<title>Machine learning approaches for prediction</title>
<p>Researchers have shown that ensembles of ML models can effectively map and predict yields by learning from past yield statistics and environmental data. For example, Neumann and Furbank (2021) compiled a continent-wide yield database covering ten major crops and extensive environmental and management variables. Using an ensemble of ML methods, they achieved cross-continental yield prediction accuracies exceeding R&#xb2; = 0.8 for some crops, despite the high environmental and management diversity (<xref ref-type="bibr" rid="B97">Newman and Furbank, 2021</xref>). The progress in using remote sensing and ML for detecting and managing Invasive Plant Species (IPS)-High-resolution imagery from UAVs, satellites, and hyperspectral sensors, combined with ML approaches has improved mapping accuracy and ecological assessment of IPS (<xref ref-type="bibr" rid="B155">Zaka and Samat, 2024</xref>). Integrating technologies such as LiDAR and cross-disciplinary modeling can improve monitoring and conservation under climate and urbanization pressures. UAV-derived vegetation indices combined with ML have been used for accurate corn yield prediction with limited training data. Using multispectral data from V6 and R5 stages under different soil treatments, support vector regression (SVR) and k-nearest neighbor (KNN) performed best, with red-edge and chlorophyll indices as key predictors (<xref ref-type="bibr" rid="B69">Kumar et&#xa0;al., 2023</xref>). Such studies indicates when enough historical data are available, ML approaches can capture yield determinants across climate zones and farming systems in ways.</p>
<p>At local scales, researchers use freely available satellite imagery and weather data to estimate yields without relying on additional field surveys. In mango orchards, Torgbor et&#xa0;al. developed a time-series modeling framework for yield forecasting without in-field fruit counting. Using eight years of yield data from 51 orchard blocks, together with remotely sensed vegetation indices and climatic variables, ML models (e.g., RF, SVM) achieved accurate yield predictions while reducing the need for extensive ground sampling (<xref ref-type="bibr" rid="B135">Torgbor et&#xa0;al., 2023</xref>). Genomic prediction models use genome-wide DNA markers (e.g. SNPs) to predict the breeding values of individuals termed genomic estimated breeding values (GEBVs) without extensive field testing. These ML models have been widely adopted in plant breeding and can predict cultivar performance in different environments by accounting for G&#xd7;E interactions (<xref ref-type="bibr" rid="B25">Crossa et&#xa0;al., 2025b</xref>).</p>
<p>A wide range of ML algorithms have been applied to implement models for G&#xd7;E: such as GBLUP, a linear mixed model using the G matrix (<xref ref-type="bibr" rid="B103">Nishio and Satoh, 2015</xref>). GBLUP captures additive genetic effects and is robust for large marker sets. Bayesian genomic models Methods like Bayes A, Bayes B, Bayes C, Bayesian LASSO, etc. potentially improving prediction when some QTL have large effects (<xref ref-type="bibr" rid="B125">Shi et&#xa0;al., 2021</xref>). Reproducing Kernel Hilbert Space (RKHS) methods have been widely utilized in plant breeding for their capacity to model non-linear genetic architectures, including epistatic and G&#xd7;E interactions. As one of the earliest ML approaches adopted in this field, RKHS regression projects marker genotype data into a high-dimensional function space via kernel function (<xref ref-type="bibr" rid="B41">Gianola and van Kaam, 2008</xref>; <xref ref-type="bibr" rid="B28">De Los Campos et&#xa0;al., 2010</xref>). RF had been used for genomic predictions and can handle many predictors by averaging over many tree models. It can model G&#xd7;E by including environment indicators or interactions as features (<xref ref-type="bibr" rid="B128">Stephan et&#xa0;al., 2015</xref>). SVM find optimal separating hyperplanes and can be used with kernels for regression (SVR). They have been tested on genomic datasets to capture non-linear genotype-phenotype relationships for very large datasets for genomic prediction (<xref ref-type="bibr" rid="B90">Montesinos L&#xf3;pez et&#xa0;al., 2022</xref>). All kinds of ML approach are highly scalable, since satellite data and gridded weather are available, making it feasible to extend yield prediction to regions where field surveys are impractical.</p>
</sec>
<sec id="s4_1_2">
<label>4.1.2</label>
<title>Combined deep learning for seasonal yield prediction</title>
<p>Neural networks, such as Multi-Layer Perceptrons (MLPs) and CNNs, can automatically capture complex feature interactions, particularly when handling large-scale datasets. As extensively reviewed, DL methods are capable of integrating diverse data sources, including image-based HTP and genomic markers to enhance G&#xd7;E predictions. Furthermore, emerging architectures like the DNN Genomic Prediction (DNNGP) framework set a precedent for fusing multi-omics data (<xref ref-type="bibr" rid="B144">Wang et&#xa0;al., 2023</xref>). A pioneering example of a multi-source architecture for yield prediction is DeepAgroNet, which employs a multi-branch neural network to handle different data types: one branch is a CNN that processes spatial features from satellite images, another branch is Long Short-Term Memory (LSTM) that captures temporal patterns in sequential weather data, and a third branch is a feed-forward network for static variables. By merging these branches, the model can acquire a unified depiction of the environment and agricultural circumstances for different place and period (<xref ref-type="bibr" rid="B6">Ashfaq et&#xa0;al., 2025b</xref>).</p>
<p>The FL-AGRN model predicts crop yields by integrating attention-based GNNs and RNNs inside a FL architecture (<xref ref-type="bibr" rid="B101">Nirosha and Vennila, 2025a</xref>). A GNN-RNN framework for predicting crop yields across the country clearly considered both geographical and temporal correlations. The model included data from more than 2,000 U.S. counties from 1981 to 2019. It did better than linear, tree-based, and DL baselines (<xref ref-type="bibr" rid="B35">Fan et&#xa0;al., 2022</xref>). The model attained great accuracy (Mean Absolute Error (MAE) = 1.23, Root Mean Square Error (RMSE) = 1.45, R&#xb2; = 0.99) through distributed training with a Federated Gaussian Average aggregator and meticulous data pretreatment, all while ensuring data confidentiality and scalability. It did better than traditional methods and made the yield forecasts more accurate (<xref ref-type="bibr" rid="B102">Nirosha and Vennila, 2025b</xref>). The Agri-GNN also established to simulate crop yield by explicitly addressing both spatial and genotypic interactions among farming plots. Agri-GNN is built on a GraphSAGE architecture and combines vegetation indices, genotype, time, and location data. It works better than ML methods, showing that graph-based neural architectures have a lot of potential for agriculture (<xref ref-type="bibr" rid="B47">Gupta and Singh, 2023</xref>).</p>
<p>Temporal dynamics are especially important in yield formation: crops progress through growth stages, and mid-season indicators (e.g. a severe drought during the flowering stage, or a period of low NDVI for stress) strongly influence final yield. AI models enabling early detection. Recurrent neural networks like LSTM and GRU are natural choices to handle time-series input. In winter wheat yield prediction, researchers showed that using multi-month NDVI time series as input to an LSTM improved yield estimates compared with single-image or cumulative index approaches. The sequential data enabled the model to account for early growth patterns and stress (<xref ref-type="bibr" rid="B124">Sharma et&#xa0;al., 2020</xref>; <xref ref-type="bibr" rid="B159">Zhang et&#xa0;al., 2024</xref>). Moreover, LSTM-based models can integrate information across the entire season (from emergence to grain filling) A recent study combined meta-transformer for multi-source feature fusion with temporal graph neural networks for sequential modeling, achieving nearly 97% accuracy on the EPFL dataset and outperforming LSTM, CNN, and standard Transformer baselines (<xref ref-type="bibr" rid="B119">Sarkar et&#xa0;al., 2024</xref>). Furthermore, convolutional temporal networks and even 1D CNNs have been used to extract features from weather time-series for yield prediction, summer crops using Landsat EVI time series, which can outperform RNN, RF, and SVM in accuracy and F1 score (<xref ref-type="bibr" rid="B118">Saha et&#xa0;al., 2025</xref>).</p>
<p>Transformer-based models, which excel at long-sequence tasks have been applied to multi-year yield forecasting. Transformer&#x2019;s attention mechanism can in principle learn to weight the most critical periods (e.g., weeks of drought stress) more heavily. In one study, a transformer model outperformed LSTMs for early-season county-level yield prediction in the U.S., likely due to its ability to flexibly incorporate historical information and early-season imagery with varying importance; Multi-Modal Spatial-Temporal Vision Transformer (MMST-ViT), which integrates satellite imagery with meteorological data through multi-modal, spatial, and temporal transformer modules. This framework accounts for both short-term weather variability and long-term climate influences. Experiments conducted in more than 200 U.S. counties showed that MMST-ViT achieved better performance comparable methods on multiple evaluation metrics (<xref ref-type="bibr" rid="B75">Lin et&#xa0;al., 2023</xref>). Najjar et&#xa0;al. submitted a paper to ICLR 2025 that was later withdrawn. It looked at how easy it is to explain Multimodal learning frameworks for predicting crop yields. The research utilized Transformer-based architectures to amalgamate many modalities, including satellite imagery and environmental data, while implementing self-attention and feature attribution techniques to improve interpretability and perhaps yield more robust predictions (<xref ref-type="bibr" rid="B94">Najjar et&#xa0;al., 2025</xref>).</p>
</sec>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Modeling temporal-spatio dynamics</title>
<p>Combining both spatial and temporal dimensions is a key problem in crop modeling. Topography, soil heterogeneity, microclimates, the timing of stresses (drought, heat, pest outbreaks), and the path of crop growth (e.g., rapid early growth vs. slow start) can all affect how crops grow and how much they yield.</p>
<p>In traffic field previous studies of TedTrajRec, is proposed to improve recovery performance. PD-GNN is employed to model periodic and topologically aware traffic patterns before, while TedFormer, a time-aware Transformer with neural ordinary differential equations, effectively learns temporal dependencies from irregularly sampled data (<xref ref-type="bibr" rid="B129">Sun et&#xa0;al., 2025</xref>). In AI-Plant field, AI-based approaches address spatial heterogeneity by incorporating remote sensing at multiple resolutions, such as combining coarse satellite data for regional climate context with high-resolution drone images for fine-grained within-field features. AI can extract spatial features (textures, patterns) that correlate with crop health, while graphical AI applied to geospatial grids can capture interactions between neighboring areas (e.g. disease spread or irrigation patterns).</p>
<sec id="s4_2_1">
<label>4.2.1</label>
<title>Spatio-temporal dynamics</title>
<p>Early spatio-temporal rice mapping mainly used phenology- or feature-based analysis of optical and SAR time-series data with thresholding and rule-based classification. A phenology-based algorithm (ILMP) combining Landsat and Moderate-Resolution Imaging Spectroradiometer (MODIS) achieved over 93% accuracy in Nanchang (2015), performing better than NLCD in fragmented cropland (<xref ref-type="bibr" rid="B32">Ding et&#xa0;al., 2020</xref>). The PKI method using Sentinel-1 SAR time series achieved 97.99% accuracy in paddy rice mapping, outperforming phenology-based methods and enabling large-scale application (<xref ref-type="bibr" rid="B76">Lin et&#xa0;al., 2024</xref>).</p>
<p>Early rule-based methods laid the basis for DL approaches that capture complex spatiotemporal signals. Crop mapping has since shifted from phenology analysis to data-driven feature extraction. Lightweight CNN combined with parcel-based image analysis was developed for crop classification using Sentinel-2 time-series imagery. Applied to two regions in T&#xfc;rkiye, the model achieved overall accuracies of 89.3% and 88.3%, outperforming VGG-16, ResNet-50, and U-Net, and demonstrating its cost-efficient crop mapping (<xref ref-type="bibr" rid="B2">Altun and Turker, 2025</xref>). Another approach for spatio-temporal modeling is the use of encoder-decoder frameworks that take sequences of input images (or weather maps) and predict sequences of outputs (like maps of predicted yield). CNN capturing spatial features and LSTM modeling phenological dynamics. A recent soybean study proposed a CNN-LSTM model for county-level soybean yield prediction across CONUS, integrating weather and MODIS data via Google Earth Engine. The hybrid model outperformed alone CNN or LSTM approaches for both in-season and end-of-season prediction, showing promise for broader crop applications (<xref ref-type="bibr" rid="B130">Sun et&#xa0;al., 2019</xref>).</p>
<p>An attention-based Geo-CBAM-CNN model was developed for crop classification using Sentinel-2 time-series imagery. By integrating geographic information into an attention module, it effectively mitigated spatial heterogeneity and enhanced spectral-spatial feature extraction. Validated across multiple U.S. regions, it achieved 97.82% overall accuracy across multiple U.S. regions and showed strong spatial adaptability for large-scale applications (<xref ref-type="bibr" rid="B147">Wang et&#xa0;al., 2021</xref>). To address the challenges of nonlinear spatiotemporal dependencies in yield prediction, a knowledge-guided Spatial-Temporal Attention Graph Network (KSTAGE) was proposed. By integrating spectral features with prior knowledge through temporal attention and spatial graph modeling, KSTAGE achieved significant improvements over baseline models in county-level yield prediction tasks both in China and the U.S (<xref ref-type="bibr" rid="B110">Qiao et&#xa0;al., 2023</xref>). The CNN- Self-Attention based model effectively captures fine spatial and long-term temporal patterns from high-resolution imagery. DL framework SepHRNet combining HRNet and Self-Attention was proposed for crop mapping, achieving 97.5% accuracy and outperforming state-of-the-art models on the Zuericrop dataset (<xref ref-type="bibr" rid="B44">Goyal et&#xa0;al., 2025</xref>). For large-scale corn yield estimation, a deep spatiotemporal framework (DeepCropNet, DCN) was developed that combines Attention-Based LSTM (RNN) for temporal features and multitask learning for spatial features. Applied to U.S. Corn Belt data (1981&#x2013;2016), DCN outperformed LASSO and RF (RMSE = 0.82 vs. 1.14 and 1.05 Mg ha<sup>-</sup>&#xb9;), effectively capturing temporal effects and patterns (<xref ref-type="bibr" rid="B77">Lin et&#xa0;al., 2020</xref>).</p>
<p>Both Cropformer and AgriFM are Transformer-based models, with Cropformer focusing on multi-scenario crop classification and AgriFM extending to large-scale, multi-source spatiotemporal crop mapping. Cropformer, a self-supervised and fine-tuned Transformer-based model, enables accurate and transferable multi-scenario crop classification from time-series remote sensing data (<xref ref-type="bibr" rid="B146">Wang et&#xa0;al., 2025</xref>). AgriFM, a multi-source remote sensing foundation model designed for crop mapping through unified multi-scale spatiotemporal feature extraction. Built on a modified Swin Transformer and pre-trained on over 25 million samples from MODIS, Landsat, and Sentinel-2, it supports diverse tasks such as cropland mapping, boundary delineation, and early-season crop classification. It outperforms remote sensing foundation models (RSFMs), showing strong scalability and adaptability (<xref ref-type="bibr" rid="B73">Li et&#xa0;al., 2025</xref>).</p>
<p>The new trend is Spatial-Temporal (ST) is GNN, by structuring the data as a graph, these approaches can incorporate domain knowledge (e.g. adjacency or shared water sources) and have the model learn how information flows in both space and time. The Spatial-Temporal Synchronous Graph Convolutional Network (STSGCN) was proposed to address complex localized spatiotemporal correlations and heterogeneities in network data forecasting (<xref ref-type="bibr" rid="B127">Song et&#xa0;al., 2020</xref>). A study introduced a hyperspectral maize nitrogen prediction model that combines a dynamic spectral-spatiotemporal attention mechanism with a Graph Neural Network (GNN). The technique attained great accuracy (R&#xb2; = 0.96) and surpassed models like SVM, RF, ResNet, and ViT, exhibiting strong adaptability throughout developmental stages and geographical contexts (<xref ref-type="bibr" rid="B81">Lu et&#xa0;al., 2025</xref>).</p>
<p>These modifications indicate employing advanced AI techniques brings a transition from rule-based phenological mapping to integrated spatiotemporal modeling.</p>
</sec>
<sec id="s4_2_2">
<label>4.2.2</label>
<title>Operational deployment and decision support</title>
<p>Remote sensing and spatially explicit yield prediction can help in making decisions about farming and policy. Satellite-based AI models allow the United Nations (UN) Food and Agriculture Organization (FAO) to monitor drought&#x2019;s effects on crops worldwide in real time. The Agricultural Stress Index System (ASIS) from the FAO combines digital innovation with decades of satellite data, such as long-term AVHRR-derived vegetation health indicators. ASIS helps with anticipatory action, crop insurance, and drought management both on a national and regional scale (<xref ref-type="bibr" rid="B139">Van Hoolst et&#xa0;al., 2016</xref>; <xref ref-type="bibr" rid="B115">Rojas, 2021</xref>).</p>
</sec>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>Representative applications with case studies</title>
<p>From research to practice, data-driven yield and trait prediction has proven effective for crop management. Some of the important uses are:</p>
<sec id="s4_3_1">
<label>4.3.1</label>
<title>Field trial analytics</title>
<p>Field trials at multiple sites generate thousands of experimental crop genotypes. AI models are utilized to guess important qualities like yield and drought tolerance, as well as meteorological conditions and soil data from each of the location. A study on maize used multitask DL to predict multiple traits from UAV imagery simultaneously, improving the selection of promising hybrids GNN have also been employed to leverage genetic relatedness information alongside environment data, enhancing the prediction of genotype performance across trial sites (<xref ref-type="bibr" rid="B161">Zhou et&#xa0;al., 2025</xref>).</p>
</sec>
<sec id="s4_3_2">
<label>4.3.2</label>
<title>Regional yield forecasts</title>
<p>Prior to harvest, governments and commodity markets seek reliable yield forecasts. AI models ingest satellite time-series for crop cover and weather data to forecast yields at district or county levels. Ashfaq et&#xa0;al. demonstrated forecast for wheat in South Asia using an LSTM on NDVI sequences (<xref ref-type="bibr" rid="B5">Ashfaq et&#xa0;al., 2025a</xref>), and similar approaches have been tested for corn and soy in the U.S. Corn Belt (<xref ref-type="bibr" rid="B48">Han et&#xa0;al., 2024</xref>). These forecasts can inform logistics (e.g. storage and transportation needs) and market pricing, therefore predict early interventions.</p>
</sec>
<sec id="s4_3_3">
<label>4.3.3</label>
<title>Stress and resilience trait prediction</title>
<p>AI can predict traits like drought tolerance, disease outbreak, or nutritional deficiencies using remote sensing data, in addition to yield. Models have been taught to forecast drought stress indices for each pixel in a satellite image, which shows where crops are having trouble getting enough water (<xref ref-type="bibr" rid="B30">Desloires et&#xa0;al., 2024</xref>). ML also forecasts when pests and diseases will happen by noticing small changes in the canopy before they happen, which allows for early management actions.</p>
<p>In these applications and case studies, typical problems include making models work in novel situations and making forecasts easier for end users to understand. SHAP (SHapley Additive exPlanations) is one of the methods used to figure out which elements had the biggest effect on a prediction (<xref ref-type="bibr" rid="B82">Lundberg and Lee, 2017</xref>; <xref ref-type="bibr" rid="B23">Crane-Droesch, 2018</xref>). With advancements in hardware and data infrastructure, the obstacles to large-scale deployment of these models are diminishing.</p>
</sec>
</sec>
<sec id="s4_4">
<label>4.4</label>
<title>Field-based AI-enabled yield forecasting</title>
<p>Nowadays, a growing body of evidence demonstrates that AI methods substantially improve yield forecasting across remote-sensing and environmental data streams. Among these approaches, both ML and DL methods have shown clear advantages in capturing complex spatial and temporal patterns for yield forecasting.</p>
<sec id="s4_4_1">
<label>4.4.1</label>
<title>Machine learning based yield forecasting</title>
<p>Accurate yield estimations are increasingly rely on remote sensing, and the growing data volume has made machine learning essential for handling complex, nonlinear information and improving prediction accuracy in modern agricultural systems (<xref ref-type="bibr" rid="B19">Chlingaryan et&#xa0;al., 2018</xref>). A global assessment across wheat, maize, and potato showed that RF provide notably higher yield-prediction accuracy than linear regression, with RF achieving RMSE of 6-14% compared to 14-49% for linear models. Moreover, yield data was from various sources and regions for model training and testing: gridded global wheat grain yield, maize grain yield from US counties over thirty years, and potato tuber/maize silage yield from the northeastern seaboard region for large-scale climate-driven yield forecasting (<xref ref-type="bibr" rid="B56">Jeong et&#xa0;al., 2016</xref>). ML ensemble to forecast maize yields across three U.S. Corn Belt states using both complete and partial in-season weather information. Weighted and average ensemble models achieved high accuracy, and were able to generate reliable early-season forecasts as early as June 1, RRMSE of 9.2% (<xref ref-type="bibr" rid="B123">Shahhosseini et&#xa0;al., 2020</xref>).</p>
</sec>
<sec id="s4_4_2">
<label>4.4.2</label>
<title>Deep learning based yield forecasting</title>
<p>Using multi-temporal MODIS surface reflectance data on satellite across major U.S. soybean-producing states, You et&#xa0;al. developed a deep Gaussian process framework that reduced county-level RMSE by approximately 30% relative to the best traditional remote-sensing models and delivered earlier-season forecasts with steadily improving skill. When aggregated to the national level, the model also achieved about 15% lower MAPE than USDA survey-based estimates in August and September, demonstrating competitive performance well before harvest (<xref ref-type="bibr" rid="B58">Jiaxuan et&#xa0;al., 2017</xref>). A DL framework combining CNN and RNN architectures has been used to forecast corn and soybean yields across the U.S. Corn Belt, integrating environmental and management time-series data. The CNN-RNN model achieved notably low errors (8-9% RMSE of average yield), outperforming RF and LASSO, and demonstrated strong generalization to unseen environments. Its design enables the extraction of temporal environmental signals allowing attribution analyses to quantify the influence of weather, soil conditions, and management practices on yield variation (<xref ref-type="bibr" rid="B63">Khaki and Wang, 2019</xref>). A transformer-based Informer model has been applied to rice yield forecasting across the Indo-Gangetic Plains by integrating time-series satellite data, environmental variables, and 2001&#x2013;2016 yield records. The model outperformed multiple ML and DL baselines (R&#xb2; = 0.81; RMSE = 0.41 t ha<sup>-</sup>&#xb9;) and achieved stable within-season accuracy (R&#xb2; &#x2248; 0.78) as early as two months before maturity. NIRV and late-season growth stages emerged as dominant predictors, reinforcing the model&#x2019;s predictive strength and interpretability (<xref ref-type="bibr" rid="B79">Liu et&#xa0;al., 2022</xref>).</p>
<p>Collectively, these case studies provide convergent and quantitative evidence that AI enables more accurate, earlier, and more transferable yield forecasting than traditional statistical or process-based models across satellite, UAV, and environmental data sources.</p>
</sec>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Challenges of deploying AI in real agricultural environments</title>
<p>Even though there have been big improvements in finding traits, detecting the environment, and predicting yields, using AI in modern farming systems is still make challenge in real conditions.</p>
<sec id="s5_1">
<label>5.1</label>
<title>Data limitations and domain shift</title>
<sec id="s5_1_1">
<label>5.1.1</label>
<title>Data limitations</title>
<p>Agricultural production systems exhibit substantial spatial and temporal variability, and this complexity poses major challenges for developing reliable AI models. plant traits change across environmental conditions, reflecting the strong phenotypic plasticity of plants. Variations in day length, water supply, nutrient availability, or light intensity can lead to substantial differences in plant architecture, making the quantification of structural and developmental variation a central task in phenomics research (<xref ref-type="bibr" rid="B107">Poorter et&#xa0;al., 2019</xref>). Crop yield is influenced by interacting factors such as climate, soil conditions, fertilizer inputs, and varietal differences (<xref ref-type="bibr" rid="B3">Ansarifar et&#xa0;al., 2021</xref>). Data availability and sharing remain uneven across crops, regions, and production scales (<xref ref-type="bibr" rid="B150">Wu et&#xa0;al., 2023</xref>). HTP platforms provide valuable sensor-based phenotypic data, but most available datasets cover few sites and seasons and lack consistent formats. These limitations reduce comparability and broader use (<xref ref-type="bibr" rid="B26">Danilevicz et&#xa0;al., 2021</xref>).</p>
</sec>
<sec id="s5_1_2">
<label>5.1.2</label>
<title>Domain shift</title>
<p>These data limitations exacerbate domain shift between the data-rich environments used for model training and the data-poor systems where predictions are often needed. Accurate yield estimation for major U.S. crops such as corn and soybean has become increasingly important under growing climate variability and production uncertainty. Yet these crops are grown across highly diverse agroecological zones that differ in climate patterns, soil properties and management practices. Such regional heterogeneity introduces domain, meaning that models trained in one production environment may not generalize well to others (<xref ref-type="bibr" rid="B85">Ma et&#xa0;al., 2021</xref>; <xref ref-type="bibr" rid="B84">Ma et&#xa0;al., 2023</xref>). Smallholder systems in South Asia exhibit severe spatial heterogeneity (fields are small, fragmented and managed with diverse practices), making their distribution fundamentally different from that of the large, uniform fields typically used to train remote-sensing models. Studies in the region show reduced yield-mapping accuracy under these mismatches, illustrating a clear form of spatial domain shift (<xref ref-type="bibr" rid="B54">Jain et&#xa0;al., 2016</xref>).</p>
</sec>
</sec>
<sec id="s5_2">
<label>5.2</label>
<title>Operational, and sensor constraints</title>
<p>Many high-performing deep learning models require computational resources, stable data pipelines, and technical expertise that are often absent in serve farmers. Farmers and agents generally need simplified, low-maintenance decision-support tools rather than complex, opaque systems. If AI become complicated, agronomists and breeders are reluctant to trust or adopt them, and it becomes difficult to diagnose whether errors arise from sensor noise, environmental heterogeneity, or broader domain-shift effects (<xref ref-type="bibr" rid="B38">Gardezi et&#xa0;al., 2024</xref>). Sensor instability further reduces the reliability of model inputs. Imagery from UAV, proximal, and satellite platforms frequently varies in spatial resolution, spectral fidelity, temporal coverage, and calibration quality, while sensor aging and calibration drift can introduce systematic biases into repeated measurements. To mitigate these inconsistencies, combining data from multiple sensing platforms and leveraging sensor-fusion or ensemble approaches can improve robustness by averaging across individual sensor limitations (<xref ref-type="bibr" rid="B62">Karmakar et&#xa0;al., 2024</xref>; <xref ref-type="bibr" rid="B145">Wang et&#xa0;al., 2025</xref>).</p>
<p>Taken together, these operational, sensor, and interpretability constraints help explain why models that perform well in research environments often fail to generalize in real-world production systems.</p>
</sec>
<sec id="s5_3">
<label>5.3</label>
<title>Future directions</title>
<p>Future progress will depend on several key areas (<xref ref-type="bibr" rid="B138">van Dijk et&#xa0;al., 2021</xref>): model interpretability, which will allow predictions and explanatory outputs to align more closely with management needs (<xref ref-type="bibr" rid="B112">Rajpurkar et&#xa0;al., 2022</xref>); the incorporation of methods into multiple crop monitoring systems, enabling more consistent and timely field assessment (<xref ref-type="bibr" rid="B60">Jumper et&#xa0;al., 2021</xref>); the adoption of FAIR data principles, which will support broader multi-site datasets and improve generalizability across diverse field conditions; and (<xref ref-type="bibr" rid="B14">Bodnar et&#xa0;al., 2025</xref>) opportunities for future advancement, including streamlined architectures that can be deployed in resource-limited settings and integrative approaches that enhance operational decision support.</p>
<p>Looking ahead, continued improvements in data quality, model design, and agronomic integration will further enhance the practical use of AI in agriculture. Expanded multi-site datasets will support cross-regional generalization, lightweight architectures will ease deployment constraints, and advances in interpretability will allow models to provide explanations that better match real management requirements.</p>
</sec>
</sec>
<sec id="s6" sec-type="conclusions">
<label>6</label>
<title>Conclusion</title>
<p>The amalgamation of AI with imaging technologies and multi-source data is revolutionizing plant phenotyping from a conventional low-throughput, labor-intensive constraint into an intelligent and data-driven high-throughput framework. Progress in various AI architectures has significantly improved the precision and flexibility of contemporary agriculture.</p>
</sec>
<sec id="s7" sec-type="discussion">
<label>7</label>
<title>Discussion for perspectives</title>
<p>This review combines modeling methods with plant phenotyping and multi-source remote sensing to provide predictions. There are still both challenges and perspectives for future.</p>
<sec id="s7_1">
<label>7.1</label>
<title>Model interpretability</title>
<p>Techniques such as gradient-weighted class activation mapping (Grad-CAM) (<xref ref-type="bibr" rid="B120">Selvaraju et&#xa0;al., 2017</xref>), and SHAP analysis (<xref ref-type="bibr" rid="B30">Desloires et&#xa0;al., 2024</xref>) help to clarify the impact of specific characteristics or regions on model predictions. Although the lack of consistent data, limited computing power, and poor model interpretability continue to be major obstacles to the widespread use of modeling approaches in agriculture. To get over these problems, people are now trying to lightweight analytical tools that are easier to test and use, as well as more flexible and understandable for different types of production systems, which, will meet the needs of real-world agricultural systems.</p>
</sec>
<sec id="s7_2">
<label>7.2</label>
<title>Multiple crop monitoring systems for remote sensing</title>
<p>CMS have been established worldwide to support agricultural management and food security assessment. Representative platforms include FAO&#x2019;s ASIS, the European Commission&#x2019;s Anomaly Hotspots of Agricultural Production (ASAP) and Monitoring Agricultural Resources (MARS), Group on Earth Observations Global Agricultural Monitoring Initiative (GEOGLAM)&#x2019;s Crop Monitor, NASA Harvest&#x2019;s Global Agricultural Monitoring (GLAM), China&#x2019;s CropWatch, USDA VegScape, and several national systems such as FASAL in India and VEGA in Russia. These platforms share a common focus on monitoring crop growth conditions, assessing environmental drivers (<xref ref-type="bibr" rid="B150">Wu et&#xa0;al., 2023</xref>).</p>
<p>Other platforms provide a comprehensive and flexible foundation for agricultural and environmental observation data. Satellite platforms (MODIS, Sentinel-2, Landsat-8/9, PlanetScope, WorldView, GF series) offer broad spatial coverage and time series. UAV-based platforms can deliver centimeter-level spatial resolution and flexible flight scheduling, making fine-scale structures detection. Radar and microwave platforms (Sentinel-1 SAR, RADARSAT, ALOS PALSAR, SMAP) provide all-weather observation that are unaffected by cloud cover or illumination conditions, soil moisture mapping and crop structural analysis.</p>
<p>Most systems rely on time-series profiling, and spatial analysis to evaluate crop status. However, differences in classification standards continue to limit the accuracy of these systems across countries, so collaboration will promote standardized protocols and data sharing, enabling more transparent and comparable crop monitoring practices globally.</p>
</sec>
<sec id="s7_3">
<label>7.3</label>
<title>FAIR data principles</title>
<p>Data sharing and standardization is necessary to accelerate plant phenotyping and sensing. Sensing technologies and methodologies have progressed, but sensed data often differ between platforms and thus may be fragmentary and inconsistent. Shared standards like FAIR, making data comparable so that modeling traits different crop types, regions, and environmental conditions be possible.</p>
<p>Setting up a unified open data platform will facilitate the sharing and permanent preservation of research products. There will be more shared code in programming languages, more shared common parameters for models used in research, and a single metadata framework to express all the datasets of projects accumulated with time in different locations. So we will certainly face fewer technical barriers when working and collaborating with each other in different locations.</p>
<p>For future, internationally standardized agricultural data interoperable can be developed for reproducible research, use the information permanently in order to get the best value from these agricultural data sources through long-term use.</p>
</sec>
<sec id="s7_4">
<label>7.4</label>
<title>Opportunities for future advancement</title>
<p>Future research in plant phenotyping and yield prediction is expected to place increasing emphasis on integration, standardization, and interpretability, ensuring that phenotypic, environmental, and genomic data are findable, accessible, interoperable, and reusable across platforms and institutions, which will facilitate large-scale collaboration, enable more rigorous cross-study comparisons, and reduce barriers for data reuse.</p>
<p>Moreover, AI is entering the era of foundation models. These models are widely used, to truly understand and manage foundation models, close interdisciplinary collaboration is needed, involving not only technical aspects but also in social, ethical, and legal dimensions (<xref ref-type="bibr" rid="B15">Bommasani et&#xa0;al., 2022</xref>). In plant research, it is insufficient to concentrate alone on attaining lightweight model correctness; equal emphasis must be placed on guaranteeing the transparency and interpretability of the analytical process. It is necessary to have both of them. Finally, computer scientists, plant biologists, agronomists, and policymakers will need to work closely together to make this vision a reality. All of these approaches could help make production systems that last longer.</p>
<p>In summary, we propose to build an integrated architecture for global agriculture. Achieving this vision requires not only a unified, lightweight AI model but also broad collaboration. Building on open and new agricultural ecosystems with AI-based phenotyping and prediction will be certainly to revolutionize the process of generating brand new systems within the next several decades.</p>
</sec>
</sec>
</body>
<back>
<sec id="s8" sec-type="author-contributions">
<title>Author contributions</title>
<p>TW: Conceptualization, Investigation, Resources, Supervision, Writing &#x2013; original draft, Writing &#x2013; review &amp; editing. RT:&#xa0;Conceptualization, Writing &#x2013; original draft, Writing &#x2013; review &amp; editing. TX: Resources, Writing &#x2013; review &amp; editing. YL: Writing &#x2013; review &amp; editing. YC: Writing &#x2013; review &amp; editing.</p></sec>
<sec id="s10" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
<p>The author TW declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.</p></sec>
<sec id="s11" sec-type="ai-statement">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p></sec>
<sec id="s12" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p></sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Akkus</surname> <given-names>C.</given-names></name>
<name><surname>Chu</surname> <given-names>L.</given-names></name>
<name><surname>Djakovic</surname> <given-names>V.</given-names></name>
<name><surname>Jauch-Walser</surname> <given-names>S.</given-names></name>
<name><surname>Koch</surname> <given-names>P.</given-names></name>
<name><surname>Loss</surname> <given-names>G.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>Multimodal Deep Learning. arXiv Prepr arXiv230104856</article-title>.
</mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Altun</surname> <given-names>M.</given-names></name>
<name><surname>Turker</surname> <given-names>M.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Integration of convolutional neural networks with parcel-based image analysis for crop type mapping from time-series images</article-title>. <source>Earth Sci. Inf.</source> <volume>18</volume>, <fpage>303</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s12145-025-01819-8</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ansarifar</surname> <given-names>J.</given-names></name>
<name><surname>Wang</surname> <given-names>L.</given-names></name>
<name><surname>Archontoulis</surname> <given-names>S. V.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>An interaction regression model for crop yield prediction</article-title>. <source>Sci. Rep.</source> <volume>11</volume>, <fpage>17754</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-021-97221-7</pub-id>, PMID: <pub-id pub-id-type="pmid">34493778</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Araus</surname> <given-names>J. L.</given-names></name>
<name><surname>Cairns</surname> <given-names>J. E.</given-names></name>
</person-group> (<year>2014</year>). 
<article-title>Field high-throughput phenotyping: the new crop breeding frontier</article-title>. <source>Trends Plant Sci.</source> <volume>19</volume>, <fpage>52</fpage>&#x2013;<lpage>61</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.tplants.2013.09.008</pub-id>, PMID: <pub-id pub-id-type="pmid">24139902</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ashfaq</surname> <given-names>M.</given-names></name>
<name><surname>Khan</surname> <given-names>I.</given-names></name>
<name><surname>Afzal</surname> <given-names>R. F.</given-names></name>
<name><surname>Shah</surname> <given-names>D.</given-names></name>
<name><surname>Ali</surname> <given-names>S.</given-names></name>
<name><surname>Tahir</surname> <given-names>M.</given-names></name>
</person-group> (<year>2025</year>a). 
<article-title>Enhanced wheat yield prediction through integrated climate and satellite data using advanced AI techniques</article-title>. <source>Sci. Rep.</source> <volume>15</volume>, <fpage>18093</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-025-02700-w</pub-id>, PMID: <pub-id pub-id-type="pmid">40413272</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ashfaq</surname> <given-names>M.</given-names></name>
<name><surname>Khan</surname> <given-names>I.</given-names></name>
<name><surname>Shah</surname> <given-names>D.</given-names></name>
<name><surname>Ali</surname> <given-names>S.</given-names></name>
<name><surname>Tahir</surname> <given-names>M.</given-names></name>
</person-group> (<year>2025</year>b). 
<article-title>Predicting wheat yield using deep learning and multi-source environmental data</article-title>. <source>Sci. Rep.</source> <volume>15</volume>, <fpage>26446</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-025-11780-7</pub-id>, PMID: <pub-id pub-id-type="pmid">40691684</pub-id>
</mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal">(<year>2024</year>). 
<article-title>Advancing geoscience with AI</article-title>. <source>Nat. Geosci</source> <volume>17</volume>, <fpage>947</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41561-024-01572-5</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Bahdanau</surname> <given-names>D.</given-names></name>
<name><surname>Cho</surname> <given-names>K.</given-names></name>
<name><surname>Bengio</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2014</year>). &#x201c;
<article-title>Neural machine translation by jointly learning to align and translate</article-title>,&#x201d; in <source>CoRR</source>. abs/1409.0. arXiv preprint arXiv:1409.0473. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1409.0473</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Baltrusaitis</surname> <given-names>T.</given-names></name>
<name><surname>Ahuja</surname> <given-names>C.</given-names></name>
<name><surname>Morency</surname> <given-names>L. P.</given-names></name>
</person-group> (<year>2019</year>). 
<article-title>Multimodal machine learning: A survey and taxonomy</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>41</volume>, <fpage>423</fpage>&#x2013;<lpage>443</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2018.2798607</pub-id>, PMID: <pub-id pub-id-type="pmid">29994351</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Barr&#xe9;</surname> <given-names>P.</given-names></name>
<name><surname>St&#xf6;ver</surname> <given-names>B. C.</given-names></name>
<name><surname>M&#xfc;ller</surname> <given-names>K. F.</given-names></name>
<name><surname>Steinhage</surname> <given-names>V.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>LeafNet: A computer vision system for automatic plant species identification</article-title>. <source>Ecol. Inform</source> <volume>40</volume>, <fpage>50</fpage>&#x2013;<lpage>56</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ecoinf.2017.05.005</pub-id>
</mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Beikmohammadi</surname> <given-names>A.</given-names></name>
<name><surname>Faez</surname> <given-names>K.</given-names></name>
<name><surname>Motallebi</surname> <given-names>A.</given-names></name>
</person-group> (<year>2022</year>). 
<article-title>SWP-LeafNET: A novel multistage approach for plant leaf identification based on deep CNN</article-title>. <source>Expert Syst. Appl.</source> <volume>202</volume>, <elocation-id>117470</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.eswa.2022.117470</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Belgiu</surname> <given-names>M.</given-names></name>
<name><surname>Dr&#x103;gu&#x163;</surname> <given-names>L.</given-names></name>
</person-group> (<year>2016</year>). 
<article-title>Random forest in remote sensing: A review of applications and future directions</article-title>. <source>ISPRS J. Photogramm Remote Sens</source> <volume>114</volume>, <fpage>24</fpage>&#x2013;<lpage>31</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.isprsjprs.2016.01.011</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Bi</surname> <given-names>L.</given-names></name>
<name><surname>Wally</surname> <given-names>O.</given-names></name>
<name><surname>Hu</surname> <given-names>G.</given-names></name>
<name><surname>Tenuta</surname> <given-names>A. U.</given-names></name>
<name><surname>Kandel</surname> <given-names>Y. R.</given-names></name>
<name><surname>Mueller</surname> <given-names>D. S.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>A transformer-based approach for early prediction of soybean yield using time-series images</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1173036</pub-id>, PMID: <pub-id pub-id-type="pmid">37409295</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Bodnar</surname> <given-names>C.</given-names></name>
<name><surname>Bruinsma</surname> <given-names>W. P.</given-names></name>
<name><surname>Lucic</surname> <given-names>A.</given-names></name>
<name><surname>Stanley</surname> <given-names>M.</given-names></name>
<name><surname>Allen</surname> <given-names>A.</given-names></name>
<name><surname>Brandstetter</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>A foundation model for the Earth system</article-title>. <source>Nature.</source> <volume>641</volume>, <fpage>1180</fpage>&#x2013;<lpage>1187</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41586-025-09005-y</pub-id>, PMID: <pub-id pub-id-type="pmid">40399684</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Bommasani</surname> <given-names>R.</given-names></name>
<name><surname>Hudson</surname> <given-names>D. A.</given-names></name>
<name><surname>Adeli</surname> <given-names>E.</given-names></name>
<name><surname>Altman</surname> <given-names>R.</given-names></name>
<name><surname>Arora</surname> <given-names>S.</given-names></name>
<name><surname>von Arx</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2022</year>). <source>On the opportunities and risks of foundation models</source>. <fpage>1</fpage>&#x2013;<lpage>214</lpage>, arXiv preprint arXiv: 2108.07258. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2108.07258</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Breiman</surname> <given-names>L.</given-names></name>
</person-group> (<year>2001</year>). 
<article-title>Random forests</article-title>. <source>Mach. Learn</source> <volume>45</volume>, <fpage>5</fpage>&#x2013;<lpage>32</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1023/A:1010933404324</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Chao</surname> <given-names>D.</given-names></name>
<name><surname>Wang</surname> <given-names>H.</given-names></name>
<name><surname>Wan</surname> <given-names>F.</given-names></name>
<name><surname>Yan</surname> <given-names>S.</given-names></name>
<name><surname>Fang</surname> <given-names>W.</given-names></name>
<name><surname>Yang</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>MtCro: multi-task deep learning framework improves multi-trait genomic prediction of crops</article-title>. <source>Plant Methods</source> <volume>21</volume>, <fpage>12</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-024-01321-0</pub-id>, PMID: <pub-id pub-id-type="pmid">39910577</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Cheng</surname> <given-names>Y.</given-names></name>
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Lan</surname> <given-names>G.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<name><surname>Chen</surname> <given-names>R.</given-names></name>
<name><surname>Huang</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Edge_MVSFormer: edge-aware multi-view stereo plant reconstruction based on transformer networks</article-title>. <source>Sensors</source> <volume>25</volume>, <elocation-id>2177</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s25072177</pub-id>, PMID: <pub-id pub-id-type="pmid">40218688</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Chlingaryan</surname> <given-names>A.</given-names></name>
<name><surname>Sukkarieh</surname> <given-names>S.</given-names></name>
<name><surname>Whelan</surname> <given-names>B.</given-names></name>
</person-group> (<year>2018</year>). 
<article-title>Machine learning approaches for crop yield prediction and nitrogen status estimation in precision agriculture: A review</article-title>. <source>Comput. Electron Agric.</source> <volume>151</volume>, <fpage>61</fpage>&#x2013;<lpage>69</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2018.05.012</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Cho</surname> <given-names>K.</given-names></name>
<name><surname>van Merri&#xeb;nboer</surname> <given-names>B.</given-names></name>
<name><surname>Gulcehre</surname> <given-names>C.</given-names></name>
<name><surname>Bahdanau</surname> <given-names>D.</given-names></name>
<name><surname>Bougares</surname> <given-names>F.</given-names></name>
<name><surname>Schwenk</surname> <given-names>H.</given-names></name>
<etal/>
</person-group>. (<year>2014</year>). &#x201c;
<article-title>Learning phrase representations using {RNN} Encoder{--}Decoder for statistical machine translation</article-title>,&#x201d; in <source>Proceedings of the 2014 conference on empirical methods in natural language processing ({EMNLP})</source>. Eds. 
<person-group person-group-type="editor">
<name><surname>Moschitti</surname> <given-names>A.</given-names></name>
<name><surname>Pang</surname> <given-names>B.</given-names></name>
<name><surname>Daelemans</surname> <given-names>W.</given-names></name>
</person-group> (
<publisher-name>Association for Computational Linguistics</publisher-name>, <publisher-loc>Doha, Qatar</publisher-loc>), <fpage>1724</fpage>&#x2013;<lpage>1734</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3115/v1/D14-1179</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Clark</surname> <given-names>S. A.</given-names></name>
<name><surname>van der Werf</surname> <given-names>J.</given-names></name>
</person-group> (<year>2013</year>). 
<article-title>Genomic best linear unbiased prediction (gBLUP) for the estimation of genomic breeding values</article-title>. <source>Methods Mol. Biol.</source> <volume>1019</volume>, <fpage>321</fpage>&#x2013;<lpage>330</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-1-62703-447-0_13</pub-id>, PMID: <pub-id pub-id-type="pmid">23756897</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Cortes</surname> <given-names>C.</given-names></name>
<name><surname>Vapnik</surname> <given-names>V.</given-names></name>
</person-group> (<year>1995</year>). 
<article-title>Support-vector networks</article-title>. <source>Mach. Learn</source> <volume>20</volume>, <fpage>273</fpage>&#x2013;<lpage>297</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/BF00994018</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Crane-Droesch</surname> <given-names>A.</given-names></name>
</person-group> (<year>2018</year>). 
<article-title>Machine learning methods for crop yield prediction and climate change impact assessment in agriculture</article-title>. <source>Environ. Res. Lett.</source> <volume>13</volume>, <elocation-id>114003</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1088/1748-9326/aae159</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
<name><surname>Martini</surname> <given-names>J. W. R.</given-names></name>
<name><surname>Vitale</surname> <given-names>P.</given-names></name>
<name><surname>P&#xe9;rez-Rodr&#xed;guez</surname> <given-names>P.</given-names></name>
<name><surname>Costa-Neto</surname> <given-names>G.</given-names></name>
<name><surname>Fritsche-Neto</surname> <given-names>R.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>a). 
<article-title>Expanding genomic prediction in plant breeding: harnessing big data, machine learning, and advanced software</article-title>. <source>Trends Plant Sci.</source> <volume>30</volume>, <fpage>756</fpage>&#x2013;<lpage>774</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.tplants.2024.12.009</pub-id>, PMID: <pub-id pub-id-type="pmid">39890501</pub-id>
</mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
<name><surname>Montesinos-Lopez</surname> <given-names>O. A.</given-names></name>
<name><surname>Costa-Neto</surname> <given-names>G.</given-names></name>
<name><surname>Vitale</surname> <given-names>P.</given-names></name>
<name><surname>Martini</surname> <given-names>J. W. R.</given-names></name>
<name><surname>Runcie</surname> <given-names>D.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>b). 
<article-title>Machine learning algorithms translate big data into predictive breeding accuracy</article-title>. <source>Trends Plant Sci.</source> <volume>30</volume>, <fpage>167</fpage>&#x2013;<lpage>184</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.tplants.2024.09.011</pub-id>, PMID: <pub-id pub-id-type="pmid">39462718</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Danilevicz</surname> <given-names>M. F.</given-names></name>
<name><surname>Bayer</surname> <given-names>P. E.</given-names></name>
<name><surname>Nestor</surname> <given-names>B. J.</given-names></name>
<name><surname>Bennamoun</surname> <given-names>M.</given-names></name>
<name><surname>Edwards</surname> <given-names>D.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Resources for image-based high-throughput phenotyping in crops and data sharing challenges</article-title>. <source>Plant Physiol.</source> <volume>187</volume>, <fpage>699</fpage>&#x2013;<lpage>715</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/plphys/kiab301</pub-id>, PMID: <pub-id pub-id-type="pmid">34608963</pub-id>
</mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>David</surname> <given-names>E.</given-names></name>
<name><surname>Madec</surname> <given-names>S.</given-names></name>
<name><surname>Sadeghi-Tehran</surname> <given-names>P.</given-names></name>
<name><surname>Aasen</surname> <given-names>H.</given-names></name>
<name><surname>Zheng</surname> <given-names>B.</given-names></name>
<name><surname>Liu</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2020</year>). 
<article-title>Global wheat head detection (GWHD) dataset: A large and diverse dataset of high-resolution RGB-labelled images to develop and benchmark wheat head detection methods</article-title>. <source>Plant Phenomics</source> <volume>2020</volume>, <elocation-id>3521852</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2020/3521852</pub-id>, PMID: <pub-id pub-id-type="pmid">33313551</pub-id>
</mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>De Los Campos</surname> <given-names>G.</given-names></name>
<name><surname>Gianola</surname> <given-names>D.</given-names></name>
<name><surname>Rosa</surname> <given-names>G. J. M.</given-names></name>
<name><surname>Weigel</surname> <given-names>K. A.</given-names></name>
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
</person-group> (<year>2010</year>). 
<article-title>Semi-parametric genomic-enabled prediction of genetic values using reproducing kernel Hilbert spaces methods</article-title>. <source>Genet. Res. (Camb).</source> <volume>92</volume>, <fpage>295</fpage>&#x2013;<lpage>308</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1017/S0016672310000285</pub-id>, PMID: <pub-id pub-id-type="pmid">20943010</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Depaepe</surname> <given-names>T.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Harder, better, faster, stronger, and with less annotated data: ESGAN and plant sciences</article-title>. <source>Plant Physiol.</source> <volume>198</volume>, <elocation-id>kiaf171</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/plphys/kiaf171</pub-id>, PMID: <pub-id pub-id-type="pmid">40314165</pub-id>
</mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Desloires</surname> <given-names>J.</given-names></name>
<name><surname>Ienco</surname> <given-names>D.</given-names></name>
<name><surname>Botrel</surname> <given-names>A.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Early season forecasting of corn yield at field level from multi-source satellite time series data</article-title>. <source>Remote Sens</source> <volume>16</volume>, <elocation-id>1573</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs16091573</pub-id>
</mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Devlin</surname> <given-names>J.</given-names></name>
<name><surname>Chang</surname> <given-names>M. W.</given-names></name>
<name><surname>Lee</surname> <given-names>K.</given-names></name>
<name><surname>Toutanova</surname> <given-names>K.</given-names></name>
</person-group> (<year>2019</year>). 
<article-title>BERT: Pre-training of deep bidirectional transformers for language understanding</article-title>. <source>NAACL HLT 2019&#x2013;2019 Conf North Am. Chapter Assoc. Comput. Linguist Hum. Lang Technol. - Proc. Conf.</source> <volume>1</volume>, <fpage>4171</fpage>&#x2013;<lpage>4186</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.18653/v1/N19-1423</pub-id>
</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ding</surname> <given-names>M.</given-names></name>
<name><surname>Guan</surname> <given-names>Q.</given-names></name>
<name><surname>Li</surname> <given-names>L.</given-names></name>
<name><surname>Zhang</surname> <given-names>H.</given-names></name>
<name><surname>Liu</surname> <given-names>C.</given-names></name>
<name><surname>Zhang</surname> <given-names>L.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Phenology-based rice paddy mapping using multi-source satellite imagery and a fusion algorithm applied to the poyang lake plain, southern China</article-title>. <source>Remote Sens</source> <volume>12</volume>, <elocation-id>1022</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs12061022</pub-id>
</mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Dosovitskiy</surname> <given-names>A.</given-names></name>
<name><surname>Beyer</surname> <given-names>L.</given-names></name>
<name><surname>Kolesnikov</surname> <given-names>A.</given-names></name>
<name><surname>Weissenborn</surname> <given-names>D.</given-names></name>
<name><surname>Zhai</surname> <given-names>X.</given-names></name>
<name><surname>Unterthiner</surname> <given-names>T.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). &#x201c;
<article-title>an image is worth 16X16 words: transformers for image recognition at scale</article-title>,&#x201d; in <source>ICLR 2021 - 9th int conf learn represent</source>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2010.11929</pub-id>
</mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Fahlgren</surname> <given-names>N.</given-names></name>
<name><surname>Gehan</surname> <given-names>M. A.</given-names></name>
<name><surname>Baxter</surname> <given-names>I.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>Lights, camera, action: high-throughput plant phenotyping is ready for a close-up</article-title>. <source>Curr. Opin. Plant Biol.</source> <volume>24</volume>, <fpage>93</fpage>&#x2013;<lpage>99</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.pbi.2015.02.006</pub-id>, PMID: <pub-id pub-id-type="pmid">25733069</pub-id>
</mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Fan</surname> <given-names>J.</given-names></name>
<name><surname>Bai</surname> <given-names>J.</given-names></name>
<name><surname>Li</surname> <given-names>Z.</given-names></name>
<name><surname>Ortiz-Bobea</surname> <given-names>A.</given-names></name>
<name><surname>Gomes</surname> <given-names>C. P.</given-names></name>
</person-group> (<year>2022</year>). 
<article-title>A GNN-RNN approach for harnessing geospatial and temporal information: application to crop yield prediction</article-title>. <source>Proc. AAAI Conf Artif. Intell.</source> <volume>36</volume>, <fpage>11873</fpage>&#x2013;<lpage>11881</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1609/aaai.v36i11.21444</pub-id>
</mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Fu</surname> <given-names>H.</given-names></name>
<name><surname>Lu</surname> <given-names>J.</given-names></name>
<name><surname>Cui</surname> <given-names>G.</given-names></name>
<name><surname>Nie</surname> <given-names>J.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>She</surname> <given-names>W.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Advanced plant phenotyping: unmanned aerial vehicle remote sensing and cimageA software technology for precision crop growth monitoring</article-title>. <source>Agronomy</source> <volume>14</volume>, <elocation-id>2534</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy14112534</pub-id>
</mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Gaillard</surname> <given-names>M.</given-names></name>
<name><surname>Miao</surname> <given-names>C.</given-names></name>
<name><surname>Schnable</surname> <given-names>J. C.</given-names></name>
<name><surname>Benes</surname> <given-names>B.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Voxel carving-based 3D reconstruction of sorghum identifies genetic determinants of light interception efficiency</article-title>. <source>Plant Direct.</source> <volume>4</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/pld3.255</pub-id>, PMID: <pub-id pub-id-type="pmid">33073164</pub-id>
</mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Gardezi</surname> <given-names>M.</given-names></name>
<name><surname>Joshi</surname> <given-names>B.</given-names></name>
<name><surname>Rizzo</surname> <given-names>D. M.</given-names></name>
<name><surname>Ryan</surname> <given-names>M.</given-names></name>
<name><surname>Prutzer</surname> <given-names>E.</given-names></name>
<name><surname>Brugler</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Artificial intelligence in farming: Challenges and opportunities for building trust</article-title>. <source>Agron. J.</source> <volume>116</volume>, <fpage>1217</fpage>&#x2013;<lpage>1228</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/agj2.21353</pub-id>
</mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Gehan</surname> <given-names>M. A.</given-names></name>
<name><surname>Fahlgren</surname> <given-names>N.</given-names></name>
<name><surname>Abbasi</surname> <given-names>A.</given-names></name>
<name><surname>Berry</surname> <given-names>J. C.</given-names></name>
<name><surname>Callen</surname> <given-names>S. T.</given-names></name>
<name><surname>Chavez</surname> <given-names>L.</given-names></name>
<etal/>
</person-group>. (<year>2017</year>). 
<article-title>PlantCV v2: Image analysis software for high-throughput plant phenotyping</article-title>. <source>PeerJ.</source> <volume>5</volume>, <elocation-id>e4088</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.7717/peerj.4088</pub-id>, PMID: <pub-id pub-id-type="pmid">29209576</pub-id>
</mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ghanbari</surname> <given-names>A.</given-names></name>
<name><surname>Shirdel</surname> <given-names>G. H.</given-names></name>
<name><surname>Maleki</surname> <given-names>F.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Semi-self-supervised domain adaptation: developing deep learning models with limited annotated data for wheat head segmentation</article-title>. <source>Algorithms.</source> <volume>17</volume>, <fpage>1</fpage>&#x2013;<lpage>12</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/a17060267</pub-id>
</mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Gianola</surname> <given-names>D.</given-names></name>
<name><surname>van Kaam</surname> <given-names>JBCHM.</given-names></name>
</person-group> (<year>2008</year>). 
<article-title>Reproducing kernel hilbert spaces regression methods for genomic assisted prediction of quantitative traits</article-title>. <source>Genetics.</source> <volume>178</volume>, <fpage>2289</fpage>&#x2013;<lpage>2303</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1534/genetics.107.084285</pub-id>, PMID: <pub-id pub-id-type="pmid">18430950</pub-id>
</mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Gilmer</surname> <given-names>J.</given-names></name>
<name><surname>Schoenholz</surname> <given-names>S. S.</given-names></name>
<name><surname>Riley</surname> <given-names>P. F.</given-names></name>
<name><surname>Vinyals</surname> <given-names>O.</given-names></name>
<name><surname>Dahl</surname> <given-names>G. E.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>Neural message passing for Quantum chemistry</article-title>,&#x201d; in <source>Proceedings of the 34th international conference on machine learning</source>, vol. <volume>70</volume>. (<publisher-loc>Sydney, New South Wales, Australia</publisher-loc>: 
<publisher-name>JMLR.org</publisher-name>), <fpage>1263</fpage>&#x2013;<lpage>1272</lpage>. (ICML&#x2019;17). doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1704.01212</pub-id>
</mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Gopalan</surname> <given-names>B.</given-names></name>
<name><surname>Nascimento</surname> <given-names>N.</given-names></name>
<name><surname>Monga</surname> <given-names>V.</given-names></name>
</person-group> (<year>2025</year>). <source>Modular transformer architecture for precision agriculture imaging</source>, arXiv preprint arXiv:2508.03751. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2508.03751</pub-id>
</mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Goyal</surname> <given-names>P.</given-names></name>
<name><surname>Mitra</surname> <given-names>A.</given-names></name>
<name><surname>Sinha</surname> <given-names>M.</given-names></name>
</person-group> (<year>2025</year>). &#x201c;
<article-title>SepHRNet: Generating High-Resolution Crop Maps from Remote Sensing imagery using HRNet with Separable Convolution</article-title>,&#x201d; in <conf-name>Proceedings of the 8th International Conference on Data Science and Management of Data (12th ACM IKDD CODS and 30th COMAD)</conf-name>. <publisher-loc>Jodhpur, India</publisher-loc>: 
<publisher-name>Association for Computing Machinery (ACM)</publisher-name>, <fpage>27</fpage>&#x2013;<lpage>34</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1145/3703323.3703327</pub-id>
</mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Guo</surname> <given-names>W.</given-names></name>
<name><surname>Carroll</surname> <given-names>M. E.</given-names></name>
<name><surname>Singh</surname> <given-names>A.</given-names></name>
<name><surname>Swetnam</surname> <given-names>T. L.</given-names></name>
<name><surname>Merchant</surname> <given-names>N.</given-names></name>
<name><surname>Sarkar</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>UAS-based plant phenotyping for research and breeding applications</article-title>. <source>Plant Phenomics</source> <volume>2021</volume>, <elocation-id>9840192</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2021/9840192</pub-id>, PMID: <pub-id pub-id-type="pmid">34195621</pub-id>
</mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Guo</surname> <given-names>A.</given-names></name>
<name><surname>Sun</surname> <given-names>K.</given-names></name>
<name><surname>Wang</surname> <given-names>M.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Domain adaptive fruit detection method based on multiple alignments</article-title>. <source>J. Intell. \&amp; Fuzzy Syst.</source> <volume>45</volume>, <fpage>5837</fpage>&#x2013;<lpage>5851</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3233/JIFS-232104</pub-id>
</mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Gupta</surname> <given-names>A.</given-names></name>
<name><surname>Singh</surname> <given-names>A.</given-names></name>
</person-group> (<year>2023</year>). <source>Agri-GNN: A novel genotypic-topological graph neural network framework built on graphSAGE for optimized yield prediction</source>, Vol. <volume>1&#x2013;19</volume>, arXiv preprint arXiv:2310.13037. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2310.13037</pub-id>
</mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Han</surname> <given-names>J.</given-names></name>
<name><surname>Kwon</surname> <given-names>Y.</given-names></name>
<name><surname>Choi</surname> <given-names>Y. S.</given-names></name>
<name><surname>Kang</surname> <given-names>S.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Improving chemical reaction yield prediction using pre-trained graph neural networks</article-title>. <source>J. Cheminform</source> <volume>16</volume>, <fpage>25</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13321-024-00818-z</pub-id>, PMID: <pub-id pub-id-type="pmid">38429787</pub-id>
</mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Han</surname> <given-names>S.</given-names></name>
<name><surname>Mao</surname> <given-names>H.</given-names></name>
<name><surname>Dally</surname> <given-names>W. J.</given-names></name>
</person-group> (<year>2016</year>). &#x201c;
<article-title>Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding</article-title>,&#x201d; in <conf-name>International Conference on Learning Representations 2016 (ICLR 2016)</conf-name>. <publisher-loc>San Juan, Puerto Rico, USA</publisher-loc>: 
<publisher-name>OpenReview / ICLR</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>.
</mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>He</surname> <given-names>K.</given-names></name>
<name><surname>Zhang</surname> <given-names>X.</given-names></name>
<name><surname>Ren</surname> <given-names>S.</given-names></name>
<name><surname>Sun</surname> <given-names>J.</given-names></name>
</person-group> (<year>2016</year>). &#x201c;
<article-title>Deep residual learning for image recognition</article-title>,&#x201d; in <source>2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016)</source>. <publisher-loc>Las Vegas, Nevada, USA</publisher-loc>: 
<publisher-name>IEEE</publisher-name>, <fpage>770</fpage>&#x2013;<lpage>778</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2016.90</pub-id>
</mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hochreiter</surname> <given-names>S.</given-names></name>
<name><surname>Schmidhuber</surname> <given-names>J.</given-names></name>
</person-group> (<year>1997</year>). 
<article-title>Long short-term memory</article-title>. <source>Neural Comput.</source> <volume>9</volume>, <fpage>1735</fpage>&#x2013;<lpage>1780</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id>, PMID: <pub-id pub-id-type="pmid">9377276</pub-id>
</mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Howard</surname> <given-names>A.</given-names></name>
<name><surname>Sandler</surname> <given-names>M.</given-names></name>
<name><surname>Chen</surname> <given-names>B.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>Chen</surname> <given-names>L. C.</given-names></name>
<name><surname>Tan</surname> <given-names>M.</given-names></name>
<etal/>
</person-group>. (<year>2019</year>). &#x201c;
<article-title>Searching for mobileNetV3</article-title>,&#x201d; in <conf-name>Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV 2019)</conf-name>. <publisher-loc>Seoul, Republic of Korea</publisher-loc>: 
<publisher-name>IEEE/CVF</publisher-name>, <fpage>1314</fpage>&#x2013;<lpage>1324</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICCV.2019.00140</pub-id>
</mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Hu</surname> <given-names>B.</given-names></name>
<name><surname>Jiang</surname> <given-names>W.</given-names></name>
<name><surname>Zeng</surname> <given-names>J.</given-names></name>
<name><surname>Cheng</surname> <given-names>C.</given-names></name>
<name><surname>He</surname> <given-names>L.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>FOTCA: hybrid transformer-CNN architecture using AFNO for accurate plant leaf disease image recognition</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>, <elocation-id>1&#x2013;12</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1231903</pub-id>, PMID: <pub-id pub-id-type="pmid">37771483</pub-id>
</mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jain</surname> <given-names>M.</given-names></name>
<name><surname>Srivastava</surname> <given-names>A. K.</given-names></name>
<name><surname>Balwinder-Singh</surname></name>
<name><surname>Joon</surname> <given-names>R. K.</given-names></name>
<name><surname>McDonald</surname> <given-names>A.</given-names></name>
<name><surname>Royal</surname> <given-names>K.</given-names></name>
<etal/>
</person-group>. (<year>2016</year>). 
<article-title>Mapping smallholder wheat yields and sowing dates using micro-satellite data</article-title>. <source>Remote Sens</source> <volume>8</volume>, <elocation-id>860</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs8100860</pub-id>
</mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>James</surname> <given-names>C.</given-names></name>
<name><surname>Chandra</surname> <given-names>S. S.</given-names></name>
<name><surname>Chapman</surname> <given-names>S. C.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>A scalable and efficient UAV-based pipeline and deep learning framework for phenotyping sorghum panicle morphology from point clouds</article-title>. <source>Plant Phenomics</source> <volume>7</volume>, <elocation-id>100050</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.plaphe.2025.100050</pub-id>, PMID: <pub-id pub-id-type="pmid">41415167</pub-id>
</mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jeong</surname> <given-names>J. H.</given-names></name>
<name><surname>Resop</surname> <given-names>J. P.</given-names></name>
<name><surname>Mueller</surname> <given-names>N. D.</given-names></name>
<name><surname>Fleisher</surname> <given-names>D. H.</given-names></name>
<name><surname>Yun</surname> <given-names>K.</given-names></name>
<name><surname>Butler</surname> <given-names>E. E.</given-names></name>
<etal/>
</person-group>. (<year>2016</year>). 
<article-title>Random forests for global and regional crop yield predictions</article-title>. <source>PloS One</source> <volume>11</volume>, <elocation-id>e0156571</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1371/journal.pone.0156571</pub-id>, PMID: <pub-id pub-id-type="pmid">27257967</pub-id>
</mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jiang</surname> <given-names>Y.</given-names></name>
<name><surname>Li</surname> <given-names>C.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Convolutional neural networks for image-based high-throughput plant phenotyping: A review</article-title>. <source>Plant Phenomics.</source> <volume>2020</volume>, <elocation-id>4152816</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2020/4152816</pub-id>, PMID: <pub-id pub-id-type="pmid">33313554</pub-id>
</mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jiaxuan</surname> <given-names>Y.</given-names></name>
<name><surname>Xiaocheng</surname> <given-names>L.</given-names></name>
<name><surname>Melvin</surname> <given-names>L.</given-names></name>
<name><surname>David Lobell</surname> <given-names>S. E.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>Deep gaussian process for crop yield prediction based on remote sensing data</article-title>. <source>Proc. Thirty-First AAAI Conf Artif. Intell.</source> <volume>2</volume>, <fpage>4559</fpage>&#x2013;<lpage>4565</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1609/aaai.v31i1.11172</pub-id>
</mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Joshi</surname> <given-names>A.</given-names></name>
<name><surname>Guevara</surname> <given-names>D.</given-names></name>
<name><surname>Earles</surname> <given-names>M.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Standardizing and centralizing datasets for efficient training of agricultural deep learning models</article-title>. <source>Plant Phenomics</source> <volume>5</volume>, <elocation-id>84</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0084</pub-id>, PMID: <pub-id pub-id-type="pmid">37680999</pub-id>
</mixed-citation>
</ref>
<ref id="B60">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Jumper</surname> <given-names>J.</given-names></name>
<name><surname>Evans</surname> <given-names>R.</given-names></name>
<name><surname>Pritzel</surname> <given-names>A.</given-names></name>
<name><surname>Green</surname> <given-names>T.</given-names></name>
<name><surname>Figurnov</surname> <given-names>M.</given-names></name>
<name><surname>Ronneberger</surname> <given-names>O.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Highly accurate protein structure prediction with AlphaFold</article-title>. <source>Nature</source> <volume>596</volume>, <fpage>583</fpage>&#x2013;<lpage>589</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41586-021-03819-2</pub-id>, PMID: <pub-id pub-id-type="pmid">34265844</pub-id>
</mixed-citation>
</ref>
<ref id="B61">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Kairouz</surname> <given-names>P.</given-names></name>
<name><surname>McMahan</surname> <given-names>H. B.</given-names></name>
<name><surname>Avent</surname> <given-names>B.</given-names></name>
<name><surname>Bellet</surname> <given-names>A.</given-names></name>
<name><surname>Bennis</surname> <given-names>M.</given-names></name>
<name><surname>Bhagoji</surname> <given-names>A. N.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Advances and open problems in federated learning</article-title>. <source>Foundations Trends Mach. Learn.</source> <volume>14</volume>, <fpage>1</fpage>&#x2013;<lpage>210</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1561/2200000083</pub-id>
</mixed-citation>
</ref>
<ref id="B62">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Karmakar</surname> <given-names>P.</given-names></name>
<name><surname>Teng</surname> <given-names>S. W.</given-names></name>
<name><surname>Murshed</surname> <given-names>M.</given-names></name>
<name><surname>Pang</surname> <given-names>S.</given-names></name>
<name><surname>Li</surname> <given-names>Y.</given-names></name>
<name><surname>Lin</surname> <given-names>H.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Crop monitoring by multimodal remote sensing: A review</article-title>. <source>Remote Sens Appl. Soc. Environ.</source> <volume>33</volume>, <elocation-id>101093</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.rsase.2023.101093</pub-id>
</mixed-citation>
</ref>
<ref id="B63">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Khaki</surname> <given-names>S.</given-names></name>
<name><surname>Wang</surname> <given-names>L.</given-names></name>
</person-group> (<year>2019</year>). 
<article-title>Crop yield prediction using deep neural networks</article-title>. <source>Front. Plant Sci.</source> <volume>10</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2019.00621</pub-id>, PMID: <pub-id pub-id-type="pmid">31191564</pub-id>
</mixed-citation>
</ref>
<ref id="B64">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Khuimphukhieo</surname> <given-names>I.</given-names></name>
<name><surname>da Silva</surname> <given-names>J. A.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Unmanned aerial systems (UAS)-based field high throughput phenotyping (HTP) as plant breeders&#x2019; toolbox: A comprehensive review</article-title>. <source>Smart Agric. Technol.</source> <volume>11</volume>, <elocation-id>100888</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.atech.2025.100888</pub-id>
</mixed-citation>
</ref>
<ref id="B65">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Kipf</surname> <given-names>T. N.</given-names></name>
<name><surname>Welling</surname> <given-names>M.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>Semi-Supervised classfication with graph convolutional networks</article-title>,&#x201d; in <source>5th International Conference on Learning Representations (ICLR 2017)</source>. <publisher-loc>Palais des Congr&#xe8;s Neptune, Toulon, France</publisher-loc>: 
<publisher-name>ICLR</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. arXiv preprint arXiv: 1609.02907. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1609.02907</pub-id>
</mixed-citation>
</ref>
<ref id="B66">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Knecht</surname> <given-names>A. C.</given-names></name>
<name><surname>Campbell</surname> <given-names>M. T.</given-names></name>
<name><surname>Caprez</surname> <given-names>A.</given-names></name>
<name><surname>Swanson</surname> <given-names>D. R.</given-names></name>
<name><surname>Walia</surname> <given-names>H.</given-names></name>
</person-group> (<year>2016</year>). 
<article-title>Image Harvest: an open-source platform for high-throughput plant image processing and analysis</article-title>. <source>J. Exp. Bot.</source> <volume>67</volume>, <fpage>3587</fpage>&#x2013;<lpage>3599</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/jxb/erw176</pub-id>, PMID: <pub-id pub-id-type="pmid">27141917</pub-id>
</mixed-citation>
</ref>
<ref id="B67">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Krizhevsky</surname> <given-names>A.</given-names></name>
<name><surname>Sutskever</surname> <given-names>I.</given-names></name>
<name><surname>Hinton</surname> <given-names>G. E.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>ImageNet classification with deep convolutional neural networks</article-title>. <source>Commun. ACM</source> <volume>60</volume>, <fpage>84</fpage>&#x2013;<lpage>90</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1145/3065386</pub-id>
</mixed-citation>
</ref>
<ref id="B68">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Kumar</surname> <given-names>A.</given-names></name>
<name><surname>Khare</surname> <given-names>S.</given-names></name>
<name><surname>Rossi</surname> <given-names>S.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>PhenoAI: A deep learning Python framework to process close-range time-lapse PhenoCam data</article-title>. <source>Ecol. Inform</source> <volume>88</volume>, <elocation-id>103134</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ecoinf.2025.103134</pub-id>
</mixed-citation>
</ref>
<ref id="B69">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Kumar</surname> <given-names>C.</given-names></name>
<name><surname>Mubvumba</surname> <given-names>P.</given-names></name>
<name><surname>Huang</surname> <given-names>Y.</given-names></name>
<name><surname>Dhillon</surname> <given-names>J.</given-names></name>
<name><surname>Reddy</surname> <given-names>K.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Multi-stage corn yield prediction using high-resolution UAV multispectral data and machine learning models</article-title>. <source>Agronomy</source> <volume>13</volume>, <elocation-id>1277</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy13051277</pub-id>
</mixed-citation>
</ref>
<ref id="B70">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lan</surname> <given-names>Y.</given-names></name>
<name><surname>Huang</surname> <given-names>K.</given-names></name>
<name><surname>Yang</surname> <given-names>C.</given-names></name>
<name><surname>Lei</surname> <given-names>L.</given-names></name>
<name><surname>Ye</surname> <given-names>J.</given-names></name>
<name><surname>Zhang</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Real-time identification of rice weeds by UAV low-altitude remote sensing based on improved semantic segmentation model</article-title>. <source>Remote Sens</source> <volume>13</volume>, <elocation-id>4370</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs13214370</pub-id>
</mixed-citation>
</ref>
<ref id="B71">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lecun</surname> <given-names>Y.</given-names></name>
<name><surname>Bottou</surname> <given-names>L.</given-names></name>
<name><surname>Bengio</surname> <given-names>Y.</given-names></name>
<name><surname>Haffner</surname> <given-names>P.</given-names></name>
</person-group> (<year>1998</year>). 
<article-title>Gradient-based learning applied to document recognition</article-title>. <source>Proc. IEEE.</source> <volume>86</volume>, <fpage>2278</fpage>&#x2013;<lpage>2324</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/5.726791</pub-id>
</mixed-citation>
</ref>
<ref id="B72">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>D.</given-names></name>
<name><surname>Li</surname> <given-names>J.</given-names></name>
<name><surname>Xiang</surname> <given-names>S.</given-names></name>
<name><surname>Pan</surname> <given-names>A.</given-names></name>
</person-group> (<year>2022</year>). 
<article-title>PSegNet: simultaneous semantic and instance segmentation for point clouds of plants</article-title>. <source>Plant phenomics (Washington DC).</source> <volume>2022</volume>, <elocation-id>9787643</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/2022/9787643</pub-id>, PMID: <pub-id pub-id-type="pmid">35693119</pub-id>
</mixed-citation>
</ref>
<ref id="B73">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>W.</given-names></name>
<name><surname>Liang</surname> <given-names>S.</given-names></name>
<name><surname>Chen</surname> <given-names>K.</given-names></name>
<name><surname>Chen</surname> <given-names>Y.</given-names></name>
<name><surname>Ma</surname> <given-names>H.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). <source>AgriFM: A multi-source temporal remote sensing foundation model for crop mapping</source>, arXiv preprint arXiv:2505.21357. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2505.21357</pub-id>
</mixed-citation>
</ref>
<ref id="B74">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Li</surname> <given-names>G.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Zhao</surname> <given-names>Q.</given-names></name>
<name><surname>Yuan</surname> <given-names>P.</given-names></name>
<name><surname>Chang</surname> <given-names>B.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>PMVT: a lightweight vision transformer for plant disease identification on mobile devices</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2023.1256773</pub-id>, PMID: <pub-id pub-id-type="pmid">37822342</pub-id>
</mixed-citation>
</ref>
<ref id="B75">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Lin</surname> <given-names>F.</given-names></name>
<name><surname>Crawford</surname> <given-names>S.</given-names></name>
<name><surname>Guillot</surname> <given-names>K.</given-names></name>
<name><surname>Zhang</surname> <given-names>Y.</given-names></name>
<name><surname>Chen</surname> <given-names>Y.</given-names></name>
<name><surname>Yuan</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). &#x201c;
<article-title>MMST-viT: climate change-aware crop yield prediction via multi-modal spatial-temporal vision transformer</article-title>,&#x201d; in <conf-name>2023 IEEE/CVF International Conference on Computer Vision (ICCV 2023)</conf-name>. <publisher-loc>Paris, France</publisher-loc>: 
<publisher-name>IEEE/CVF</publisher-name>, <fpage>5751</fpage>&#x2013;<lpage>5761</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICCV51070.2023.00531</pub-id>
</mixed-citation>
</ref>
<ref id="B76">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lin</surname> <given-names>S.</given-names></name>
<name><surname>Qi</surname> <given-names>Z.</given-names></name>
<name><surname>Li</surname> <given-names>X.</given-names></name>
<name><surname>Zhang</surname> <given-names>H.</given-names></name>
<name><surname>Lv</surname> <given-names>Q.</given-names></name>
<name><surname>Huang</surname> <given-names>D.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>A phenological-knowledge-independent method for automatic paddy rice mapping with time series of polarimetric SAR images</article-title>. <source>ISPRS J. Photogramm Remote Sens</source> <volume>218</volume>, <fpage>628</fpage>&#x2013;<lpage>644</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.isprsjprs.2024.09.035</pub-id>
</mixed-citation>
</ref>
<ref id="B77">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lin</surname> <given-names>T.</given-names></name>
<name><surname>Zhong</surname> <given-names>R.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<name><surname>Jiang</surname> <given-names>H.</given-names></name>
<name><surname>Xu</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2020</year>). <volume>15</volume>, 
<article-title>DeepCropNet: a deep spatial-temporal learning framework for county-level corn yield estimation DeepCropNet: a deep spatial-temporal learning framework for county-level corn yield estimation</article-title>. <source>Environ. Res. Lett.</source><elocation-id>034016</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1088/1748-9326/ab66cb</pub-id>
</mixed-citation>
</ref>
<ref id="B78">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Lin</surname> <given-names>Y.</given-names></name>
<name><surname>Cao</surname> <given-names>Y.</given-names></name>
<name><surname>Hu</surname> <given-names>H.</given-names></name>
<name><surname>Wei</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). &#x201c;
<article-title>Swin transformer: hierarchical vision transformer using shifted windows</article-title>,&#x201d; in <conf-name>2021 IEEE/CVF International Conference on Computer Vision (ICCV 2021)</conf-name>. <publisher-loc>Montreal, QC, Canada</publisher-loc>: 
<publisher-name>IEEE/CVF</publisher-name>, <fpage>9992</fpage>&#x2013;<lpage>10002</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICCV48922.2021.00986</pub-id>
</mixed-citation>
</ref>
<ref id="B79">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Wang</surname> <given-names>S.</given-names></name>
<name><surname>Chen</surname> <given-names>J.</given-names></name>
<name><surname>Chen</surname> <given-names>B.</given-names></name>
<name><surname>Wang</surname> <given-names>X.</given-names></name>
<name><surname>Hao</surname> <given-names>D.</given-names></name>
<etal/>
</person-group>. (<year>2022</year>). 
<article-title>Rice yield prediction and model interpretation based on satellite and climatic indicators using a transformer method</article-title>. <source>Remote Sens.</source> <volume>14</volume>, <elocation-id>5045</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs14195045</pub-id>
</mixed-citation>
</ref>
<ref id="B80">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Lu</surname> <given-names>J.</given-names></name>
<name><surname>Batra</surname> <given-names>D.</given-names></name>
<name><surname>Parikh</surname> <given-names>D.</given-names></name>
<name><surname>Lee</surname> <given-names>S.</given-names></name>
</person-group> (<year>2019</year>). &#x201c;
<article-title>ViLBERT: pretraining task-agnostic visiolinguistic representations for vision-and-language tasks</article-title>,&#x201d; in <source>Proceedings of the 33rd international conference on neural information processing systems</source> (
<publisher-name>Curran Associates Inc</publisher-name>, <publisher-loc>Red Hook, NY, USA</publisher-loc>). doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1908.02265</pub-id>
</mixed-citation>
</ref>
<ref id="B81">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Lu</surname> <given-names>F.</given-names></name>
<name><surname>Zhang</surname> <given-names>B.</given-names></name>
<name><surname>Hou</surname> <given-names>Y.</given-names></name>
<name><surname>Xiong</surname> <given-names>X.</given-names></name>
<name><surname>Dong</surname> <given-names>C.</given-names></name>
<name><surname>Lu</surname> <given-names>W.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>A spatiotemporal attention-guided graph neural network for precise hyperspectral estimation of corn nitrogen content</article-title>. <source>Agronomy</source> <volume>15</volume>, <elocation-id>1041</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/agronomy15051041</pub-id>
</mixed-citation>
</ref>
<ref id="B82">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Lundberg</surname> <given-names>S. M.</given-names></name>
<name><surname>Lee</surname> <given-names>S. I.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>A unified approach to interpreting model predictions</article-title>,&#x201d; in <source>Proceedings of the 31st international conference on neural information processing systems</source> (
<publisher-name>Curran Associates Inc</publisher-name>, <publisher-loc>Red Hook, NY, USA</publisher-loc>), <fpage>4768</fpage>&#x2013;<lpage>4777</lpage>.
</mixed-citation>
</ref>
<ref id="B83">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Luo</surname> <given-names>J.</given-names></name>
<name><surname>Liu</surname> <given-names>Z.</given-names></name>
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Tang</surname> <given-names>A.</given-names></name>
<name><surname>Zuo</surname> <given-names>H.</given-names></name>
<name><surname>Han</surname> <given-names>P.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Efficient small object detection you only look once: A small object detection algorithm for aerial images</article-title>. <source>Sensors.</source> <volume>24</volume>, <elocation-id>7067</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s24217067</pub-id>, PMID: <pub-id pub-id-type="pmid">39517964</pub-id>
</mixed-citation>
</ref>
<ref id="B84">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ma</surname> <given-names>Y.</given-names></name>
<name><surname>Yang</surname> <given-names>Z.</given-names></name>
<name><surname>Huang</surname> <given-names>Q.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Improving the transferability of deep learning models for crop yield prediction: A partial domain adaptation approach</article-title>. <source>Remote Sensing.</source> <volume>15</volume>, <elocation-id>4562</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs15184562</pub-id>
</mixed-citation>
</ref>
<ref id="B85">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ma</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<name><surname>Yang</surname> <given-names>H. L.</given-names></name>
<name><surname>Yang</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>An adaptive adversarial domain adaptation approach for corn yield prediction</article-title>. <source>Comput. Electron Agric.</source> <volume>187</volume>, <elocation-id>106314</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2021.106314</pub-id>
</mixed-citation>
</ref>
<ref id="B86">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>McMahan</surname> <given-names>B.</given-names></name>
<name><surname>Moore</surname> <given-names>E.</given-names></name>
<name><surname>Ramage</surname> <given-names>D.</given-names></name>
<name><surname>Hampson</surname> <given-names>S.</given-names></name>
<name><surname>Arcas</surname> <given-names>B.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>Communication-efficient learning of deep networks from decentralized data</article-title>,&#x201d; in <source>Proceedings of the 20th international conference on artificial intelligence and statistics</source>, vol. <volume>54</volume> . Eds. 
<person-group person-group-type="editor">
<name><surname>Singh</surname> <given-names>A.</given-names></name>
<name><surname>Zhu</surname> <given-names>J.</given-names></name>
</person-group> (<publisher-loc>Fort Lauderdale, Florida, USA</publisher-loc>: 
<publisher-name>Proceedings of Machine Learning Research</publisher-name>), <fpage>1273</fpage>&#x2013;<lpage>1282</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1602.05629</pub-id>
</mixed-citation>
</ref>
<ref id="B87">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Mehdipour</surname> <given-names>S.</given-names></name>
<name><surname>Mirroshandel</surname> <given-names>S. A.</given-names></name>
<name><surname>Tabatabaei</surname> <given-names>S. A.</given-names></name>
</person-group> (<year>2025</year>). <source>Vision transformers in precision agriculture: A comprehensive survey</source>, arXiv preprint arXiv: 2504.21706. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2504.21706</pub-id>
</mixed-citation>
</ref>
<ref id="B88">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Minervini</surname> <given-names>M.</given-names></name>
<name><surname>Abdelsamea</surname> <given-names>M. M.</given-names></name>
<name><surname>Tsaftaris</surname> <given-names>S. A.</given-names></name>
</person-group> (<year>2014</year>). 
<article-title>Image-based plant phenotyping with incremental learning and active contours</article-title>. <source>Ecol. Inform</source> <volume>23</volume>, <fpage>35</fpage>&#x2013;<lpage>48</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ecoinf.2013.07.004</pub-id>
</mixed-citation>
</ref>
<ref id="B89">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Mohanty</surname> <given-names>S. P.</given-names></name>
<name><surname>Hughes</surname> <given-names>D. P.</given-names></name>
<name><surname>Salath&#xe9;</surname> <given-names>M.</given-names></name>
</person-group> (<year>2016</year>). 
<article-title>Using deep learning for image-based plant disease detection</article-title>. <source>Front. Plant Sci.</source> <volume>7</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2016.01419</pub-id>, PMID: <pub-id pub-id-type="pmid">27713752</pub-id>
</mixed-citation>
</ref>
<ref id="B90">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Montesinos L&#xf3;pez</surname> <given-names>O. A.</given-names></name>
<name><surname>Montesinos L&#xf3;pez</surname> <given-names>A.</given-names></name>
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
</person-group> (<year>2022</year>). &#x201c;
<article-title>Support vector machines and support vector regression BT</article-title>,&#x201d; in <source>Multivariate statistical machine learning methods for genomic prediction</source>. Eds. 
<person-group person-group-type="editor">
<name><surname>Montesinos L&#xf3;pez</surname> <given-names>O. A.</given-names></name>
<name><surname>Montesinos L&#xf3;pez</surname> <given-names>A.</given-names></name>
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
</person-group> (
<publisher-name>Springer International Publishing</publisher-name>, <publisher-loc>Cham</publisher-loc>), <fpage>337</fpage>&#x2013;<lpage>378</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-030-89010-0_9</pub-id>
</mixed-citation>
</ref>
<ref id="B91">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Mostafa</surname> <given-names>S.</given-names></name>
<name><surname>Mondal</surname> <given-names>D.</given-names></name>
<name><surname>Panjvani</surname> <given-names>K.</given-names></name>
<name><surname>Kochian</surname> <given-names>L.</given-names></name>
<name><surname>Stavness</surname> <given-names>I.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Explainable deep learning in plant phenotyping</article-title>. <source>Front. Artif. Intell.</source> <volume>6</volume>, <elocation-id>1203546</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/frai.2023.1203546</pub-id>, PMID: <pub-id pub-id-type="pmid">37795496</pub-id>
</mixed-citation>
</ref>
<ref id="B92">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Murphy</surname> <given-names>K. M.</given-names></name>
<name><surname>Ludwig</surname> <given-names>E.</given-names></name>
<name><surname>Gutierrez</surname> <given-names>J.</given-names></name>
<name><surname>Gehan</surname> <given-names>M. A.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Deep learning in image-based plant phenotyping</article-title>. <source>Annu. Rev. Plant Biol.</source> <volume>75</volume>, <fpage>771</fpage>&#x2013;<lpage>795</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1146/annurev-arplant-070523-042828</pub-id>, PMID: <pub-id pub-id-type="pmid">38382904</pub-id>
</mixed-citation>
</ref>
<ref id="B93">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Nabwire</surname> <given-names>S.</given-names></name>
<name><surname>Suh</surname> <given-names>H. K.</given-names></name>
<name><surname>Kim</surname> <given-names>M. S.</given-names></name>
<name><surname>Baek</surname> <given-names>I.</given-names></name>
<name><surname>Cho</surname> <given-names>B. K.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Review: Application of artificial intelligence in phenomics</article-title>. <source>Sensors.</source> <volume>21</volume>, <fpage>1</fpage>&#x2013;<lpage>19</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s21134363</pub-id>, PMID: <pub-id pub-id-type="pmid">34202291</pub-id>
</mixed-citation>
</ref>
<ref id="B94">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Najjar</surname> <given-names>H.</given-names></name>
<name><surname>Pathak</surname> <given-names>D.</given-names></name>
<name><surname>Nuske</surname> <given-names>M.</given-names></name>
<name><surname>Dengel</surname> <given-names>A.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Intrinsic explainability of multimodal learning for crop yield prediction</article-title>. <source>Comput. Electron Agric.</source> <volume>239</volume>, <elocation-id>111003</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2025.111003</pub-id>
</mixed-citation>
</ref>
<ref id="B95">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Nawaz</surname> <given-names>U.</given-names></name>
<name><surname>Zaheer</surname> <given-names>M. Z.</given-names></name>
<name><surname>Khan</surname> <given-names>F. S.</given-names></name>
<name><surname>Cholakkal</surname> <given-names>H.</given-names></name>
<name><surname>Khan</surname> <given-names>S.</given-names></name>
<name><surname>Anwer</surname> <given-names>R. M.</given-names></name>
</person-group> (<year>2025</year>). <source>AI in agriculture: A survey of deep learning techniques for crops, fisheries and livestock</source>. <fpage>1</fpage>&#x2013;<lpage>42</lpage>, arXiv preprint arXiv: 2507.22101. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2507.22101</pub-id>
</mixed-citation>
</ref>
<ref id="B96">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Nevavuori</surname> <given-names>P.</given-names></name>
<name><surname>Narra</surname> <given-names>N.</given-names></name>
<name><surname>Linna</surname> <given-names>P.</given-names></name>
<name><surname>Lipping</surname> <given-names>T.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Crop yield prediction using multitemporal UAV data and spatio-temporal deep learning models</article-title>. <source>Remote Sens</source> <volume>12</volume>, <elocation-id>4000</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs12234000</pub-id>
</mixed-citation>
</ref>
<ref id="B97">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Newman</surname> <given-names>S. J.</given-names></name>
<name><surname>Furbank</surname> <given-names>R. T.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Explainable machine learning models of major crop traits from satellite-monitored continent-wide field trial data</article-title>. <source>Nat. plants.</source> <volume>7</volume>, <fpage>1354</fpage>&#x2013;<lpage>1363</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41477-021-01001-0</pub-id>, PMID: <pub-id pub-id-type="pmid">34608272</pub-id>
</mixed-citation>
</ref>
<ref id="B98">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ngiam</surname> <given-names>J.</given-names></name>
<name><surname>Khosla</surname> <given-names>A.</given-names></name>
<name><surname>Kim</surname> <given-names>M.</given-names></name>
<name><surname>Nam</surname> <given-names>J.</given-names></name>
<name><surname>Lee</surname> <given-names>H.</given-names></name>
<name><surname>Ng</surname> <given-names>A. Y.</given-names></name>
<etal/>
</person-group>. (<year>2011</year>). 
<article-title>Multimodal deep learning. In Proceedings of the 28th International Conference on Machine Learning (ICML 2011)</article-title>. <publisher-loc>Bellevue, Washington, USA</publisher-loc>: 
<publisher-name>Microtome Publishing on behalf of the International Machine Learning Society (IMLS)</publisher-name>. <fpage>689</fpage>&#x2013;<lpage>696</lpage>.
</mixed-citation>
</ref>
<ref id="B99">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Nguyen</surname> <given-names>C.</given-names></name>
<name><surname>Sagan</surname> <given-names>V.</given-names></name>
<name><surname>Bhadra</surname> <given-names>S.</given-names></name>
<name><surname>Moose</surname> <given-names>S.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>UAV multisensory data fusion and multi-task deep learning for high-throughput maize phenotyping</article-title>. <source>Sensors</source> <volume>23</volume>, <elocation-id>1827</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s23041827</pub-id>, PMID: <pub-id pub-id-type="pmid">36850425</pub-id>
</mixed-citation>
</ref>
<ref id="B100">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ninomiya</surname> <given-names>S.</given-names></name>
</person-group> (<year>2022</year>). 
<article-title>High-throughput field crop phenotyping: current status and challenges</article-title>. <source>Breed Sci.</source> <volume>72</volume>, <fpage>3</fpage>&#x2013;<lpage>18</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1270/jsbbs.21069</pub-id>, PMID: <pub-id pub-id-type="pmid">36045897</pub-id>
</mixed-citation>
</ref>
<ref id="B101">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Nirosha</surname> <given-names>U.</given-names></name>
<name><surname>Vennila</surname> <given-names>G.</given-names></name>
</person-group> (<year>2025</year>a). &#x201c;
<article-title>Federated learning-powered hybrid model for crop prediction: combining graph neural networks with RNNs</article-title>,&#x201d; in <conf-name>2025 Global Conference in Emerging Technology (GINOTECH 2025)</conf-name>. <publisher-loc>PUNE, India</publisher-loc>: 
<publisher-name>IEEE</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>6</lpage>.
</mixed-citation>
</ref>
<ref id="B102">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Nirosha</surname> <given-names>U.</given-names></name>
<name><surname>Vennila</surname> <given-names>G.</given-names></name>
</person-group> (<year>2025</year>b). 
<article-title>Enhancing crop yield prediction for agriculture productivity using federated learning integrating with graph and recurrent neural networks model</article-title>. <source>Expert Syst. Appl.</source> <volume>289</volume>, <elocation-id>128312</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.eswa.2025.12832</pub-id>
</mixed-citation>
</ref>
<ref id="B103">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Nishio</surname> <given-names>M.</given-names></name>
<name><surname>Satoh</surname> <given-names>M.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>Genomic best linear unbiased prediction method including imprinting effects for genomic evaluation</article-title>. <source>Genet. Sel Evol.</source> <volume>47</volume>, <fpage>32</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s12711-015-0091-y</pub-id>, PMID: <pub-id pub-id-type="pmid">25928098</pub-id>
</mixed-citation>
</ref>
<ref id="B104">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Osco</surname> <given-names>L. P.</given-names></name>
<name><surname>Marcato Junior</surname> <given-names>J.</given-names></name>
<name><surname>Marques Ramos</surname> <given-names>A. P.</given-names></name>
<name><surname>de Castro Jorge</surname> <given-names>L. A.</given-names></name>
<name><surname>Fatholahi</surname> <given-names>S. N.</given-names></name>
<name><surname>de Andrade Silva</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>a). 
<article-title>A review on deep learning in UAV remote sensing</article-title>. <source>Int. J. Appl. Earth Obs Geoinf</source> <volume>102</volume>, <elocation-id>102456</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.jag.2021.102456</pub-id>
</mixed-citation>
</ref>
<ref id="B105">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Osco</surname> <given-names>L. P.</given-names></name>
<name><surname>Nogueira</surname> <given-names>K.</given-names></name>
<name><surname>Marques Ramos</surname> <given-names>A. P.</given-names></name>
<name><surname>Faita Pinheiro</surname> <given-names>M. M.</given-names></name>
<name><surname>Furuya</surname> <given-names>D. E. G.</given-names></name>
<name><surname>Gon&#xe7;alves</surname> <given-names>W. N.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>b). 
<article-title>Semantic segmentation of citrus-orchard using deep neural networks and multispectral UAV-based imagery</article-title>. <source>Precis Agric.</source> <volume>22</volume>, <fpage>1171</fpage>&#x2013;<lpage>1188</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s11119-020-09777-5</pub-id>
</mixed-citation>
</ref>
<ref id="B106">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Papoutsoglou</surname> <given-names>E. A.</given-names></name>
<name><surname>Athanasiadis</surname> <given-names>I. N.</given-names></name>
<name><surname>Visser</surname> <given-names>R. G. F.</given-names></name>
<name><surname>Finkers</surname> <given-names>R.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>The benefits and struggles of FAIR data: the case of reusing plant phenotyping data</article-title>. <source>Sci. data.</source> <volume>10</volume>, <fpage>457</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41597-023-02364-z</pub-id>, PMID: <pub-id pub-id-type="pmid">37443110</pub-id>
</mixed-citation>
</ref>
<ref id="B107">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Poorter</surname> <given-names>H.</given-names></name>
<name><surname>Niinemets</surname> <given-names>&#xdc;.</given-names></name>
<name><surname>Ntagkas</surname> <given-names>N.</given-names></name>
<name><surname>Siebenk&#xe4;s</surname> <given-names>A.</given-names></name>
<name><surname>M&#xe4;enp&#xe4;&#xe4;</surname> <given-names>M.</given-names></name>
<name><surname>Matsubara</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2019</year>). 
<article-title>A meta-analysis of plant responses to light intensity for 70 traits ranging from molecules to whole plant performance</article-title>. <source>New Phytol.</source> <volume>223</volume>, <fpage>1073</fpage>&#x2013;<lpage>1105</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1111/nph.15754</pub-id>, PMID: <pub-id pub-id-type="pmid">30802971</pub-id>
</mixed-citation>
</ref>
<ref id="B108">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Pound</surname> <given-names>M. P.</given-names></name>
<name><surname>Atkinson</surname> <given-names>J. A.</given-names></name>
<name><surname>Townsend</surname> <given-names>A. J.</given-names></name>
<name><surname>Wilson</surname> <given-names>M. H.</given-names></name>
<name><surname>Griffiths</surname> <given-names>M.</given-names></name>
<name><surname>Jackson</surname> <given-names>A. S.</given-names></name>
<etal/>
</person-group>. (<year>2017</year>). 
<article-title>Deep machine learning provides state-of-the-art performance in image-based plant phenotyping</article-title>. <source>Gigascience.</source> <volume>6</volume>, <fpage>1</fpage>&#x2013;<lpage>10</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/gigascience/gix083</pub-id>, PMID: <pub-id pub-id-type="pmid">29020747</pub-id>
</mixed-citation>
</ref>
<ref id="B109">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Pyzer-Knapp</surname> <given-names>E. O.</given-names></name>
<name><surname>Pitera</surname> <given-names>J. W.</given-names></name>
<name><surname>Staar</surname> <given-names>P. W. J.</given-names></name>
<name><surname>Takeda</surname> <given-names>S.</given-names></name>
<name><surname>Laino</surname> <given-names>T.</given-names></name>
<name><surname>Sanders</surname> <given-names>D. P.</given-names></name>
<etal/>
</person-group>. (<year>2022</year>). 
<article-title>Accelerating materials discovery using artificial intelligence, high performance computing and robotics</article-title>. <source>NPJ Comput. Mater</source> <volume>8</volume>, <fpage>84</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41524-022-00765-z</pub-id>
</mixed-citation>
</ref>
<ref id="B110">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Qiao</surname> <given-names>M.</given-names></name>
<name><surname>He</surname> <given-names>X.</given-names></name>
<name><surname>Cheng</surname> <given-names>X.</given-names></name>
<name><surname>Li</surname> <given-names>P.</given-names></name>
<name><surname>Zhao</surname> <given-names>Q.</given-names></name>
<name><surname>Zhao</surname> <given-names>C.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>KSTAGE: A knowledge-guided spatial-temporal attention graph learning network for crop yield prediction</article-title>. <source>Inf Sci. (Ny)</source> <volume>619</volume>, <fpage>19</fpage>&#x2013;<lpage>37</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.ins.2022.10.112</pub-id>
</mixed-citation>
</ref>
<ref id="B111">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Radford</surname> <given-names>A.</given-names></name>
<name><surname>Kim</surname> <given-names>J. W.</given-names></name>
<name><surname>Hallacy</surname> <given-names>C.</given-names></name>
<name><surname>Ramesh</surname> <given-names>A.</given-names></name>
<name><surname>Goh</surname> <given-names>G.</given-names></name>
<name><surname>Agarwal</surname> <given-names>S.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Learning transferable visual models from natural language supervision</article-title>. <source>Proc. Mach. Learn Res.</source> <volume>139</volume>, <fpage>8748</fpage>&#x2013;<lpage>8763</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2103.00020</pub-id>
</mixed-citation>
</ref>
<ref id="B112">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Rajpurkar</surname> <given-names>P.</given-names></name>
<name><surname>Chen</surname> <given-names>E.</given-names></name>
<name><surname>Banerjee</surname> <given-names>O.</given-names></name>
<name><surname>Topol</surname> <given-names>E. J.</given-names></name>
</person-group> (<year>2022</year>). 
<article-title>AI in health and medicine</article-title>. <source>Nat. Med.</source> <volume>28</volume>, <fpage>31</fpage>&#x2013;<lpage>38</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41591-021-01614-0</pub-id>, PMID: <pub-id pub-id-type="pmid">35058619</pub-id>
</mixed-citation>
</ref>
<ref id="B113">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Reddy</surname> <given-names>J.</given-names></name>
<name><surname>Niu</surname> <given-names>H.</given-names></name>
<name><surname>Scott</surname> <given-names>J. L. L.</given-names></name>
<name><surname>Bhandari</surname> <given-names>M.</given-names></name>
<name><surname>Landivar</surname> <given-names>J. A.</given-names></name>
<name><surname>Bednarz</surname> <given-names>C. W.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Cotton yield prediction via UAV-based cotton boll image segmentation using YOLO model and segment anything model (SAM)</article-title>. <source>Remote Sens</source> <volume>16</volume>, <elocation-id>4346</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs16234346</pub-id>
</mixed-citation>
</ref>
<ref id="B114">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ren</surname> <given-names>S.</given-names></name>
<name><surname>He</surname> <given-names>K.</given-names></name>
<name><surname>Girshick</surname> <given-names>R.</given-names></name>
<name><surname>Sun</surname> <given-names>J.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>Faster R-CNN: towards real-time object detection with region proposal networks</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>39</volume>, <fpage>1137</fpage>&#x2013;<lpage>1149</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TPAMI.2016.2577031</pub-id>, PMID: <pub-id pub-id-type="pmid">27295650</pub-id>
</mixed-citation>
</ref>
<ref id="B115">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Rojas</surname> <given-names>O.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Next generation agricultural stress index system (ASIS) for agricultural drought monitoring</article-title>. <source>Remote Sens</source> <volume>13</volume>, <elocation-id>959</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs13050959</pub-id>
</mixed-citation>
</ref>
<ref id="B116">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Ronneberger</surname> <given-names>O.</given-names></name>
<name><surname>Fischer</surname> <given-names>P.</given-names></name>
<name><surname>Brox</surname> <given-names>T.</given-names></name>
</person-group> (<year>2015</year>). &#x201c;
<article-title>U-net: convolutional networks for biomedical image segmentation BT</article-title>,&#x201d; in <source>Medical image computing and computer-assisted intervention &#x2013; MICCAI 2015</source>. Eds. 
<person-group person-group-type="editor">
<name><surname>Navab</surname> <given-names>N.</given-names></name>
<name><surname>Hornegger</surname> <given-names>J.</given-names></name>
<name><surname>Wells</surname> <given-names>W. M.</given-names></name>
<name><surname>Frangi</surname> <given-names>A. F.</given-names></name>
</person-group> (
<publisher-name>Springer International Publishing</publisher-name>, <publisher-loc>Cham</publisher-loc>), <fpage>234</fpage>&#x2013;<lpage>241</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id>
</mixed-citation>
</ref>
<ref id="B117">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sabo</surname> <given-names>F.</given-names></name>
<name><surname>Meroni</surname> <given-names>M.</given-names></name>
<name><surname>Waldner</surname> <given-names>F.</given-names></name>
<name><surname>Rembold</surname> <given-names>F.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Is deeper always better? Evaluating deep learning models for yield forecasting with small data</article-title>. <source>Environ. Monit Assess.</source> <volume>195</volume>, <fpage>1153</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s10661-023-11609-8</pub-id>, PMID: <pub-id pub-id-type="pmid">37672152</pub-id>
</mixed-citation>
</ref>
<ref id="B118">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Saha</surname> <given-names>S.</given-names></name>
<name><surname>Kucher</surname> <given-names>O. D.</given-names></name>
<name><surname>Utkina</surname> <given-names>A. O.</given-names></name>
<name><surname>Rebouh</surname> <given-names>N. Y.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Precision agriculture for improving crop yield predictions: a literature review</article-title>. <source>Front. Agron.</source> <volume>7</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fagro.2025.1566201</pub-id>
</mixed-citation>
</ref>
<ref id="B119">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sarkar</surname> <given-names>S.</given-names></name>
<name><surname>Dey</surname> <given-names>A.</given-names></name>
<name><surname>Pradhan</surname> <given-names>R.</given-names></name>
<name><surname>Sarkar</surname> <given-names>U. M.</given-names></name>
<name><surname>Chatterjee</surname> <given-names>C.</given-names></name>
<name><surname>Mondal</surname> <given-names>A.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Crop yield prediction using multimodal meta-transformer and temporal graph neural networks</article-title>. <source>IEEE Trans. AgriFood Electron.</source> <volume>2</volume>, <fpage>545</fpage>&#x2013;<lpage>553</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TAFE.2024.3438330</pub-id>
</mixed-citation>
</ref>
<ref id="B120">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Selvaraju</surname> <given-names>R. R.</given-names></name>
<name><surname>Cogswell</surname> <given-names>M.</given-names></name>
<name><surname>Das</surname> <given-names>A.</given-names></name>
<name><surname>Vedantam</surname> <given-names>R.</given-names></name>
<name><surname>Parikh</surname> <given-names>D.</given-names></name>
<name><surname>Batra</surname> <given-names>D.</given-names></name>
</person-group> (<year>2017</year>). &#x201c;
<article-title>Grad-CAM: visual explanations from deep networks via gradient-based localization</article-title>,&#x201d; in <conf-name>2017 IEEE International Conference on Computer Vision (ICCV)</conf-name>. <publisher-loc>Venice, Italy</publisher-loc>: 
<publisher-name>IEEE/CVF</publisher-name>, <fpage>618</fpage>&#x2013;<lpage>626</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICCV.2017.74</pub-id>
</mixed-citation>
</ref>
<ref id="B121">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Serrano</surname> <given-names>D. R.</given-names></name>
<name><surname>Luciano</surname> <given-names>F. C.</given-names></name>
<name><surname>Anaya</surname> <given-names>B. J.</given-names></name>
<name><surname>Ongoren</surname> <given-names>B.</given-names></name>
<name><surname>Kara</surname> <given-names>A.</given-names></name>
<name><surname>Molina</surname> <given-names>G.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Artificial intelligence (AI) applications in drug discovery and drug delivery: revolutionizing personalized medicine</article-title>. <source>Pharmaceutics</source> <volume>16</volume>, <elocation-id>1328</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/pharmaceutics16101328</pub-id>, PMID: <pub-id pub-id-type="pmid">39458657</pub-id>
</mixed-citation>
</ref>
<ref id="B122">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Shaban</surname> <given-names>A.</given-names></name>
<name><surname>Yousefi</surname> <given-names>S.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Multimodal deep learning</article-title>. <source>Springer Optim Its Appl.</source> <volume>211</volume>, <fpage>209</fpage>&#x2013;<lpage>219</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/978-3-031-53092-0_10</pub-id>
</mixed-citation>
</ref>
<ref id="B123">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Shahhosseini</surname> <given-names>M.</given-names></name>
<name><surname>Hu</surname> <given-names>G.</given-names></name>
<name><surname>Archontoulis</surname> <given-names>S. V.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Forecasting corn yield with machine learning ensembles</article-title>. <source>Front. Plant Sci.</source> <volume>11</volume>, <fpage>1</fpage>&#x2013;<lpage>16</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2020.01120</pub-id>, PMID: <pub-id pub-id-type="pmid">32849688</pub-id>
</mixed-citation>
</ref>
<ref id="B124">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Sharma</surname> <given-names>S.</given-names></name>
<name><surname>Rai</surname> <given-names>S.</given-names></name>
<name><surname>Krishnan</surname> <given-names>N. C.</given-names></name>
</person-group> (<year>2020</year>). <source>Wheat crop yield prediction using deep LSTM model</source>, arXiv preprint arXiv:2011.01498v1. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.2011.01498</pub-id>
</mixed-citation>
</ref>
<ref id="B125">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Shi</surname> <given-names>S.</given-names></name>
<name><surname>Li</surname> <given-names>X.</given-names></name>
<name><surname>Fang</surname> <given-names>L.</given-names></name>
<name><surname>Liu</surname> <given-names>A.</given-names></name>
<name><surname>Su</surname> <given-names>G.</given-names></name>
<name><surname>Zhang</surname> <given-names>Y.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Genomic prediction using bayesian regression models with global&#x2013;local prior</article-title>. <source>Front. Genet.</source> <volume>12</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fgene.2021.628205</pub-id>, PMID: <pub-id pub-id-type="pmid">33936162</pub-id>
</mixed-citation>
</ref>
<ref id="B126">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Simonyan</surname> <given-names>K.</given-names></name>
<name><surname>Zisserman</surname> <given-names>A.</given-names></name>
</person-group> (<year>2015</year>). &#x201c;
<article-title>Very deep convolutional networks for large-scale image recognition</article-title>,&#x201d; in <conf-name>3rd International Conference on Learning Representations (ICLR 2015)</conf-name>. <publisher-loc>San Diego, CA, USA</publisher-loc>: 
<publisher-name>ICLR</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>14</lpage>. arXiv preprint arXiv: 1409.1556.
</mixed-citation>
</ref>
<ref id="B127">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Song</surname> <given-names>C.</given-names></name>
<name><surname>Lin</surname> <given-names>Y.</given-names></name>
<name><surname>Guo</surname> <given-names>S.</given-names></name>
<name><surname>Wan</surname> <given-names>H.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting</article-title>. <source>Proc. AAAI Conf Artif. Intell.</source> <volume>34</volume>, <fpage>914</fpage>&#x2013;<lpage>921</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1609/aaai.v34i01.5438</pub-id>
</mixed-citation>
</ref>
<ref id="B128">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Stephan</surname> <given-names>J.</given-names></name>
<name><surname>Stegle</surname> <given-names>O.</given-names></name>
<name><surname>Beyer</surname> <given-names>A.</given-names></name>
</person-group> (<year>2015</year>). 
<article-title>A random forest approach to capture genetic effects in the presence of population structure</article-title>. <source>Nat. Commun.</source> <volume>6</volume>, <fpage>7432</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/ncomms8432</pub-id>, PMID: <pub-id pub-id-type="pmid">26109276</pub-id>
</mixed-citation>
</ref>
<ref id="B129">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sun</surname> <given-names>T.</given-names></name>
<name><surname>Chen</surname> <given-names>Y.</given-names></name>
<name><surname>Zheng</surname> <given-names>B.</given-names></name>
<name><surname>Sun</surname> <given-names>W.</given-names></name>
<name><surname>May</surname> <given-names>L. G.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Learning spatio-temporal dynamics for trajectory recovery via time-aware transformer</article-title>. <source>IEEE Trans. Intelligent Transportation Syst.</source> <volume>26</volume>, <fpage>16584</fpage>&#x2013;<lpage>16601</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TITS.2025.3574100</pub-id>
</mixed-citation>
</ref>
<ref id="B130">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Sun</surname> <given-names>J.</given-names></name>
<name><surname>Di</surname> <given-names>L.</given-names></name>
<name><surname>Sun</surname> <given-names>Z.</given-names></name>
<name><surname>Shen</surname> <given-names>Y.</given-names></name>
<name><surname>Lai</surname> <given-names>Z.</given-names></name>
</person-group> (<year>2019</year>). 
<article-title>County-level soybean yield prediction using deep CNN-LSTM model</article-title>. <source>Sensors</source> <volume>19</volume>, <elocation-id>4363</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s19204363</pub-id>, PMID: <pub-id pub-id-type="pmid">31600963</pub-id>
</mixed-citation>
</ref>
<ref id="B131">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Tan</surname> <given-names>H.</given-names></name>
<name><surname>Bansal</surname> <given-names>M.</given-names></name>
</person-group> (<year>2019</year>). &#x201c;
<article-title>Lxmert: learning cross-modality encoder representations from transformers</article-title>,&#x201d; in <source>Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP)</source>. Eds. 
<person-group person-group-type="editor">
<name><surname>Inui</surname> <given-names>K.</given-names></name>
<name><surname>Jiang</surname> <given-names>J.</given-names></name>
<name><surname>Ng</surname> <given-names>V.</given-names></name>
<name><surname>Wan</surname> <given-names>X.</given-names></name>
</person-group> (
<publisher-name>Association for Computational Linguistics</publisher-name>, <publisher-loc>Hong Kong, China</publisher-loc>), <fpage>5100</fpage>&#x2013;<lpage>5111</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.18653/v1/D19-1514</pub-id>
</mixed-citation>
</ref>
<ref id="B132">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Tang</surname> <given-names>Z.</given-names></name>
<name><surname>Parajuli</surname> <given-names>A.</given-names></name>
<name><surname>Chen</surname> <given-names>C. J.</given-names></name>
<name><surname>Hu</surname> <given-names>Y.</given-names></name>
<name><surname>Revolinski</surname> <given-names>S.</given-names></name>
<name><surname>Medina</surname> <given-names>C. A.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Validation of UAV-based alfalfa biomass predictability using photogrammetry with fully automatic plot segmentation</article-title>. <source>Sci. Rep.</source> <volume>11</volume>, <fpage>3336</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-021-82797-x</pub-id>, PMID: <pub-id pub-id-type="pmid">33558558</pub-id>
</mixed-citation>
</ref>
<ref id="B133">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Tardieu</surname> <given-names>F.</given-names></name>
<name><surname>Cabrera-Bosquet</surname> <given-names>L.</given-names></name>
<name><surname>Pridmore</surname> <given-names>T.</given-names></name>
<name><surname>Bennett</surname> <given-names>M.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>Plant phenomics, from sensors to knowledge</article-title>. <source>Curr. Biol.</source> <volume>27</volume>, <fpage>R770</fpage>&#x2013;<lpage>R783</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.cub.2017.05.055</pub-id>, PMID: <pub-id pub-id-type="pmid">28787611</pub-id>
</mixed-citation>
</ref>
<ref id="B134">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Tausen</surname> <given-names>M.</given-names></name>
<name><surname>Clausen</surname> <given-names>M.</given-names></name>
<name><surname>Moeskj&#xe6;r</surname> <given-names>S.</given-names></name>
<name><surname>Shihavuddin</surname> <given-names>A.</given-names></name>
<name><surname>Dahl</surname> <given-names>A. B.</given-names></name>
<name><surname>Janss</surname> <given-names>L.</given-names></name>
<etal/>
</person-group>. (<year>2020</year>). 
<article-title>Greenotyper: image-based plant phenotyping using distributed computing and deep learning</article-title>. <source>Front. Plant Sci.</source> <volume>11</volume>, <elocation-id>1181</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2020.01181</pub-id>, PMID: <pub-id pub-id-type="pmid">32849731</pub-id>
</mixed-citation>
</ref>
<ref id="B135">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Torgbor</surname> <given-names>B. A.</given-names></name>
<name><surname>Rahman</surname> <given-names>M. M.</given-names></name>
<name><surname>Brinkhoff</surname> <given-names>J.</given-names></name>
<name><surname>Sinha</surname> <given-names>P.</given-names></name>
<name><surname>Robson</surname> <given-names>A.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>Integrating remote sensing and weather variables for mango yield prediction using a machine learning approach</article-title>. <source>Remote Sens</source> <volume>15</volume>, <elocation-id>3075</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs15123075</pub-id>
</mixed-citation>
</ref>
<ref id="B136">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ubbens</surname> <given-names>J.</given-names></name>
<name><surname>Cieslak</surname> <given-names>M.</given-names></name>
<name><surname>Prusinkiewicz</surname> <given-names>P.</given-names></name>
<name><surname>Stavness</surname> <given-names>I.</given-names></name>
</person-group> (<year>2018</year>). 
<article-title>The use of plant models in deep learning: an application to leaf counting in rosette plants</article-title>. <source>Plant Methods</source> <volume>14</volume>, <elocation-id>6</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1186/s13007-018-0273-z</pub-id>, PMID: <pub-id pub-id-type="pmid">29375647</pub-id>
</mixed-citation>
</ref>
<ref id="B137">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Ubbens</surname> <given-names>J. R.</given-names></name>
<name><surname>Stavness</surname> <given-names>I.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>Deep plant phenomics: A deep learning platform for complex plant phenotyping tasks</article-title>. <source>Front. Plant Sci.</source> <volume>8</volume>, <elocation-id>1190</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2017.01190</pub-id>, PMID: <pub-id pub-id-type="pmid">28736569</pub-id>
</mixed-citation>
</ref>
<ref id="B138">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>van Dijk</surname> <given-names>M.</given-names></name>
<name><surname>Morley</surname> <given-names>T.</given-names></name>
<name><surname>Rau</surname> <given-names>M. L.</given-names></name>
<name><surname>Saghai</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>A meta-analysis of projected global food demand and population at risk of hunger for the period 2010&#x2013;2050</article-title>. <source>Nat. Food</source> <volume>2</volume>, <fpage>494</fpage>&#x2013;<lpage>501</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s43016-021-00322-9</pub-id>, PMID: <pub-id pub-id-type="pmid">37117684</pub-id>
</mixed-citation>
</ref>
<ref id="B139">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Van Hoolst</surname> <given-names>R.</given-names></name>
<name><surname>Eerens</surname> <given-names>H.</given-names></name>
<name><surname>Haesen</surname> <given-names>D.</given-names></name>
<name><surname>Royer</surname> <given-names>A.</given-names></name>
<name><surname>Bydekerke</surname> <given-names>L.</given-names></name>
<name><surname>Rojas</surname> <given-names>O.</given-names></name>
<etal/>
</person-group>. (<year>2016</year>). 
<article-title>FAO&#x2019;s AVHRR-based Agricultural Stress Index System ASIS for global drought monitoring</article-title>. <source>Int. J. Remote Sens</source> <volume>37</volume>, <fpage>418</fpage>&#x2013;<lpage>439</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1080/01431161.2015.1126378</pub-id>
</mixed-citation>
</ref>
<ref id="B140">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>van Klompenburg</surname> <given-names>T.</given-names></name>
<name><surname>Kassahun</surname> <given-names>A.</given-names></name>
<name><surname>Catal</surname> <given-names>C.</given-names></name>
</person-group> (<year>2020</year>). 
<article-title>Crop yield prediction using machine learning: A systematic literature review</article-title>. <source>Comput. Electron Agric.</source> <volume>177</volume>, <elocation-id>105709</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2020.105709</pub-id>
</mixed-citation>
</ref>
<ref id="B141">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Varela</surname> <given-names>S.</given-names></name>
<name><surname>Zheng</surname> <given-names>X.</given-names></name>
<name><surname>Njuguna</surname> <given-names>J.</given-names></name>
<name><surname>Sacks</surname> <given-names>E.</given-names></name>
<name><surname>Allen</surname> <given-names>D.</given-names></name>
<name><surname>Ruhter</surname> <given-names>J.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Breaking the barrier of human-annotated training data for machine learning-aided plant research using aerial imagery</article-title>. <source>Plant Physiol.</source> <volume>197</volume>, <elocation-id>kiaf132</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/plphys/kiaf132</pub-id>, PMID: <pub-id pub-id-type="pmid">40265604</pub-id>
</mixed-citation>
</ref>
<ref id="B142">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Vaswani</surname> <given-names>A.</given-names></name>
<name><surname>Shazeer</surname> <given-names>N.</given-names></name>
<name><surname>Parmar</surname> <given-names>N.</given-names></name>
<name><surname>Uszkoreit</surname> <given-names>J.</given-names></name>
<name><surname>Jones</surname> <given-names>L.</given-names></name>
<name><surname>Gomez</surname> <given-names>A. N.</given-names></name>
<etal/>
</person-group>. (<year>2017</year>). &#x201c;
<article-title>Attention is all you need</article-title>,&#x201d; in <source>Neural Information Processing Systems 2017 (NeurIPS 2017)</source>. <publisher-loc>Long Beach, CA, USA</publisher-loc>: 
<publisher-name>Curran Associates, Inc.</publisher-name>, <fpage>5999</fpage>&#x2013;<lpage>6009</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.5555/3295222.3295349</pub-id>
</mixed-citation>
</ref>
<ref id="B143">
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name><surname>Velickovic</surname> <given-names>P.</given-names></name>
<name><surname>Cucurull</surname> <given-names>G.</given-names></name>
<name><surname>Casanova</surname> <given-names>A.</given-names></name>
<name><surname>Romero</surname> <given-names>A.</given-names></name>
<name><surname>Lio&#x2019;</surname> <given-names>P.</given-names></name>
<name><surname>Bengio</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2017</year>). <source>Graph attention networks</source>, arXiv preprint arXiv:1710.10903. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1710.10903</pub-id>
</mixed-citation>
</ref>
<ref id="B144">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>K.</given-names></name>
<name><surname>Abid</surname> <given-names>M. A.</given-names></name>
<name><surname>Rasheed</surname> <given-names>A.</given-names></name>
<name><surname>Crossa</surname> <given-names>J.</given-names></name>
<name><surname>Hearne</surname> <given-names>S.</given-names></name>
<name><surname>Li</surname> <given-names>H.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>DNNGP, a deep neural network-based method for genomic prediction using multi-omics data in plants</article-title>. <source>Mol. Plant</source> <volume>16</volume>, <fpage>279</fpage>&#x2013;<lpage>293</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.molp.2022.11.004</pub-id>, PMID: <pub-id pub-id-type="pmid">36366781</pub-id>
</mixed-citation>
</ref>
<ref id="B145">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>C.</given-names></name>
<name><surname>Ling</surname> <given-names>L.</given-names></name>
<name><surname>Kuai</surname> <given-names>J.</given-names></name>
<name><surname>Xie</surname> <given-names>J.</given-names></name>
<name><surname>Ma</surname> <given-names>N.</given-names></name>
<name><surname>You</surname> <given-names>L.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Integrating UAV and satellite LAI data into a modified DSSAT-rapeseed model to improve yield predictions</article-title>. <source>F Crop Res.</source> <volume>327</volume>, <elocation-id>109883</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.fcr.2025.109883</pub-id>
</mixed-citation>
</ref>
<ref id="B146">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>H.</given-names></name>
<name><surname>Yan</surname> <given-names>S.</given-names></name>
<name><surname>Wang</surname> <given-names>W.</given-names></name>
<name><surname>Chen</surname> <given-names>Y.</given-names></name>
<name><surname>Hong</surname> <given-names>J.</given-names></name>
<name><surname>He</surname> <given-names>Q.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Cropformer: An interpretable deep learning framework for crop genomic prediction</article-title>. <source>Plant Commun.</source> <volume>6</volume>, <elocation-id>101223</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.xplc.2024.101223</pub-id>, PMID: <pub-id pub-id-type="pmid">39690739</pub-id>
</mixed-citation>
</ref>
<ref id="B147">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wang</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>Z.</given-names></name>
<name><surname>Feng</surname> <given-names>L.</given-names></name>
<name><surname>Ma</surname> <given-names>Y.</given-names></name>
<name><surname>Du</surname> <given-names>Q.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>A new attention-based CNN approach for crop mapping using time series Sentinel-2 images</article-title>. <source>Comput. Electron Agric.</source> <volume>184</volume>, <elocation-id>106090</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compag.2021.106090</pub-id>
</mixed-citation>
</ref>
<ref id="B148">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Waqas</surname> <given-names>M.</given-names></name>
<name><surname>Naseem</surname> <given-names>A.</given-names></name>
<name><surname>Humphries</surname> <given-names>U. W.</given-names></name>
<name><surname>Hlaing</surname> <given-names>P. T.</given-names></name>
<name><surname>Dechpichai</surname> <given-names>P.</given-names></name>
<name><surname>Wangwongchai</surname> <given-names>A.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>Applications of machine learning and deep learning in agriculture: A comprehensive review</article-title>. <source>Green Technol. Sustain</source> <volume>3</volume>, <elocation-id>100199</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.grets.2025.100199</pub-id>
</mixed-citation>
</ref>
<ref id="B149">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wu</surname> <given-names>H.</given-names></name>
<name><surname>Wu</surname> <given-names>W.</given-names></name>
<name><surname>Huang</surname> <given-names>Y.</given-names></name>
<name><surname>Liu</surname> <given-names>S.</given-names></name>
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Zhang</surname> <given-names>N.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>FEWheat-YOLO: A lightweight improved algorithm for wheat spike detection</article-title>. <source>Plants</source> <volume>14</volume>, <elocation-id>3058</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/plants14193058</pub-id>, PMID: <pub-id pub-id-type="pmid">41095199</pub-id>
</mixed-citation>
</ref>
<ref id="B150">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Wu</surname> <given-names>B.</given-names></name>
<name><surname>Zhang</surname> <given-names>M.</given-names></name>
<name><surname>Zeng</surname> <given-names>H.</given-names></name>
<name><surname>Tian</surname> <given-names>F.</given-names></name>
<name><surname>Potgieter</surname> <given-names>A. B.</given-names></name>
<name><surname>Qin</surname> <given-names>X.</given-names></name>
<etal/>
</person-group>. (<year>2023</year>). 
<article-title>Challenges and opportunities in remote sensing-based crop monitoring: a review</article-title>. <source>Natl. Sci. Rev.</source> <volume>10</volume>, <elocation-id>nwac290</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1093/nsr/nwac290</pub-id>, PMID: <pub-id pub-id-type="pmid">36960224</pub-id>
</mixed-citation>
</ref>
<ref id="B151">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Xie</surname> <given-names>W.</given-names></name>
<name><surname>Zhao</surname> <given-names>M.</given-names></name>
<name><surname>Liu</surname> <given-names>Y.</given-names></name>
<name><surname>Yang</surname> <given-names>D.</given-names></name>
<name><surname>Huang</surname> <given-names>K.</given-names></name>
<name><surname>Fan</surname> <given-names>C.</given-names></name>
<etal/>
</person-group>. (<year>2024</year>). 
<article-title>Recent advances in Transformer technology for agriculture: A comprehensive survey</article-title>. <source>Eng. Appl. Artif. Intell.</source> <volume>138</volume>, <elocation-id>109412</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.engappai.2024.109412</pub-id>
</mixed-citation>
</ref>
<ref id="B152">
<mixed-citation publication-type="confproc">
<person-group person-group-type="author">
<name><surname>Yang</surname> <given-names>C.</given-names></name>
<name><surname>Baireddy</surname> <given-names>S.</given-names></name>
<name><surname>Cai</surname> <given-names>E.</given-names></name>
<name><surname>Crawford</surname> <given-names>M.</given-names></name>
<name><surname>Delp</surname> <given-names>E. J.</given-names></name>
</person-group> (<year>2021</year>). &#x201c;
<article-title>Field-based plot extraction using UAV RGB images</article-title>,&#x201d; in <conf-name>2021 IEEE/CVF International Conference on Computer Vision (ICCV 2021)</conf-name>. <publisher-loc>Montreal, QC, Canada</publisher-loc>: 
<publisher-name>IEEE/CVF</publisher-name>, <fpage>1390</fpage>&#x2013;<lpage>1398</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ICCVW54120.2021.00160</pub-id>
</mixed-citation>
</ref>
<ref id="B153">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yang</surname> <given-names>G.</given-names></name>
<name><surname>Liu</surname> <given-names>J.</given-names></name>
<name><surname>Zhao</surname> <given-names>C.</given-names></name>
<name><surname>Li</surname> <given-names>Z.</given-names></name>
<name><surname>Huang</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2017</year>). 
<article-title>Unmanned aerial vehicle remote sensing for field-based crop phenotyping: current status and perspectives</article-title>. <source>Front. Plant Sci.</source> <volume>8</volume>, <elocation-id>1111</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2017.01111</pub-id>, PMID: <pub-id pub-id-type="pmid">28713402</pub-id>
</mixed-citation>
</ref>
<ref id="B154">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Yu</surname> <given-names>S.</given-names></name>
<name><surname>Xie</surname> <given-names>L.</given-names></name>
<name><surname>Dai</surname> <given-names>L.</given-names></name>
</person-group> (<year>2025</year>). 
<article-title>ST-CFI: Swin Transformer with convolutional feature interactions for identifying plant diseases</article-title>. <source>Sci. Rep.</source> <volume>15</volume>, <fpage>25000</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41598-025-08673-0</pub-id>, PMID: <pub-id pub-id-type="pmid">40640259</pub-id>
</mixed-citation>
</ref>
<ref id="B155">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zaka</surname> <given-names>M. M.</given-names></name>
<name><surname>Samat</surname> <given-names>A.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Advances in remote sensing and machine learning methods for invasive plants study: A comprehensive review</article-title>. <source>Remote Sens</source> <volume>16</volume>, <elocation-id>3781</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/rs16203781</pub-id>
</mixed-citation>
</ref>
<ref id="B156">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zenkl</surname> <given-names>R.</given-names></name>
<name><surname>Timofte</surname> <given-names>R.</given-names></name>
<name><surname>Kirchgessner</surname> <given-names>N.</given-names></name>
<name><surname>Roth</surname> <given-names>L.</given-names></name>
<name><surname>Hund</surname> <given-names>A.</given-names></name>
<name><surname>Van Gool</surname> <given-names>L.</given-names></name>
<etal/>
</person-group>. (<year>2021</year>). 
<article-title>Outdoor plant segmentation with deep learning for high-throughput field phenotyping on a diverse wheat dataset</article-title>. <source>Front. Plant Sci.</source> <volume>12</volume>, <elocation-id>774068</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2021.774068</pub-id>, PMID: <pub-id pub-id-type="pmid">35058948</pub-id>
</mixed-citation>
</ref>
<ref id="B157">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhan</surname> <given-names>Y.</given-names></name>
<name><surname>Zhou</surname> <given-names>Y.</given-names></name>
<name><surname>Bai</surname> <given-names>G.</given-names></name>
<name><surname>Ge</surname> <given-names>Y.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>Bagging improves the performance of deep learning-based semantic segmentation with limited labeled images: A case study of crop segmentation for high-throughput plant phenotyping</article-title>. <source>Sensors.</source> <volume>24</volume>, <elocation-id>3420</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3390/s24113420</pub-id>, PMID: <pub-id pub-id-type="pmid">38894212</pub-id>
</mixed-citation>
</ref>
<ref id="B158">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>W.</given-names></name>
<name><surname>Chen</surname> <given-names>K.</given-names></name>
<name><surname>Wang</surname> <given-names>J.</given-names></name>
<name><surname>Shi</surname> <given-names>Y.</given-names></name>
<name><surname>Guo</surname> <given-names>W.</given-names></name>
</person-group> (<year>2021</year>). 
<article-title>Easy domain adaptation method for filling the species gap in deep learning-based fruit detection</article-title>. <source>Hortic. Res.</source> <volume>8</volume>, <fpage>119</fpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1038/s41438-021-00553-8</pub-id>, PMID: <pub-id pub-id-type="pmid">34059636</pub-id>
</mixed-citation>
</ref>
<ref id="B159">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>L.</given-names></name>
<name><surname>Li</surname> <given-names>C.</given-names></name>
<name><surname>Wu</surname> <given-names>X.</given-names></name>
<name><surname>Xiang</surname> <given-names>H.</given-names></name>
<name><surname>Jiao</surname> <given-names>Y.</given-names></name>
<name><surname>Chai</surname> <given-names>H.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>BO-CNN-BiLSTM deep learning model integrating multisource remote sensing data for improving winter wheat yield estimation</article-title>. <source>Front. Plant Sci.</source> <volume>15</volume>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fpls.2024.1500499</pub-id>, PMID: <pub-id pub-id-type="pmid">39759241</pub-id>
</mixed-citation>
</ref>
<ref id="B160">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhang</surname> <given-names>W.</given-names></name>
<name><surname>Zheng</surname> <given-names>C.</given-names></name>
<name><surname>Wang</surname> <given-names>C.</given-names></name>
<name><surname>Guo</surname> <given-names>W.</given-names></name>
</person-group> (<year>2024</year>). 
<article-title>DomAda-fruitDet: domain-adaptive anchor-free fruit detection model for auto labeling</article-title>. <source>Plant phenomics (Washington DC).</source> <volume>6</volume>, <elocation-id>135</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.34133/plantphenomics.0135</pub-id>, PMID: <pub-id pub-id-type="pmid">38273867</pub-id>
</mixed-citation>
</ref>
<ref id="B161">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhou</surname> <given-names>H.</given-names></name>
<name><surname>Huang</surname> <given-names>F.</given-names></name>
<name><surname>Lou</surname> <given-names>W.</given-names></name>
<name><surname>Gu</surname> <given-names>Q.</given-names></name>
<name><surname>Ye</surname> <given-names>Z.</given-names></name>
<name><surname>Hu</surname> <given-names>H.</given-names></name>
<etal/>
</person-group>. (<year>2025</year>). 
<article-title>Yield prediction through UAV-based multispectral imaging and deep learning in rice breeding trials</article-title>. <source>Agric. Syst.</source> <volume>223</volume>, <elocation-id>104214</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.agsy.2024.104214</pub-id>
</mixed-citation>
</ref>
<ref id="B162">
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name><surname>Zhou</surname> <given-names>H.</given-names></name>
<name><surname>Jiang</surname> <given-names>F.</given-names></name>
<name><surname>Lu</surname> <given-names>H.</given-names></name>
</person-group> (<year>2023</year>). 
<article-title>SSDA-YOLO: Semi-supervised domain adaptive YOLO for cross-domain object detection</article-title>. <source>Comput. Vis. Image Underst.</source> <volume>229</volume>, <elocation-id>103649</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.cviu.2023.103649</pub-id>
</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn id="n1" fn-type="custom" custom-type="edited-by">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/313887">Ruslan Kalendar</ext-link>, University of Helsinki, Finland</p></fn>
<fn id="n2" fn-type="custom" custom-type="reviewed-by">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/441946">Chien Van Ha</ext-link>, Texas Tech University, United States</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/609983">Krishna Kumar Rai</ext-link>, Amity University, India</p></fn>
</fn-group>
<fn-group>
<fn fn-type="abbr" id="abbrev1">
<label>Abbreviations:</label>
<p>MLP, Multi-Layer Perceptron; MMST-ViT, Multi-Modal Spatial-Temporal Vision Transformer; NDVI, Normalized Difference Vegetation Index; QTL, Quantitative Trait Locus/Loci; RF, Random Forest; RKHS, Reproducing Kernel Hilbert Space; RMSE, Root Mean Square Error; RNN, Recurrent Neural Network; RSFM, Remote Sensing Foundation Model; SIFT, Scale-Invariant Feature Transform; SHAP, SHapley Additive exPlanations; SOS/EOS, Start/End of Season; STSGCN, Spatial-Temporal Synchronous Graph Convolutional Network; Swin Transformer, Shifted Window Transformer; SVM/SVR, Support Vector Machine/Support Vector Regression; TFT, Temporal Fusion Transformer; UAS, Unmanned aerial system; UAV, Unmanned Aerial Vehicle; UAV-RSPs, UAV Remote Sensing Platforms; UN, United Nations; ViT, Vision Transformer.</p>
</fn>
</fn-group>
</back>
</article>