<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" dtd-version="1.3" article-type="research-article">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2026.1750108</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Transformer model to determine spatio-temporal relationships of variables, and interpretability for soybean seed yield, oil, and protein prediction</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Ayanlade</surname> <given-names>Timilehin T.</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Software" vocab-term-identifier="https://credit.niso.org/contributor-roles/software/">Software</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<uri xlink:href="https://loop.frontiersin.org/people/2981363"/>
</contrib>
<contrib contrib-type="author" equal-contrib="yes">
<name><surname>Van der Laan</surname> <given-names>Liza</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="author-notes" rid="fn001"><sup>&#x02020;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Visualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/visualization/">Visualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<uri xlink:href="https://loop.frontiersin.org/people/1622873"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Liu</surname> <given-names>Qisai</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Formal analysis" vocab-term-identifier="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
<uri xlink:href="https://loop.frontiersin.org/people/1887720"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Gangopadhyay</surname> <given-names>Tryambak</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff3"><sup>3</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
</contrib>
<contrib contrib-type="author">
<name><surname>Shook</surname> <given-names>Johnathon</given-names></name>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
</contrib>
<contrib contrib-type="author">
<name><surname>Singh</surname> <given-names>Arti</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<uri xlink:href="https://loop.frontiersin.org/people/224402"/>
</contrib>
<contrib contrib-type="author">
<name><surname>Ganapathysubramanian</surname> <given-names>Baskar</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<uri xlink:href="https://loop.frontiersin.org/people/483726"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Sarkar</surname> <given-names>Soumik</given-names></name>
<xref ref-type="aff" rid="aff1"><sup>1</sup></xref>
<xref ref-type="aff" rid="aff4"><sup>4</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<uri xlink:href="https://loop.frontiersin.org/people/159962"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name><surname>Singh</surname> <given-names>Asheesh K.</given-names></name>
<xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<xref ref-type="corresp" rid="c001"><sup>&#x0002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; review &amp; editing" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing &#x2013; review &#x00026; editing</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<uri xlink:href="https://loop.frontiersin.org/people/433094"/>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Department of Mechanical Engineering, Iowa State University</institution>, <city>Ames, IA</city>, <country country="us">United States</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Agronomy, Iowa State University</institution>, <city>Ames, IA</city>, <country country="us">United States</country></aff>
<aff id="aff3"><label>3</label><institution>Amazon Web Services, Amazon</institution>, <city>Santa Clara, CA</city>, <country country="us">United States</country></aff>
<aff id="aff4"><label>4</label><institution>Department of Computer Science, Iowa State University</institution>, <city>Ames, IA</city>, <country country="us">United States</country></aff>
<author-notes>
<corresp id="c001"><label>&#x0002A;</label>Correspondence: Soumik Sarkar, <email xlink:href="mailto:soumiks@iastate.edu">soumiks@iastate.edu</email>; Asheesh K. Singh, <email xlink:href="mailto:singhak@iastate.edu">singhak@iastate.edu</email></corresp>
<fn fn-type="equal" id="fn001"><label>&#x02020;</label><p>These authors share first authorship</p></fn></author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-26">
<day>26</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>9</volume>
<elocation-id>1750108</elocation-id>
<history>
<date date-type="received">
<day>07</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>02</day>
<month>02</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>03</day>
<month>02</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x000A9; 2026 Ayanlade, Van der Laan, Liu, Gangopadhyay, Shook, Singh, Ganapathysubramanian, Sarkar and Singh.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Ayanlade, Van der Laan, Liu, Gangopadhyay, Shook, Singh, Ganapathysubramanian, Sarkar and Singh</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-26">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Accurate in-season prediction of seed yield and seed composition traits such as oil and protein are useful for gaining accuracy and efficiency in soybean breeding. These predictions can also inform farmers, enabling them to improve their field management practices, and guide their market decisions. We report a Transformer-based deep learning framework built on 30 years of multi-environment performance data from the Northern and Southern Uniform Soybean Tests (UST) across North America. Unlike earlier studies on seed yield, oil and protein prediction that focus on limited years, regions, single modalities, we utilized a comprehensive dataset that includes weather, genotype, and management factors, ensuring a more holistic approach to soybean yield, oil, and protein prediction. Our model integrates multivariate time-series weather data with genotypic relationship information, maturity group, and geographic location, to predict variety performance in diverse environments. Our model captures complex temporal patterns associated with trait variability; showing high predictive accuracy (R2) of 77.6 &#x000B1; 0.2%, 63.9 &#x000B1; 4.7%, and 79.3 &#x000B1; 2.3% for seed yield, oil, and protein, respectively. Additionally, for seed yield, we also evaluated multiple interpretability methods to assess feature importance for predictor variables and critical growing timepoints, and solar radiation and temperature were noted as the key predictors. Overall, these results demonstrate the usefulness of a Transformer-based model in trait predictions, and the utility of large cooperative datasets from breeding programs.</p></abstract>
<kwd-group>
<kwd>interpretability</kwd>
<kwd>seed oil</kwd>
<kwd>seed protein</kwd>
<kwd>soybean</kwd>
<kwd>spatio-temporal analysis</kwd>
<kwd>transformer</kwd>
<kwd>yield prediction</kwd>
</kwd-group>
<funding-group>
  <funding-statement>The author(s) declared that financial support was received for this work and/or its publication. This work was supported by: (a) AI Institute for Resilient Agriculture (USDA-NIFA &#x00023;2021-67021-35329), (b) COALESCE: COntext Aware LEarning for Sustainable CybEr-Agricultural Systems (NSF CPS Frontier &#x00023;1954556), (c) Smart Integrated Farm Network for Rural Agricultural Communities (SIRAC) (NSF S &#x00026; CC &#x00023;1952045), (d) USDA CRIS Project IOW04714, (e) Plant Sciences Institute, (f) Iowa Soybean Association, (g) North Central Soybean Research Program, (h) R. F. Baker Center for Plant Breeding.</funding-statement>
</funding-group>
<counts>
<fig-count count="6"/>
<table-count count="2"/>
<equation-count count="4"/>
<ref-count count="65"/>
<page-count count="12"/>
<word-count count="8657"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>AI in Food, Agriculture and Water</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Predicting cultivar performance with high accuracy and interpretability is essential for optimizing plant breeding strategies and mitigating the risks posed by climate variability. Crop performance, such as yield, is influenced by complex interactions between genetic factors and environmental conditions (GxE), making it challenging to develop reliable predictive models. The prediction of crop performance in untested environments is an enduring challenge in plant breeding, and is an active area of research with numerous approaches being investigated, such as genomic prediction (<xref ref-type="bibr" rid="B34">Meuwissen et al., 2001</xref>; <xref ref-type="bibr" rid="B26">Jannink et al., 2010</xref>; <xref ref-type="bibr" rid="B55">Tolhurst et al., 2022</xref>), phenomic prediction (<xref ref-type="bibr" rid="B38">Parmley et al., 2019</xref>; <xref ref-type="bibr" rid="B63">Winn et al., 2023</xref>; <xref ref-type="bibr" rid="B50">Singh et al., 2021</xref>; <xref ref-type="bibr" rid="B59">Van der Laan et al., 2025</xref>), and crop modeling (<xref ref-type="bibr" rid="B24">Jagtap and Jones, 2002</xref>; <xref ref-type="bibr" rid="B33">Messina et al., 2006</xref>; <xref ref-type="bibr" rid="B43">Sakamoto, 2020</xref>), as well as the integration of these different methods (<xref ref-type="bibr" rid="B53">Technow et al., 2015</xref>; <xref ref-type="bibr" rid="B12">Cooper et al., 2020</xref>; <xref ref-type="bibr" rid="B18">Gonz&#x000E1;lez et al., 2024</xref>; <xref ref-type="bibr" rid="B13">Crossa et al., 2025</xref>). Traditional crop modeling approaches, including process-based and statistical models, have been widely used but often require extensive datasets and struggle to capture the intricate, non-linear relationships between variables (<xref ref-type="bibr" rid="B27">Jones et al., 2003</xref>; <xref ref-type="bibr" rid="B8">Boote et al., 1996</xref>).</p>
<p>Machine learning (ML), particularly deep learning, offers an alternative for yield prediction but has historically faced limitations in interpretability, which is critical for domain experts making breeding and management decisions (<xref ref-type="bibr" rid="B1">Ashfaq et al., 2025</xref>; <xref ref-type="bibr" rid="B32">Mahesh and Soundrapandiyan, 2024</xref>). Additionally, training ML models requires extensive datasets, which are often costly and time-consuming to generate.</p>
<p>Soybean is one of the most grown crops in North America, and breeding efforts have been extensive for the crop. As part of the breeding efforts for North American soybeans, the annual Uniform Soybean Tests (UST) have been coordinated through the United States Department of Agriculture (USDA) since 1941 (<xref ref-type="bibr" rid="B57">USDA-ARS, 2025a</xref>,<xref ref-type="bibr" rid="B58">b</xref>). Starting out as yield trials initially, the UST is a partnership with public breeders in universities and government organizations across the US and Canada. These trials are meant to allow public breeding programs and small private programs to test current and experimental varieties in a greater number of new environments within the intended range of adaptation. This allows public breeding programs to expand the area of testing available with less associated costs, and also serve as a valuable source of historical and current soybean data. Because these data are publicly available, they are an affordable option for research in predicting crop performance (<xref ref-type="bibr" rid="B29">Krause et al., 2023</xref>). Such large collaborative networks involving multiple breeders and researchers generating large multi-year, multi-environment datasets has been utilized for comprehensive studies (<xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>; <xref ref-type="bibr" rid="B62">Wartha et al., 2025</xref>; <xref ref-type="bibr" rid="B11">Chiozza et al., 2025</xref>), and these valuable datasets provide tremendous opportunities for ML-based model development for agriculture and breeding applications.</p>
<p>Recent advances in attention-based models, particularly transformers, have demonstrated significant improvements in time-series prediction by efficiently capturing long-range dependencies and highlighting critical input features (<xref ref-type="bibr" rid="B60">Vaswani et al., 2017</xref>; <xref ref-type="bibr" rid="B49">Singh et al., 2025</xref>). Unlike long short-term memory models (LSTMs) (<xref ref-type="bibr" rid="B21">Hochreiter and Schmidhuber, 1997</xref>), which process sequences sequentially, transformers utilize self-attention mechanisms to weigh the importance of different time steps and variables simultaneously. This makes them particularly well-suited for yield prediction, where weather variability across the growing season plays a crucial role in determining crop performance (<xref ref-type="bibr" rid="B52">Tan et al., 2018</xref>; <xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>). Compared to traditional statistical models and deep learning methods such as LSTMs, transformers offer a more flexible and scalable framework for handling complex, multivariate agricultural datasets. The ability to process entire sequences in parallel allows for efficient learning of patterns across years and locations (<xref ref-type="bibr" rid="B45">Shi et al., 2024</xref>). Furthermore, these models improves interpretability by explicitly highlighting key weather variables and their temporal importance, aiding plant breeders and agronomists in understanding the drivers of yield variability. Prior studies have utilized traditional and deep learning techniques to model temporal dependencies in yield prediction, there is a need to understand the role(s) of environmental factors and time periods that most influence the critical traits (<xref ref-type="bibr" rid="B30">Lobell et al., 2013</xref>), such as soybean seed yield, oil, and protein.</p>
<p>Interpretability plays a vital role in yield prediction by aligning model outputs with agronomic expertise and supporting decision making in breeding programs. This need becomes even more pressing in multi-sensor or multi-source settings, where interpretability has often been cited as a significant gap in current approaches (<xref ref-type="bibr" rid="B3">Ayanlade et al., 2024</xref>; <xref ref-type="bibr" rid="B49">Singh et al., 2025</xref>). Prior studies have explored both intrinsic and <italic>post-hoc</italic> interpretability approaches. Intrinsic methods, such as linear regression and decision trees (<xref ref-type="bibr" rid="B35">Monteiro et al., 2022</xref>; <xref ref-type="bibr" rid="B15">Dwivedi et al., 2023</xref>; <xref ref-type="bibr" rid="B4">Balcan and Sharma, 2024</xref>; <xref ref-type="bibr" rid="B56">Torsoni et al., 2023</xref>), directly expose feature contributions through their model structures, while <italic>post-hoc</italic> methods&#x02014;such as LIME (<xref ref-type="bibr" rid="B41">Ribeiro et al., 2016</xref>), SHAP (<xref ref-type="bibr" rid="B31">Lundberg and Lee, 2017</xref>), Integrated Gradients (<xref ref-type="bibr" rid="B51">Sundararajan et al., 2017</xref>), and DeepLIFT (<xref ref-type="bibr" rid="B48">Shrikumar et al., 2017</xref>) approximate local explanations for complex deep models. In recent years, attention-based architectures have also gained prominence as interpretable modeling strategies for sequential and multimodal data, particularly in crop monitoring and yield prediction tasks where temporal attention highlights critical growth stages or stress periods (<xref ref-type="bibr" rid="B65">Xiong et al., 2024</xref>; <xref ref-type="bibr" rid="B2">Aviles Toledo et al., 2024</xref>; <xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>; <xref ref-type="bibr" rid="B61">Wang et al., 2024</xref>). However, both intrinsic and <italic>post-hoc</italic> methods often lack standardized measures of faithfulness, raising concerns about whether their explanations truly reflect the model&#x00027;s decision process rather than coincidental correlations. To address this limitation, evaluation pipelines based on quantitative faithfulness metrics, such as comprehensiveness and sufficiency (<xref ref-type="bibr" rid="B14">DeYoung et al., 2019</xref>), and their regression-oriented extension AOPCR (<xref ref-type="bibr" rid="B37">Ozyegen et al., 2022</xref>) have been established to assess how model predictions respond to systematic perturbations of the most relevant features. These frameworks enable rigorous evaluation of interpretability methods across domains, providing deeper insight into the reliability of attention- and attribution-based explanations in multi-source agronomic modeling.</p>
<p>In this study, we propose a Transformer-based deep learning model for the prediction of soybean seed yield, oil and protein using historical data from the Uniform Soybean Trials (UST), encompassing the Northern and Southern trials. Our model integrates multidecadal weather records with pedigree information, leveraging recent advances in sequence modeling to improve both predictive accuracy and model interpretability. In addition, by leveraging the self-attention mechanism of Transformer architectures, we identify the most influential time periods and variables during the growing season, providing valuable insights into how environmental factors impact yield, oil and protein levels, offering a valuable tool to potentially mitigate risks and improve agricultural and breeding decision-making. These insights will inform breeding strategies aimed at developing resilient cultivars, ultimately contributing to global food security (<xref ref-type="bibr" rid="B42">Rose et al., 2016</xref>). Previous studies that applied attention-based mechanisms to yield prediction (<xref ref-type="bibr" rid="B65">Xiong et al., 2024</xref>; <xref ref-type="bibr" rid="B2">Aviles Toledo et al., 2024</xref>; <xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>), primarily relied on LSTMs and meteorological data without integrating genetic information and crop management practices including pre-planting information. Our work bridges this gap by incorporating a comprehensive dataset including weather, genotype, and management factors, with Transformer-based modeling ensuring a more holistic approach to soybean yield, oil, and protein prediction.</p>
</sec>
<sec sec-type="materials|methods" id="s2">
<label>2</label>
<title>Materials and methods</title>
<sec>
<label>2.1</label>
<title>Data preparation and processing</title>
<sec>
<label>2.1.1</label>
<title>Performance records</title>
<p>PDF files from the 1989-2018 Uniform Soybean Tests (USTs) (<xref ref-type="bibr" rid="B57">USDA-ARS, 2025a</xref>,<xref ref-type="bibr" rid="B58">b</xref>) were downloaded. Following the protocol outlined in <xref ref-type="bibr" rid="B47">Shook et al. (2021)</xref>, all PDFs were converted to .xlsx files. Briefly, the on-line utility Zamzar (<ext-link ext-link-type="uri" xlink:href="https://zamzar.com">zamzar.com</ext-link>) for conversions to .xlsx files. Each tab of the resulting .xlsx files correspond to a single page from the original PDF file. This methodology allows for the recovery of the PDF tables with no need for error-prone human translation. Random checking of files for errors was conducted to ensure the accuracy of the data, and no instance of error was noticed. Once in .xlsx files, the tables were manually sorted to align the performance records for a given genotype-environment combination into a single row. Records with no yield data were removed from the file to ensure data integrity. To standardize the dataset, performance records were curated such that each genotype-location-year combination was consolidated into a single row. Following data cleaning, the final dataset comprised 103,955 performance records over 29 years, with 11,506 unique genotypes. Each performance record contains 214 days of multivariate time-series data, corresponding to the growing season within the year, with each record characterized by the maturity group variables.</p>
</sec>
<sec>
<label>2.1.2</label>
<title>Acquisition of weather records</title>
<p>Daily weather records for all year/location combinations were compiled from NASA Power (<xref ref-type="bibr" rid="B36">NASA, 2025</xref>) and using the closest grid point from the 1/2&#x000B0; &#x000D7; 5/8&#x000B0; global grid utilized by the MERRA-2 model implemented within NASA Power. The weather data included the maximum, minimum, and average air temperature (&#x000B0;C), relative humidity (percentage), average precipitation (mm day<sup>&#x02212;1</sup>), and total solar irradiance (MJ/m<sup>2</sup>). For model training, we downsampled the dataset to extract maximum, minimum, and average weather conditions over different time intervals throughout the growing season. This processed weather information was then concatenated with the performance records to create a complete dataset for prediction. Notably, the available performance records for oil and protein had missing records for the combined data (49.3% of total records missing), and motivated our exploration of transfer learning strategies. The missing records for oil and protein were environment specific, and are due to only a subset of environments in the UST network providing these trait data.</p>
</sec>
<sec>
<label>2.1.3</label>
<title>Genotype clustering</title>
<p>Genotype identity was preserved in the performance records so that it could be applied to the model and allow for prediction of specific genotypes. The UST program was used so that experimental genotypes can be tested on a large scale. Because of this, most of the genotypes are experimental lines which do not have molecular marker data available. Additionally, this dataset extends into a time when genotyping was cost prohibitive for many programs, further limiting the availability of molecular marker data from the older genotypes. While there are recent efforts to genotypes all lines, gaps still persists. Additionally, we wanted to explore the usefulness for crops or in situation where marker data are not available. This prevents the use of a G matrix for use in genotype clustering. As an alternative to these restrictions, we developed a connected pedigree for all genotypes that had publicly available pedigree information. This allowed for the creation of an A matrix, using the pedigree package in R. The A matrix then resulted in the formation of a 11,506 &#x000D7; 11,506 kinship matrix. The kinship matrix and genotype program of origin were used together to increase control over relatedness. This allowed for genotype clustering, which was found to improve model performance in <xref ref-type="bibr" rid="B47">Shook et al. (2021)</xref>.</p>
<p>To incorporate genotype-specific variation into the model, we applied clustering techniques based on the A matrix of kinship relationships and mean location yield across genotypes. Using the K-means clustering approach, we determined that the optimal number of clusters was 40. Details are provided in <xref ref-type="supplementary-material" rid="SM1">Supplementary material 1</xref>. This allowed genotypes to be assigned to clusters based on the correlation matrix, effectively grouping lines with similar genetic backgrounds. K-means clustering partitions data into a predefined number of clusters (n) by iteratively minimizing intra-cluster variance. Specifically, the algorithm identifies n cluster centroids and assigns each genotype to the nearest centroid, minimizing the within-cluster sum-of-squares (inertia). This method has demonstrated effectiveness in handling large datasets and is commonly applied across various scientific domains (<xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>). Each genotype was assigned a fixed cluster ID, which was subsequently incorporated into our model as an additional feature to enable the model capture genetic variation for yield prediction. All clustering analyses were implemented in Python 3.8 using the scikit-learn (v1.2) library, with NumPy, Pandas, and SciPy for data preprocessing and matrix computation.</p>
</sec>
</sec>
<sec>
<label>2.2</label>
<title>Model</title>
<p>Our approach employs a transformer encoder model for yield prediction, leveraging self-attention mechanisms to capture dependencies across time steps in parallel. The prediction task is framed as a many-to-one regression problem where the input consists of time-series weather data alongside genotype and agronomic variables. Each training sample comprises a multivariate time series, containing weather-related variables and other context variables. To structure the input for the transformer model, we constructed tokens by aggregating daily weather variables into multiple day segments (see <xref ref-type="fig" rid="F1">Figure 1</xref>), forming input tokens. The initial tokens for each variable were padded with zeros where necessary to maintain uniform sequence lengths. Additionally, the context tokens (genotype cluster, maturity group, and location&#x02014;latitude and longitude) were also tokenized individually and added to the sequence.</p>
<fig position="float" id="F1">
<label>Figure 1</label>
<caption><p>Architecture of the transformer model used for multi-trait prediction in soybean. Input features, including weather variables, genotype, maturity group, and location information, are segmented and embedded with patch and positional encodings. These are passed through a transformer encoder (4 encoder blocks) consisting of stacked layers of multi-head self-attention and feedforward neural networks. The model outputs trait predictions (yield, oil, and protein) through a regression head.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0001.tif">
<alt-text content-type="machine-generated">Flowchart diagram illustrating a transformer-based model for regression. Inputs include precipitation, air temperature, relative humidity, solar irradiance, genotype, location, and maturity group, which are converted into patch and position embeddings, processed by a transformer encoder with multi-head attention, and output regression predictions for yield, oil, and protein.</alt-text>
</graphic>
</fig>
<p>As shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, each input input sequence</p>
<disp-formula id="E1"><mml:math id="M1"><mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mo>=</mml:mo><mml:mo stretchy='false'>[</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>h</mml:mi><mml:mi>u</mml:mi><mml:mi>m</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>h</mml:mi><mml:mi>u</mml:mi><mml:mi>m</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mi>p</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>t</mml:mi><mml:mi>e</mml:mi><mml:mi>m</mml:mi><mml:mi>p</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>r</mml:mi><mml:mi>a</mml:mi><mml:mi>d</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mtext>&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;&#x02009;</mml:mtext><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>v</mml:mi><mml:mi>a</mml:mi><mml:mi>r</mml:mi><mml:mo>*</mml:mo><mml:mi>t</mml:mi><mml:mi>p</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>l</mml:mi><mml:mi>o</mml:mi><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>g</mml:mi><mml:mi>e</mml:mi><mml:mi>n</mml:mi><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mrow><mml:mi>m</mml:mi><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>]</mml:mo><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mstyle mathvariant='bold' mathsize='normal'><mml:mi>x</mml:mi></mml:mstyle><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mi>&#x0211D;</mml:mi><mml:mrow><mml:mi>T</mml:mi><mml:mo>&#x000D7;</mml:mo><mml:mi>d</mml:mi></mml:mrow></mml:msup></mml:mtd></mml:mtr></mml:mtable></mml:math></disp-formula>
<p>where <italic>T</italic> is the number of tokens, <italic>d</italic> is the dimensionality, <italic>hum</italic> represents humidity, <italic>temp</italic> represents the temperature related variables, <italic>rad</italic> represents radiation, <italic>var</italic> represents each weather variable, <italic>tp</italic> represents the token timepoint, <italic>loc</italic> represents latitude and longitude token, <italic>geno</italic> represents genotype, and <italic>mg</italic> represents the maturity group. Each token is embedded and positionally encoded:</p>
<disp-formula id="E2"><mml:math id="M2"><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>=</mml:mo><mml:mstyle mathvariant="bold"><mml:mtext>E</mml:mtext></mml:mstyle><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>x</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:math></disp-formula>
<p>where <bold>E</bold> &#x02208; &#x0211D;<sup><italic>d</italic>&#x000D7;<italic>d</italic><sup>&#x02032;</sup></sup> is a learnable projection matrix, and <inline-formula><mml:math id="M3"><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>p</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x02208;</mml:mo><mml:msup><mml:mrow><mml:mi>&#x0211D;</mml:mi></mml:mrow><mml:mrow><mml:msup><mml:mrow><mml:mi>d</mml:mi></mml:mrow><mml:mrow><mml:mi>&#x02032;</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:msup></mml:math></inline-formula> is the positional encoding for token <italic>i</italic>.</p>
<p>Similar to (<xref ref-type="bibr" rid="B64">Woo et al., 2024</xref>), the embedded sequence for all variables <inline-formula><mml:math id="M4"><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup><mml:mo>=</mml:mo><mml:mrow><mml:mo stretchy="false">[</mml:mo><mml:mrow><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup><mml:mo>,</mml:mo><mml:mo>&#x02026;</mml:mo><mml:mo>,</mml:mo><mml:msubsup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mi>T</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mn>0</mml:mn></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msubsup></mml:mrow><mml:mo stretchy="false">]</mml:mo></mml:mrow></mml:math></inline-formula> is concatenated and passed through a unified encoder consisting of <italic>N</italic> stacked transformer encoder layers. Each layer consists of a multi-head self-attention module followed by a feedforward network.</p>
<p>The final token representations <bold>z</bold><sup>(<italic>N</italic>)</sup> &#x02208; &#x0211D;<sup><italic>T</italic>&#x000D7;<italic>d</italic><sup>&#x02032;</sup></sup> are flattened and passed to a regression head:</p>
<disp-formula id="E3"><mml:math id="M5"><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>W</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">reg</mml:mtext></mml:mrow></mml:msub><mml:mo>&#x000B7;</mml:mo><mml:mtext class="textrm" mathvariant="normal">concat</mml:mtext><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>z</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mrow><mml:mi>N</mml:mi></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:mrow></mml:msup></mml:mrow><mml:mo stretchy="false">)</mml:mo></mml:mrow><mml:mo>&#x0002B;</mml:mo><mml:msub><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>b</mml:mtext></mml:mstyle></mml:mrow><mml:mrow><mml:mtext class="textrm" mathvariant="normal">reg</mml:mtext></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:mtext>&#x02003;</mml:mtext><mml:mover accent="true"><mml:mrow><mml:mstyle mathvariant="bold"><mml:mtext>y</mml:mtext></mml:mstyle></mml:mrow><mml:mo>^</mml:mo></mml:mover><mml:mo>&#x02208;</mml:mo><mml:mi>&#x0211D;</mml:mi></mml:mrow></mml:math></disp-formula>
<p>The model is trained using mean squared error (MSE) loss</p>
<disp-formula id="E4"><mml:math id="M6"><mml:mrow><mml:mi>L</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi></mml:mrow></mml:mfrac><mml:mrow><mml:mo>&#x02225;</mml:mo><mml:mi>&#x00177;</mml:mi><mml:mo>-</mml:mo><mml:mi>y</mml:mi><mml:msup><mml:mrow><mml:mo>&#x02225;</mml:mo></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:mrow></mml:mrow></mml:math></disp-formula>
<p>where <italic>y</italic> is the ground truth value for desired trait and <italic>n</italic> is the number of samples.</p>
<p>For yield prediction, we trained the model end-to-end using a regression head applied to the flattened encoder outputs. In contrast, oil and protein prediction tasks had fewer available training samples. To address this data limitation, we adopted a transfer learning approach by reusing the embedding and transformer encoder pre-trained on the yield task. Specifically, we evaluated two strategies: (i) end-to-end fine-tuning of the entire model on the limited oil and protein data, and (ii) freezing the pre-trained encoder weights and training only a new regression head, each for oil and protein. This setup enabled us to assess the effectiveness of transferring temporal and environmental representations learned from the yield task to lesser data traits, seed oil, and protein.</p>
<p>All analyses were implemented using PyTorch (v1.12) and Python 3.8, with dependencies including NumPy, Pandas, SciPy, scikit-learn, and PyYAML for data preprocessing and configuration handling. Experiments were conducted on the Iowa State University Nova High-Performance Computing (HPC) cluster, running Linux (CentOS 9 / Kernel 5.14.0) with job scheduling via Slurm. Each job was executed on a single NVIDIA A100-SXM4 GPU (80 GB memory, 6912 CUDA cores, Ampere architecture) and utilized 16 CPU threads out of a total of 64 physical (128 logical) cores available per node. All training was performed using mixed-precision computation (FP16) enabled through PyTorch AMP. Model training was managed and tracked using the Weights and Biases experiment manager, with random seeds fixed to five values (0&#x02013;4) for reproducibility. Average wall-clock training time per model was approximately 200 min for pre-training and the same for finetuning.</p>
</sec>
<sec>
<label>2.3</label>
<title>Baseline models</title>
<p>To assess the effectiveness of the transformer model, we compared its performance against several baseline models: Support Vector Regression with Radial Basis Function Kernel (SVR-RBF), Least Absolute Shrinkage and Selection Operator (LASSO) regression, Bidirectional Long Short-Term Memory (BLSTM), and Bidirectional Recurrent Neural Networks (BRNN). Each baseline was trained using hyperparameter optimization to ensure fair comparison.</p>
</sec>
<sec>
<label>2.4</label>
<title>Model training and evaluation</title>
<p>The models were optimized using the Adam optimizer. The dataset was randomly split into 80% training, 10% validation, and 10% test sets. To prevent overfitting, a dropout of 0.1 was applied within the transformer layers. The model was trained for a maximum of 100 epochs with a batch size of 16. We include early stopping with a patience of 20 epoch to prevent overfitting and to ensure that the model is well trained. Each model was trained with a learning rate of 5 &#x000D7; 10<sup>&#x02212;6</sup>, and cosine learning rate adjustment strategy.</p>
<p>Model performance was evaluated using three standard regression metrics: root mean squared error (RMSE), mean absolute error (MAE), and coefficient of determination (R-squared, <italic>R</italic><sup>2</sup>) score.</p>
</sec>
<sec>
<label>2.5</label>
<title>Model interpretability</title>
<p>To analyze model interpretability, we employed <italic>post-hoc</italic> attention-based and attribution-based techniques on the trained transformer model. First, the self-attention maps were extracted from each encoder layer and head for all test set samples. Attention weights were normalized and mean aggregated across heads and layers in the encoder block to compute importance score. By aggregating attention scores across samples and across layers, we identified weather patterns that strongly influenced trait predictions. The attention scores were extracted from the trained model and visualized to assess their alignment with established domain principles in crop production. We compare the performance of the attention-based score with gradient-based attribution methods such as Integrated Gradients (<xref ref-type="bibr" rid="B51">Sundararajan et al., 2017</xref>), DeepLIFT (<xref ref-type="bibr" rid="B48">Shrikumar et al., 2017</xref>), Gradient SHAP (<xref ref-type="bibr" rid="B31">Lundberg and Lee, 2017</xref>), and LIME (<xref ref-type="bibr" rid="B41">Ribeiro et al., 2016</xref>). These methods quantify the marginal contribution of each input token to the model&#x00027;s predictions by computing local gradients or perturbation responses.</p>
</sec>
<sec>
<label>2.6</label>
<title>Methodology for faithfulness evaluation of interpretations</title>
<p>Evaluating interpretability in the absence of ground-truth explanations requires a quantitative framework that measures the internal consistency between the model&#x00027;s behavior and the importance assigned to its input features. Following the approach of <xref ref-type="bibr" rid="B37">Ozyegen et al. (2022)</xref>, <xref ref-type="bibr" rid="B14">DeYoung et al. (2019)</xref>, <xref ref-type="bibr" rid="B23">Islam et al. (2024)</xref>, we assess local interpretability using perturbation-based metrics that quantify the degree to which identified important tokens truly influence model predictions. For each interpretability method, the token relevance scores produced by the explanation algorithm were normalized and ranked in descending order to determine the most informative input positions. A fixed set of perturbation levels <italic>k</italic> &#x02208; {5, 7.5, 10, 15, 25} was used to define the top percentage of relevant tokens.</p>
<p>For each perturbation level, two complementary evaluations were performed. In the comprehensiveness setting, the top-<italic>k</italic> most relevant tokens were masked from the input while keeping all other tokens unchanged, and the model&#x00027;s new output was compared to the original prediction. The intuition is that if the explanation method correctly identifies truly important tokens, removing them should substantially degrade the model&#x00027;s confidence or alter its prediction. In contrast, the sufficiency setting retains only the top-<italic>k</italic> tokens and masks the rest, measuring whether the selected subset alone contains enough information to reconstruct the model&#x00027;s original output. A smaller difference in this case implies higher sufficiency.</p>
<p>The difference between the model&#x00027;s unperturbed and perturbed predictions was measured using both mean absolute error (MAE) and mean squared error (MSE), averaged over all test samples. The comprehensiveness and sufficiency losses were then aggregated across all perturbation levels to compute the Area Over the Perturbation Curve for Regression (AOPCR). This metric integrates the change in model performance as progressively larger fractions of the most relevant tokens are removed or retained. A higher AOPCR for comprehensiveness indicates that masking top-ranked tokens leads to stronger degradation in predictions, suggesting higher faithfulness, whereas a lower AOPCR for sufficiency implies that the explanation method successfully identifies a minimal subset of features sufficient to reproduce the model&#x00027;s decision.</p>
</sec>
</sec>
<sec sec-type="results" id="s3">
<label>3</label>
<title>Results</title>
<sec>
<label>3.1</label>
<title>Evaluation of model performance</title>
<sec>
<label>3.1.1</label>
<title>Yield prediction</title>
<p>We evaluated the performance of five models using weather variables as inputs, with results summarized in <xref ref-type="table" rid="T1">Table 1</xref>. Among traditional machine learning approaches, SVR-RBF outperformed LASSO. However, both were outperformed by deep learning models. Within the recurrent neural network family, the BLSTM achieved better results compared to the BRNN, though the differences in performance were marginal. The best performance was observed with Transformer-based models. Our proposed transformer model achieved the highest R<sup>2</sup> score (0.776 &#x000B1; 0.002), with the lowest RMSE and MAE across all tested models. This show the effectiveness of attention-based architectures to model complex temporal patterns in weather driven yield prediction, largely capturing the performance differences in performance between environments and years.</p>
<table-wrap position="float" id="T1">
<label>Table 1</label>
<caption><p>Performance of different models for yield prediction.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Model</bold></th>
<th valign="top" align="center"><bold>R<sup>2</sup> score</bold></th>
<th valign="top" align="center"><bold>RMSE</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">LASSO</td>
<td valign="top" align="center">0.732 &#x000B1; 0.000</td>
<td valign="top" align="center">7.692 &#x000B1; 0.000</td>
<td valign="top" align="center">5.832 &#x000B1; 0.000</td>
</tr>
<tr>
<td valign="top" align="left">SVR-RBF</td>
<td valign="top" align="center">0.757 &#x000B1; 0.000</td>
<td valign="top" align="center">7.319 &#x000B1; 0.000</td>
<td valign="top" align="center">5.478 &#x000B1; 0.000</td>
</tr>
<tr>
<td valign="top" align="left">BRNN</td>
<td valign="top" align="center">0.752 &#x000B1; 0.011</td>
<td valign="top" align="center">7.393 &#x000B1; 0.020</td>
<td valign="top" align="center">5.582 &#x000B1; 0.154</td>
</tr>
<tr>
<td valign="top" align="left">BLSTM</td>
<td valign="top" align="center">0.754 &#x000B1; 0.034</td>
<td valign="top" align="center">7.370 &#x000B1; 0.147</td>
<td valign="top" align="center">5.562 &#x000B1; 0.023</td>
</tr>
<tr>
<td valign="top" align="left">Transformer</td>
<td valign="top" align="center"><bold>0.776</bold> <bold>&#x000B1;0.002</bold></td>
<td valign="top" align="center"><bold>7.029</bold> <bold>&#x000B1;0.026</bold></td>
<td valign="top" align="center"><bold>5.317</bold> <bold>&#x000B1;0.019</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Bold values indicate the best performance for each metric.</p>
</table-wrap-foot>
</table-wrap>
</sec>
<sec>
<label>3.1.2</label>
<title>Oil and protein prediction</title>
<p>To assess the generalizability of representations learned during yield prediction, we evaluated two transfer learning strategies for oil and protein prediction, where we compared them to a baseline direct finetuning (vanilla) without pre-training (see Materials and Methods Section 2.2). First, we explore an end-to-end fine-tuning of a yield pre-trained Transformer model on oil and protein, and second, a variation where the encoder of the model pre-trained on yield is frozen, while training only the regression head on the target oil and protein trait. The results across five runs are shown in <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
<fig position="float" id="F2">
<label>Figure 2</label>
<caption><p>Comparison of performance of encoder training strategies (vanilla vs. frozen vs. fine-tuned) for oil and protein prediction. Bars show mean performance across five folds, with error bars representing standard deviation. Transfer learning strategies from the encoder pre-trained on yield consistently outperform direct finetuning for both traits, suggesting generalizability of the learned representations from yield. RMSE and MAE followed similar trends and are provided in the <xref ref-type="supplementary-material" rid="SM1">Supplementary material 11</xref>.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0002.tif">
<alt-text content-type="machine-generated">Bar chart comparing R-squared values for predicting oil and protein using three models: Vanilla (green), Frozen Encoder (blue), and Fine-Tuned Encoder (orange). Fine-Tuned and Frozen Encoders outperform Vanilla for both oil and protein. Error bars indicate standard deviation.</alt-text>
</graphic>
</fig>
<p>For oil prediction, an end to end finetuning of the model resulted in the best performance, achieving a R<sup>2</sup> score (63.9 &#x000B1; 4.7%), with the freezing based transfer strategy achieving comparable performance. A similar trend was seen for protein, where full finetuning the encoder led to a notable gain in R<sup>2</sup> (79.3 &#x000B1; 2.3%), in comparison to a non pre-trained finetuning, and both models saw reduction in RMSE and MAE (see <xref ref-type="supplementary-material" rid="SM1">Supplementary material 5</xref>). These findings suggest that the pre-trained model had already captured generalizable temporal patterns relevant to other seed traits. End-to-end fine-tuning, while more flexible, may have suffered given the relatively limited dataset size. The results reinforce the value of leveraging pre-trained temporal encoders for downstream trait prediction, particularly when labeled data are limited.</p>
</sec>
</sec>
<sec>
<label>3.2</label>
<title>Impact of temporal resolution on yield modeling</title>
<p>To determine the optimal temporal resolution for encoding weather inputs, we evaluated model performance using different sampling intervals, by varying the number of days each input token represents. Prediction performance varied with the size of the temporal window, with the highest R<sup>2</sup> observed at intermediate resolutions, a peak of 0.776 at a 14-day/tp sampling interval across all tasks (<xref ref-type="fig" rid="F3">Figure 3</xref>).</p>
<fig position="float" id="F3">
<label>Figure 3</label>
<caption><p>Effect of varying weather data sampling intervals on <bold>(A)</bold> yield <bold>(B)</bold> oil <bold>(C)</bold> protein <bold>(D)</bold> training time. Shaded regions show standard deviation across five runs.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0003.tif">
<alt-text content-type="machine-generated">Four-panel graphic showing data trends for token sizes over days. Panel A: Line graph in green for Yield, showing R squared values increasing up to 14 days, then plateauing or slightly decreasing with larger token sizes. Panel B: Blue line graph for Oil, displaying fluctuating R squared values across token sizes. Panel C: Orange line graph for Protein, with R squared peaking at 7&#x02013;14 days then declining as token size increases. Panel D: Log-scale line graph with three lines (Yield in green, Protein in orange, Oil in blue), showing time per epoch sharply decreasing up to 36 days and stabilizing afterward. Panels are labeled A, B, C, and D.</alt-text>
</graphic>
</fig>
<p>Very fine grained sampling (e.g., daily tokens) led to slightly lower performance, likely due to increased sequence length, redundancy and extrapolation to the embedding dimension, which may complicate learning long-term dependencies. In contrast, overly coarse resolutions reduced temporal granularity, potentially omitting meaningful short-term weather fluctuations relevant to yield outcomes. The best performance across our test set was achieved using a 14 days per token, suggesting a balance between input compression and temporal fidelity.</p>
</sec>
<sec>
<label>3.3</label>
<title>Interpretability of yield model</title>
<p>In addition to yield prediction, our model provided insights about how much focus the model put on predictive variables for yield prediction across our test dataset. Following the methods described in Section 2.6, we apply various interpretability methods on our data using the trained model version on the 14-day/token sampling interval as it led to the best performance (see <xref ref-type="fig" rid="F3">Figure 3</xref>), to investigate which methods show the best faithfulness in terms of comprehensiveness and sufficiency.</p>
<p>The attention-based method had the best faithfulness across both metrics, with the most comprehensiveness score and least sufficiency score (<xref ref-type="table" rid="T2">Table 2</xref>). Metrics for random samples in the Test set are provided in <xref ref-type="supplementary-material" rid="SM1">Supplementary material 8</xref>. We used the attention-based method for subsequent analysis in the following sections, as the overall results and randomly test samples consistently demonstrated that attention was among the top two methods.</p>
<table-wrap position="float" id="T2">
<label>Table 2</label>
<caption><p>AOPCR-based faithfulness evaluation of interpretation methods averaged across test set.</p></caption>
<table frame="box" rules="all">
<thead>
<tr>
<th valign="top" align="left"><bold>Method</bold></th>
<th valign="top" align="center" colspan="2">Comprehensiveness (&#x02191;)</th>
<th valign="top" align="center" colspan="2">Sufficiency (&#x02193;)</th>
</tr>
<tr>
<th/>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>MSE</bold></th>
<th valign="top" align="center"><bold>MAE</bold></th>
<th valign="top" align="center"><bold>MSE</bold></th>
</tr>
</thead>
<tbody>
<tr>
<td valign="top" align="left">Gradient SHAP</td>
<td valign="top" align="center">0.29</td>
<td valign="top" align="center">3.49</td>
<td valign="top" align="center">0.36</td>
<td valign="top" align="center">5.13</td>
</tr>
<tr>
<td valign="top" align="left">Integrated Gradients</td>
<td valign="top" align="center">0.31</td>
<td valign="top" align="center">4.10</td>
<td valign="top" align="center">0.33</td>
<td valign="top" align="center">4.48</td>
</tr>
<tr>
<td valign="top" align="left">LIME</td>
<td valign="top" align="center">0.05</td>
<td valign="top" align="center">0.17</td>
<td valign="top" align="center">0.44</td>
<td valign="top" align="center">7.76</td>
</tr>
<tr>
<td valign="top" align="left">DeepLIFT</td>
<td valign="top" align="center">0.35</td>
<td valign="top" align="center">4.80</td>
<td valign="top" align="center">0.36</td>
<td valign="top" align="center">5.28</td>
</tr>
<tr>
<td valign="top" align="left">Attention</td>
<td valign="top" align="center"><bold>0.39</bold></td>
<td valign="top" align="center"><bold>6.12</bold></td>
<td valign="top" align="center"><bold>0.28</bold></td>
<td valign="top" align="center"><bold>3.53</bold></td>
</tr></tbody>
</table>
<table-wrap-foot>
<p>Higher values (&#x02191;) indicate better comprehensiveness; lower values (&#x02193;) indicate better sufficiency. Best results for each metric are shown in bold.</p>
</table-wrap-foot>
</table-wrap>
<p>Next, we analyzed the attention weights from the trained transformer model. This attention-based interpretability framework highlights tokens the model deems most relevant, based on the relationship to other tokens providing insight into genotype-environment interactions and the temporal dynamics underlying yield variability. <xref ref-type="fig" rid="F4">Figure 4</xref> illustrates the average attention weights assigned to individual weather variables throughout the growing season in our 14-day/tp sampling configuration, averaged across all samples in the test set. Notably, across all sampling intervals, we see that maximum temperature and solar radiation tokens received consistently high attention weights, aligning with established agronomic understanding of their influence on yield formation. The trend for the other sampling intervals across the test set and for two examples for random time-series samples in our test dataset are provided in <xref ref-type="supplementary-material" rid="SM1">Supplementary materials 9, 10</xref>, respectively. In our higher sampling fidelity models with 1-day/tp, we observe that maximum temperature tokens had the highest attention score (see <xref ref-type="supplementary-material" rid="SM1">Supplementary material 9</xref>).</p>
<fig position="float" id="F4">
<label>Figure 4</label>
<caption><p>Average attention weights of yield predictive variables for the test set across five experimental runs for the 14 days per timepoint sampling intervals. Line plots represent the attention of time-varying variables over the growing season, while dashed horizontal lines represent static context variables such as maturity group (MG), location (Loc), and genotypic cluster (Geno).</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0004.tif">
<alt-text content-type="machine-generated">Line graph showing attention weights for nine features over sixteen time steps, each representing fourteen days. Precipitation has the highest weights, peaking near the midpoint, followed by MaxTemp and other variables. Static features of MG, Loc., and Geno. are represented as horizontal dashed lines, with MG and Loc. around 0.05 and Geno. near 0.06. The y-axis ranges from zero to 0.08, and the legend identifies all variables measured.</alt-text>
</graphic>
</fig>
<p>We also explored the quantification of the relative importance of dynamic weather features versus static context tokens, where we aggregated all time varying weather inputs into a single token and compared its average attention weight against the static variables, MG, genotype cluster, and location (see <xref ref-type="fig" rid="F5">Figure 5</xref>). The composite weather token received the highest attention overall, with larger weights than any individual context. To verify that spatio-temporal features contribute meaningfully to prediction performance, we trained multiple model variants using (i) genetic features only, (ii) genetics combined with weather, and (iii) genetics combined with weather and management features. We observe a consistent improvement in performance as additional information is incorporated to the genetic features (see <xref ref-type="supplementary-material" rid="SM1">Supplementary material 12</xref>). This reinforces the central role of in season weather dynamics in yield determination, though also shows that context features such as genotype cluster still contributed meaningfully to the prediction.</p>
<fig position="float" id="F5">
<label>Figure 5</label>
<caption><p>Average attention weights of yield predictive variables using transformer model. The weather bar represent the attention weight on the token when all weather variables (time-varying variables over the growing season) are treated as a single unit. The error bars show the standard error around the mean weights.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0005.tif">
<alt-text content-type="machine-generated">Bar chart showing attention weights for four variables: Weather, MG, Location, and Geno. Weather has the highest weight near 0.4, followed by MG, Location, and Geno, each around 0.2. Error bars indicate standard error.</alt-text>
</graphic>
</fig>
<p>To further investigate how the transformer model captured the variable interactions, we analyzed the distribution of attention weights when training and testing across different maturity groups (MGs). <xref ref-type="fig" rid="F6">Figure 6</xref> illustrates the averaged attention weights for each MG during the growing season. The radiation related tokens were observed to be have the highest attention weights across all MGs. With maximum temperature coming second in earlier-maturing groups (e.g., MG 1&#x02013;3) assigning greater attention to early-season radiation and temperature, whereas later-maturing groups (MG 7&#x02013;9) emphasized a mixture of other variables.</p>
<fig position="float" id="F6">
<label>Figure 6</label>
<caption><p>Attention weight across soybean maturity groups (MGs). Radiation related tokens were observed to have the highest values across all MGs, but the next attended tokens differing for the early vs. late MGs, where early MGs emphasized temperature.</p></caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="frai-09-1750108-g0006.tif">
<alt-text content-type="machine-generated">Nine-panel line chart compares nine groups labeled MG 1 to MG 9, plotting variables including maximum temperature, minimum temperature, average temperature, average humidity, precipitation, radiation, MG, location, and genotype across sixteen intervals. Each group displays similar trends, with radiation variable consistently peaking above others in all panels.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec sec-type="discussion" id="s4">
<label>4</label>
<title>Discussion</title>
<p>This study presents a comprehensive multimodal frameworks for soybean trait prediction that integrates genotypic relatedness, maturity group, and weather features across multiple locations and years in North America. Previous studies have largely relied on narrow environmental ranges with limited years of data, limited number of cultivars, or single modal data. In contrast, our dataset spans diverse maturity zones and climatic conditions encompassing the soybean growing region of North America, and includes thousands of distinct soybean varieties, which is representative of genetic diversity in soybean breeding programs. The transformer-based architecture leveraged the state of the art attention mechanism, capturing temporal dependencies, while simultaneously enabling cross-modal feature learning in this extensive dataset spanning decades, locations, growing conditions and genotypes.</p>
<p>Earlier research in yield prediction primarily focused on meteorological drivers and vegetation indices derived from remote sensing, often using recurrent neural networks or regression-based ensemble models (<xref ref-type="bibr" rid="B65">Xiong et al., 2024</xref>; <xref ref-type="bibr" rid="B44">Schwalbert et al., 2020</xref>; <xref ref-type="bibr" rid="B47">Shook et al., 2021</xref>; <xref ref-type="bibr" rid="B5">Barbosa Dos Santos et al., 2022</xref>; <xref ref-type="bibr" rid="B56">Torsoni et al., 2023</xref>; <xref ref-type="bibr" rid="B2">Aviles Toledo et al., 2024</xref>; <xref ref-type="bibr" rid="B17">Eugenio et al., 2020</xref>). These approaches serve a useful purpose; however, they are less generalizable and lack explainability. Studies have worked on predictions for oil and protein in addition to yield, although these have often been in the context for whole field predictions of a single cultivar (<xref ref-type="bibr" rid="B20">Hernandez et al., 2023</xref>; <xref ref-type="bibr" rid="B39">Pereyra et al., 2026</xref>), and not the performance of different cultivars as in plant breeding. While these previous plant breeding related approaches have achieved moderate success for yield estimation, their extension to seed composition traits such as oil and protein has been more limited due to data scarcity, and imbalance across environments. Herein, the cooperatively run Northern and Southern Uniform Soybean Test are an excellent community resource, and critical for ML models requiring large datasets, such as Transformers.</p>
<p>The transformer framework presented here captures long-range dependencies among environmental, genetic, and management variables, offering a unified architecture that can be adapted to multiple traits. The observed improvements in predictive accuracy show the importance of explicitly modeling spatio-temporal interactions in the data. Crop development and yield formation are inherently time dependent, with physiological processes responding dynamically to environmental and management changes (<xref ref-type="bibr" rid="B22">Hodges and French, 1985</xref>; <xref ref-type="bibr" rid="B10">Chen and Wiatrak, 2010</xref>). The transformer&#x00027;s temporal attention enables the model to modify weights on critical growth periods, where weather and soil interactions most strongly influence final outcomes. While our models did not utilize soil features as they were unavailable, future studies are needed to incorporate soil variables in yield prediction (<xref ref-type="bibr" rid="B9">Carroll et al., 2024</xref>).</p>
<p>Computational approaches, i.e., Transformer-based models, highlighted different environmental variables carrying higher attention weight in predicting yield. For example, both Northern and Southern latitudes had solar radiation and temperature as most predictive. While this is expected and follows physiological reasoning, it is important to note that our models show a high level of accuracy for genotype level yield prediction simply with weather variables and pedigree at individual locations.</p>
<p>Attention-based interpretability revealed that the variables temperature and solar radiation were most attended during the soybean growing season, confirming their central role in regulating photosynthetic activity during seed development (<xref ref-type="bibr" rid="B19">Hamed et al., 2021</xref>; <xref ref-type="bibr" rid="B28">Khaki and Wang, 2019</xref>). Air temperature has previously reported to affect seed yield, oil, and protein, with higher temperatures often shortening the growing season, thereby reducing yield and oil content (<xref ref-type="bibr" rid="B40">Piper and Boote, 1999</xref>; <xref ref-type="bibr" rid="B25">James E. Board and Modali, 2005</xref>; <xref ref-type="bibr" rid="B54">Thenveettil et al., 2024</xref>), and reducing seed quality (<xref ref-type="bibr" rid="B16">Egli et al., 2005</xref>). Solar radiation is also known to affect soybean development, including impacting biomass, seed yield, and seed composition (<xref ref-type="bibr" rid="B46">Shibles and Weber, 1966</xref>; <xref ref-type="bibr" rid="B7">Board and Kahlon, 2011</xref>; <xref ref-type="bibr" rid="B6">Bianculli et al., 2016</xref>). Although we do not interpret the attention weights as absolute measures of variable importance, they reflect the model&#x00027;s focus conditioned on interactions among each variable over each temporal sequence. Our results indicate how the attention weights reflect the importance of environmental data, and that this new method is able to identify and focus on specific variables which have been identified as important in crop physiology. Together, these analyses demonstrate that the model effectively learned physiologically meaningful associations without being implicitly programmed with pre-assigned weights. Additionally, the use of self-attention provides a mechanism to guide future modeling efforts.</p>
<p>The proposed framework advances the state of trait prediction modeling by enabling generalizable, interpretable, and multi-objective inference on genotypes across traits and locations. Nevertheless, challenges remain. Data imbalance across maturity zones and limited representation of extreme weather events may affect performance under uncertain growing weather conditions. Additionally, while attention mechanisms provide interpretability, they remain correlational rather than causal. Thus, there is a need for an integration with domain-based process models for deeper inference. There is a need for incorporating genomic data, such as SNP markers or geno expression profiles, that can enhance genotype-specific predictions and improve the model&#x00027;s ability to capture G &#x000D7; E interactions at a finer resolution. Another promising direction involves integrating other high-throughput phenotyping data such as hyperspectral indices, or canopy temperature to enrich temporal modeling with more physiological context. Both of these approaches would provide greater information on individual genotypes, allowing for greater capture of the variability caused by unique varieties which is presently lacking in this study. However, generating such datasets will take a concerted effort as these types of data are costly and time consuming to generate. Initial advancements are already noted, for example, a recent SNP dataset from the Nothern Uniform Soybean Tests (<xref ref-type="bibr" rid="B62">Wartha et al., 2025</xref>).</p>
</sec>
<sec sec-type="conclusions" id="s5">
<label>5</label>
<title>Conclusion</title>
<p>In this study, we developed and evaluated a transformer-based deep learning framework for predicting soybean yield across multi-environment trials in the U.S., integrating temporal weather data with genotypic cluster and contextual information. Our approach achieved high predictive performance, with the transformer model outperforming traditional and recurrent baselines in yield prediction. By systematically varying temporal resolutions, we identified that intermediate sampling intervals provided the best balance between temporal fidelity and model efficiency, an insight useful for designing future data embedding strategies in Transformer-based agronomic modeling. We additionally demonstrated that pre-trained temporal representations from yield prediction could be effectively transferred to downstream tasks such as oil and protein prediction, which shows the generalizability of the learned features and highlights a practical strategy for modeling traits with limited labeled data. Leveraging the self-attention mechanism inherent to transformers, our interpretability analysis shows the potential of attention-based models not only as a predictive tool but as a tool to generate actionable insights for breeding and crop management, similar to existing <italic>post-hoc</italic> interpretability methods. This can serve as an important step toward building explicitly interpretable deep learning models, although this should be further tested on the transferability to other crop species.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>TA: Writing &#x02013; review &#x00026; editing, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing &#x02013; original draft. LV: Data curation, Investigation, Methodology, Validation, Visualization, Writing &#x02013; original draft, Writing &#x02013; review &#x00026; editing. QL: Formal analysis, Methodology, Writing &#x02013; original draft. TG: Conceptualization, Data curation, Writing &#x02013; review &#x00026; editing. JS: Data curation, Writing &#x02013; review &#x00026; editing. AS: Writing &#x02013; review &#x00026; editing, Funding acquisition. BG: Funding acquisition, Writing &#x02013; review &#x00026; editing. SS: Funding acquisition, Writing &#x02013; review &#x00026; editing, Conceptualization, Project administration, Resources, Supervision. AS: Conceptualization, Funding acquisition, Project administration, Resources, Supervision, Writing &#x02013; review &#x00026; editing.</p>
</sec>
<sec sec-type="COI-statement" id="conf1">
<title>Conflict of interest</title>
<p>TG was employed by Amazon Web Services.</p>
<p>The remaining author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s9">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x00027;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="s11">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/frai.2026.1750108/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/frai.2026.1750108/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.pdf" id="SM1" mimetype="application/pdf" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ashfaq</surname> <given-names>M.</given-names></name> <name><surname>Khan</surname> <given-names>I.</given-names></name> <name><surname>Shah</surname> <given-names>D.</given-names></name> <name><surname>Ali</surname> <given-names>S.</given-names></name> <name><surname>Tahir</surname> <given-names>M.</given-names></name></person-group> (<year>2025</year>). <article-title>Predicting wheat yield using deep learning and multi-source environmental data</article-title>. <source>Sci. Rep</source>. <volume>15</volume>:<fpage>26446</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-025-11780-7</pub-id><pub-id pub-id-type="pmid">40691684</pub-id></mixed-citation>
</ref>
<ref id="B2">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Aviles Toledo</surname> <given-names>C.</given-names></name> <name><surname>Crawford</surname> <given-names>M. M.</given-names></name> <name><surname>Tuinstra</surname> <given-names>M. R.</given-names></name></person-group> (<year>2024</year>). <article-title>Integrating multi-modal remote sensing, deep learning, and attention mechanisms for yield prediction in plant breeding experiments</article-title>. <source>Front. Plant Sci</source>. <volume>15</volume>:<fpage>1408047</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2024.1408047</pub-id><pub-id pub-id-type="pmid">39119495</pub-id></mixed-citation>
</ref>
<ref id="B3">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ayanlade</surname> <given-names>T. T.</given-names></name> <name><surname>Jones</surname> <given-names>S. E.</given-names></name> <name><surname>der Laan</surname> <given-names>L. V.</given-names></name> <name><surname>Chattopadhyay</surname> <given-names>S.</given-names></name> <name><surname>Elango</surname> <given-names>D.</given-names></name> <name><surname>Raigne</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Multi-modal ai for ultra-precision agriculture</article-title>. <source>Stud. Big Data</source> <volume>161</volume>, <fpage>299</fpage>&#x02013;<lpage>334</lpage>. doi: <pub-id pub-id-type="doi">10.1007/978-981-97-7762-4_13</pub-id></mixed-citation>
</ref>
<ref id="B4">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Balcan</surname> <given-names>M.-F.</given-names></name> <name><surname>Sharma</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Learning accurate and interpretable tree-based models</article-title>. <source>arXiv preprint arXiv:2405.15911</source>.</mixed-citation>
</ref>
<ref id="B5">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Barbosa Dos Santos</surname> <given-names>V.</given-names></name> <name><surname>Moreno Ferreira Dos Santos</surname> <given-names>A.</given-names></name> <name><surname>da Silva Cabral de Moraes</surname> <given-names>J. R.</given-names></name> <name><surname>de Oliveira Vieira</surname> <given-names>I. C.</given-names></name> <name><surname>de Souza Rolim</surname> <given-names>G.</given-names></name></person-group> (<year>2022</year>). <article-title>Machine learning algorithms for soybean yield forecasting in the Brazilian Cerrado</article-title>. <source>J. Sci. Food Agric</source>. <volume>102</volume>, <fpage>3665</fpage>&#x02013;<lpage>3672</lpage>. doi: <pub-id pub-id-type="doi">10.1002/jsfa.11713</pub-id><pub-id pub-id-type="pmid">34893984</pub-id></mixed-citation>
</ref>
<ref id="B6">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bianculli</surname> <given-names>M. L.</given-names></name> <name><surname>Aguirrez&#x000E1;bal</surname> <given-names>L. A. N.</given-names></name> <name><surname>Pereyra Irujo</surname> <given-names>G. A.</given-names></name> <name><surname>Echarte</surname> <given-names>M. M.</given-names></name></person-group> (<year>2016</year>). <article-title>Contribution of incident solar radiation on leaves and pods to soybean seed weight and composition</article-title>. <source>Eur. J. Agron</source>. <volume>77</volume>, <fpage>1</fpage>&#x02013;<lpage>9</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.eja.2016.03.002</pub-id></mixed-citation>
</ref>
<ref id="B7">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Board</surname> <given-names>J. E.</given-names></name> <name><surname>Kahlon</surname> <given-names>C. S.</given-names></name></person-group> (<year>2011</year>). <article-title>Soybean yield formation: what controls it and how it can be improved</article-title>. <source>Soybean Physiol. Biochem</source>. <volume>2011</volume>, <fpage>1</fpage>&#x02013;<lpage>36</lpage>. doi: <pub-id pub-id-type="doi">10.5772/17596</pub-id></mixed-citation>
</ref>
<ref id="B8">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Boote</surname> <given-names>K. J.</given-names></name> <name><surname>Jones</surname> <given-names>J. W.</given-names></name> <name><surname>Pickering</surname> <given-names>N. B.</given-names></name></person-group> (<year>1996</year>). <article-title>Potential uses and limitations of crop models</article-title>. <source>Agric. Syst</source>. <volume>52</volume>, <fpage>213</fpage>&#x02013;<lpage>234</lpage>. doi: <pub-id pub-id-type="doi">10.2134/agronj1996.00021962008800050005x</pub-id></mixed-citation>
</ref>
<ref id="B9">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Carroll</surname> <given-names>M. E.</given-names></name> <name><surname>Riera</surname> <given-names>L. G.</given-names></name> <name><surname>Miller</surname> <given-names>B. A.</given-names></name> <name><surname>Dixon</surname> <given-names>P. M.</given-names></name> <name><surname>Ganapathysubramanian</surname> <given-names>B.</given-names></name> <name><surname>Sarkar</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Leveraging soil mapping and machine learning to improve spatial adjustments in plant breeding trials</article-title>. <source>Crop Sci</source>. <volume>61</volume>, <fpage>3135</fpage>&#x02013;<lpage>3152</lpage>. doi: <pub-id pub-id-type="doi">10.1002/csc2.21336</pub-id></mixed-citation>
</ref>
<ref id="B10">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chen</surname> <given-names>G.</given-names></name> <name><surname>Wiatrak</surname> <given-names>P.</given-names></name></person-group> (<year>2010</year>). <article-title>Soybean development and yield are influenced by planting date and environmental conditions in the southeastern coastal plain, United States</article-title>. <source>Agron. J</source>. <volume>102</volume>, <fpage>1731</fpage>&#x02013;<lpage>1737</lpage>. doi: <pub-id pub-id-type="doi">10.2134/agronj2010.0219</pub-id></mixed-citation>
</ref>
<ref id="B11">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Chiozza</surname> <given-names>M. V.</given-names></name> <name><surname>Shook</surname> <given-names>J. M.</given-names></name> <name><surname>Van der Laan</surname> <given-names>L.</given-names></name> <name><surname>Singh</surname> <given-names>A. K.</given-names></name> <name><surname>Miguez</surname> <given-names>F. E.</given-names></name></person-group> (<year>2025</year>). <article-title>Comprehensive assessment of soybean seed composition from field trials spanning 22 us states and 24 years: Predictive insights</article-title>. <source>Crop Sci</source>. <volume>65</volume>:<fpage>e70142</fpage>. doi: <pub-id pub-id-type="doi">10.1002/csc2.70142</pub-id></mixed-citation>
</ref>
<ref id="B12">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Cooper</surname> <given-names>M.</given-names></name> <name><surname>Powell</surname> <given-names>O.</given-names></name> <name><surname>Voss-Fels</surname> <given-names>K. P.</given-names></name> <name><surname>Messina</surname> <given-names>C. D.</given-names></name> <name><surname>Gho</surname> <given-names>C.</given-names></name> <name><surname>Podlich</surname> <given-names>D. W.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Modelling selection response in plant-breeding programs using crop models as mechanistic gene-to-phenotype (CGM-G2P) multi-trait link functions</article-title>. <source>In silico Plants</source> <volume>3</volume>:<fpage>diaa016</fpage>. doi: <pub-id pub-id-type="doi">10.1093/insilicoplants/diaa016</pub-id></mixed-citation>
</ref>
<ref id="B13">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Crossa</surname> <given-names>J.</given-names></name> <name><surname>Martini</surname> <given-names>J. W. R.</given-names></name> <name><surname>Vitale</surname> <given-names>P.</given-names></name> <name><surname>P&#x000E9;rez-Rodr&#x000ED;guez</surname> <given-names>P.</given-names></name> <name><surname>Costa-Neto</surname> <given-names>G.</given-names></name> <name><surname>Fritsche-Neto</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Expanding genomic prediction in plant breeding: harnessing big data, machine learning, and advanced software</article-title>. <source>Trends Plant Sci</source>. <volume>30</volume>, <fpage>756</fpage>&#x02013;<lpage>774</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.tplants.2024.12.009</pub-id><pub-id pub-id-type="pmid">39890501</pub-id></mixed-citation>
</ref>
<ref id="B14">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>DeYoung</surname> <given-names>J.</given-names></name> <name><surname>Jain</surname> <given-names>S.</given-names></name> <name><surname>Rajani</surname> <given-names>N. F.</given-names></name> <name><surname>Lehman</surname> <given-names>E.</given-names></name> <name><surname>Xiong</surname> <given-names>C.</given-names></name> <name><surname>Socher</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2019</year>). <article-title>Eraser: a benchmark to evaluate rationalized NLP models</article-title>. <source>arXiv preprint arXiv:1911.03429</source>.</mixed-citation>
</ref>
<ref id="B15">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Dwivedi</surname> <given-names>R.</given-names></name> <name><surname>Dave</surname> <given-names>D.</given-names></name> <name><surname>Naik</surname> <given-names>H.</given-names></name> <name><surname>Singhal</surname> <given-names>S.</given-names></name> <name><surname>Omer</surname> <given-names>R.</given-names></name> <name><surname>Patel</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2023</year>). <article-title>Explainable AI (xAI): core ideas, techniques, and solutions</article-title>. <source>ACM Comput. Surv</source>. <volume>55</volume>, <fpage>1</fpage>&#x02013;<lpage>33</lpage>. doi: <pub-id pub-id-type="doi">10.1145/3561048</pub-id></mixed-citation>
</ref>
<ref id="B16">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Egli</surname> <given-names>D.</given-names></name> <name><surname>TeKrony</surname> <given-names>D.</given-names></name> <name><surname>Heitholt</surname> <given-names>J.</given-names></name> <name><surname>Rupe</surname> <given-names>J.</given-names></name></person-group> (<year>2005</year>). <article-title>Air temperature during seed filling and soybean seed germination and vigor</article-title>. <source>Crop Sci</source>. <volume>45</volume>, <fpage>1329</fpage>&#x02013;<lpage>1335</lpage>. doi: <pub-id pub-id-type="doi">10.2135/cropsci2004.0029</pub-id></mixed-citation>
</ref>
<ref id="B17">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Eugenio</surname> <given-names>F. C.</given-names></name> <name><surname>Grohs</surname> <given-names>M.</given-names></name> <name><surname>Venancio</surname> <given-names>L. P.</given-names></name> <name><surname>Schuh</surname> <given-names>M.</given-names></name> <name><surname>Bottega</surname> <given-names>E. L.</given-names></name> <name><surname>Ruoso</surname> <given-names>R.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Estimation of soybean yield from machine learning techniques and multispectral RPAS imagery</article-title>. <source>Remote Sens. Appl. Soc. Environ</source>. <volume>20</volume>:<fpage>100397</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.rsase.2020.100397</pub-id></mixed-citation>
</ref>
<ref id="B18">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Gonz&#x000E1;lez</surname> <given-names>F.</given-names></name> <name><surname>Garc&#x000ED;a-Abadillo</surname> <given-names>J.</given-names></name> <name><surname>Jarqu&#x000ED;n</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>Introducing CHiDO-a no code genomic prediction software implementation for the characterization and integration of driven omics</article-title>. <source>Plant Genome</source> <volume>18</volume>:<fpage>e20519</fpage>. doi: <pub-id pub-id-type="doi">10.1002/tpg2.20519</pub-id><pub-id pub-id-type="pmid">39447214</pub-id></mixed-citation>
</ref>
<ref id="B19">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hamed</surname> <given-names>R.</given-names></name> <name><surname>Van Loon</surname> <given-names>A. F.</given-names></name> <name><surname>Aerts</surname> <given-names>J.</given-names></name> <name><surname>Coumou</surname> <given-names>D.</given-names></name></person-group> (<year>2021</year>). <article-title>Impacts of compound hot-dry extremes on us soybean yields</article-title>. <source>Earth Syst. Dyn</source>. <volume>12</volume>, <fpage>1371</fpage>&#x02013;<lpage>1391</lpage>. doi: <pub-id pub-id-type="doi">10.5194/esd-12-1371-2021</pub-id></mixed-citation>
</ref>
<ref id="B20">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hernandez</surname> <given-names>C. M.</given-names></name> <name><surname>Correndo</surname> <given-names>A.</given-names></name> <name><surname>Kyveryga</surname> <given-names>P.</given-names></name> <name><surname>Prestholt</surname> <given-names>A.</given-names></name> <name><surname>Ciampitti</surname> <given-names>I. A.</given-names></name></person-group> (<year>2023</year>). <article-title>On-farm soybean seed protein and oil prediction using satellite data</article-title>. <source>Comput. Electron. Agric</source>. <volume>212</volume>:<fpage>108096</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2023.108096</pub-id></mixed-citation>
</ref>
<ref id="B21">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hochreiter</surname> <given-names>S.</given-names></name> <name><surname>Schmidhuber</surname> <given-names>J.</given-names></name></person-group> (<year>1997</year>). <article-title>Long short-term memory</article-title>. <source>Neural Comput</source>. <volume>9</volume>, <fpage>1735</fpage>&#x02013;<lpage>1780</lpage>. doi: <pub-id pub-id-type="doi">10.1162/neco.1997.9.8.1735</pub-id></mixed-citation>
</ref>
<ref id="B22">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hodges</surname> <given-names>T.</given-names></name> <name><surname>French</surname> <given-names>V.</given-names></name></person-group> (<year>1985</year>). <article-title>Soyphen: Soybean growth stages modeled from temperature, daylength, and water availability<sup>1</sup></article-title>. <source>Agron. J</source>. <volume>77</volume>, <fpage>500</fpage>&#x02013;<lpage>505</lpage>. doi: <pub-id pub-id-type="doi">10.2134/agronj1985.00021962007700030031x</pub-id></mixed-citation>
</ref>
<ref id="B23">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Islam</surname> <given-names>M. K.</given-names></name> <name><surname>Valentine</surname> <given-names>T.</given-names></name> <name><surname>Sue</surname> <given-names>T. J.</given-names></name> <name><surname>Karmacharya</surname> <given-names>A.</given-names></name> <name><surname>Benham</surname> <given-names>L. N.</given-names></name> <name><surname>Wang</surname> <given-names>Z.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Interpreting time series transformer models and sensitivity analysis of population age groups to covid-19 infections</article-title>. <source>arXiv preprint arXiv:2401.15119</source>.</mixed-citation>
</ref>
<ref id="B24">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jagtap</surname> <given-names>S. S.</given-names></name> <name><surname>Jones</surname> <given-names>J. W.</given-names></name></person-group> (<year>2002</year>). <article-title>Adaptation and evaluation of the CROPGRO-soybean model to predict regional yield and production</article-title>. <source>Agric. Ecosyst. Environ</source>. <volume>93</volume>, <fpage>73</fpage>&#x02013;<lpage>85</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S0167-8809(01)00358-9</pub-id></mixed-citation>
</ref>
<ref id="B25">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>James</surname> <given-names>E.</given-names></name> <name><surname>Board Modali</surname> <given-names>H.</given-names></name></person-group> (<year>2005</year>). <article-title>Dry matter accumulation predictors for optimal yield in soybean</article-title>. <source>Crop Sci</source>. <volume>45</volume>, <fpage>1790</fpage>&#x02013;<lpage>1799</lpage>. doi: <pub-id pub-id-type="doi">10.2135/cropsci2004.0602</pub-id></mixed-citation>
</ref>
<ref id="B26">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jannink</surname> <given-names>J.-L.</given-names></name> <name><surname>Lorenz</surname> <given-names>A. J.</given-names></name> <name><surname>Iwata</surname> <given-names>H.</given-names></name></person-group> (<year>2010</year>). <article-title>Genomic selection in plant breeding: from theory to practice</article-title>. <source>Brief. Funct. Genomics</source> <volume>9</volume>, <fpage>166</fpage>&#x02013;<lpage>177</lpage>. doi: <pub-id pub-id-type="doi">10.1093/bfgp/elq001</pub-id><pub-id pub-id-type="pmid">20156985</pub-id></mixed-citation>
</ref>
<ref id="B27">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jones</surname> <given-names>J. W.</given-names></name> <name><surname>Hoogenboom</surname> <given-names>G.</given-names></name> <name><surname>Porter</surname> <given-names>C. H.</given-names></name> <name><surname>Boote</surname> <given-names>K. J.</given-names></name> <name><surname>Batchelor</surname> <given-names>W. D.</given-names></name> <name><surname>Hunt</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2003</year>). <article-title>The dssat cropping system model</article-title>. <source>Eur. J. Agron</source>. <volume>18</volume>, <fpage>235</fpage>&#x02013;<lpage>265</lpage>. doi: <pub-id pub-id-type="doi">10.1016/S1161-0301(02)00107-7</pub-id></mixed-citation>
</ref>
<ref id="B28">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khaki</surname> <given-names>S.</given-names></name> <name><surname>Wang</surname> <given-names>L.</given-names></name></person-group> (<year>2019</year>). <article-title>Crop yield prediction using deep neural networks</article-title>. <source>Front. Plant Sci</source>. <volume>10</volume>:<fpage>621</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2019.00621</pub-id><pub-id pub-id-type="pmid">31191564</pub-id></mixed-citation>
</ref>
<ref id="B29">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Krause</surname> <given-names>M. D.</given-names></name> <name><surname>Piepho</surname> <given-names>H.-P.</given-names></name> <name><surname>Dias</surname> <given-names>K. O. G.</given-names></name> <name><surname>Singh</surname> <given-names>A. K.</given-names></name> <name><surname>Beavis</surname> <given-names>W. D.</given-names></name></person-group> (<year>2023</year>). <article-title>Models to estimate genetic gain of soybean seed yield from annual multi-environment field trials</article-title>. <source>Theoret. Appl. Genet</source>. <volume>136</volume>:<fpage>252</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s00122-023-04470-3</pub-id><pub-id pub-id-type="pmid">37987845</pub-id></mixed-citation>
</ref>
<ref id="B30">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lobell</surname> <given-names>D. B.</given-names></name> <name><surname>Hammer</surname> <given-names>G. L.</given-names></name> <name><surname>McLean</surname> <given-names>G.</given-names></name> <name><surname>Messina</surname> <given-names>C.</given-names></name> <name><surname>Roberts</surname> <given-names>M. J.</given-names></name> <name><surname>Schlenker</surname> <given-names>W.</given-names></name></person-group> (<year>2013</year>). <article-title>The critical role of extreme heat for maize production in the United States</article-title>. <source>Nat. Clim. Chang</source>. <volume>3</volume>, <fpage>497</fpage>&#x02013;<lpage>501</lpage>. doi: <pub-id pub-id-type="doi">10.1038/nclimate1832</pub-id></mixed-citation>
</ref>
<ref id="B31">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Lundberg</surname> <given-names>S. M.</given-names></name> <name><surname>Lee</surname> <given-names>S.-I.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;A unified approach to interpreting model predictions,&#x0201D;</article-title> in <source>Advances in Neural Information Processing Systems</source>, 30.</mixed-citation>
</ref>
<ref id="B32">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Mahesh</surname> <given-names>P.</given-names></name> <name><surname>Soundrapandiyan</surname> <given-names>R.</given-names></name></person-group> (<year>2024</year>). <article-title>Yield prediction for crops by gradient-based algorithms</article-title>. <source>PLoS ONE</source> <volume>19</volume>, <fpage>1</fpage>&#x02013;<lpage>20</lpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0291928</pub-id><pub-id pub-id-type="pmid">39186769</pub-id></mixed-citation>
</ref>
<ref id="B33">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Messina</surname> <given-names>C. D.</given-names></name> <name><surname>Jones</surname> <given-names>J. W.</given-names></name> <name><surname>Boote</surname> <given-names>K. J.</given-names></name> <name><surname>Vallejos</surname> <given-names>C. E.</given-names></name></person-group> (<year>2006</year>). <article-title>A gene-based model to simulate soybean development and yield responses to environment</article-title>. <source>Crop Sci</source>. <volume>46</volume>, <fpage>456</fpage>&#x02013;<lpage>466</lpage>. doi: <pub-id pub-id-type="doi">10.2135/cropsci2005.04-0372</pub-id></mixed-citation>
</ref>
<ref id="B34">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Meuwissen</surname> <given-names>T. H.</given-names></name> <name><surname>Hayes</surname> <given-names>B. J.</given-names></name> <name><surname>Goddard</surname> <given-names>M. E.</given-names></name></person-group> (<year>2001</year>). <article-title>Prediction of total genetic value using genome-wide dense marker maps</article-title>. <source>Genetics</source> <volume>157</volume>, <fpage>1819</fpage>&#x02013;<lpage>1829</lpage>. doi: <pub-id pub-id-type="doi">10.1093/genetics/157.4.1819</pub-id><pub-id pub-id-type="pmid">11290733</pub-id></mixed-citation>
</ref>
<ref id="B35">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Monteiro</surname> <given-names>L. A.</given-names></name> <name><surname>Ramos</surname> <given-names>R. M.</given-names></name> <name><surname>Battisti</surname> <given-names>R.</given-names></name> <name><surname>Soares</surname> <given-names>J. R.</given-names></name> <name><surname>Oliveira</surname> <given-names>J. C.</given-names></name> <name><surname>Figueiredo</surname> <given-names>G. K. D. A.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Potential use of data-driven models to estimate and predict soybean yields at national scale in Brazil</article-title>. <source>Int. J. Plant Prod</source>. <volume>16</volume>, <fpage>691</fpage>&#x02013;<lpage>703</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s42106-022-00209-0</pub-id></mixed-citation>
</ref>
<ref id="B36">
<mixed-citation publication-type="web"><collab>NASA</collab> (<year>2025</year>). <source>NASA: The power project</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://power.larc.nasa.gov//">https://power.larc.nasa.gov//</ext-link> (Accessed June 10, 2022).</mixed-citation>
</ref>
<ref id="B37">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ozyegen</surname> <given-names>O.</given-names></name> <name><surname>Ilic</surname> <given-names>I.</given-names></name> <name><surname>Cevik</surname> <given-names>M.</given-names></name></person-group> (<year>2022</year>). <article-title>Evaluation of interpretability methods for multivariate time series forecasting</article-title>. <source>Appl. Intell</source>. <volume>52</volume>, <fpage>4727</fpage>&#x02013;<lpage>4743</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10489-021-02662-2</pub-id><pub-id pub-id-type="pmid">34764613</pub-id></mixed-citation>
</ref>
<ref id="B38">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Parmley</surname> <given-names>K.</given-names></name> <name><surname>Nagasubramanian</surname> <given-names>K.</given-names></name> <name><surname>Sarkar</surname> <given-names>S.</given-names></name> <name><surname>Ganapathysubramanian</surname> <given-names>B.</given-names></name> <name><surname>Singh</surname> <given-names>A. K.</given-names></name></person-group> (<year>2019</year>). <article-title>Development of optimized phenomic predictors for efficient plant breeding decisions using phenomic-assisted selection in soybean</article-title>. <source>Plant Phen</source>. <volume>58</volume>:<fpage>404</fpage>. doi: <pub-id pub-id-type="doi">10.34133/2019/5809404</pub-id><pub-id pub-id-type="pmid">33313530</pub-id></mixed-citation>
</ref>
<ref id="B39">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Pereyra</surname> <given-names>V. M.</given-names></name> <name><surname>Hernandez</surname> <given-names>C.</given-names></name> <name><surname>Hefley</surname> <given-names>T.</given-names></name> <name><surname>Sharda</surname> <given-names>A.</given-names></name> <name><surname>Prasad</surname> <given-names>P. V. V.</given-names></name> <name><surname>Puntel</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2026</year>). <article-title>Regional drivers of soybean farmers&#x00027; yield, seed protein, and oil concentration</article-title>. <source>Field Crops Res</source>. <volume>335</volume>:<fpage>110176</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.fcr.2025.110176</pub-id></mixed-citation>
</ref>
<ref id="B40">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Piper</surname> <given-names>E. L.</given-names></name> <name><surname>Boote</surname> <given-names>K. I.</given-names></name></person-group> (<year>1999</year>). <article-title>Temperature and cultivar effects on soybean seed oil and protein concentrations</article-title>. <source>J. Am. Oil Chem. Soc</source>. <volume>76</volume>, <fpage>1233</fpage>&#x02013;<lpage>1241</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11746-999-0099-y</pub-id></mixed-citation>
</ref>
<ref id="B41">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ribeiro</surname> <given-names>M. T.</given-names></name> <name><surname>Singh</surname> <given-names>S.</given-names></name> <name><surname>Guestrin</surname> <given-names>C.</given-names></name></person-group> (<year>2016</year>). <article-title>&#x0201C;&#x02018;Why should i trust you?&#x00027; Explaining the predictions of any classifier,&#x0201D;</article-title> in <source>Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>, 1135&#x02013;1144. doi: <pub-id pub-id-type="doi">10.1145/2939672.2939778</pub-id></mixed-citation>
</ref>
<ref id="B42">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rose</surname> <given-names>G.</given-names></name> <name><surname>Osborne</surname> <given-names>T.</given-names></name> <name><surname>Greatrex</surname> <given-names>H.</given-names></name> <name><surname>Wheeler</surname> <given-names>T.</given-names></name></person-group> (<year>2016</year>). <article-title>Impact of progressive global warming on the global-scale yield of maize and soybean</article-title>. <source>Clim. Change</source> <volume>134</volume>, <fpage>417</fpage>&#x02013;<lpage>428</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s10584-016-1601-9</pub-id></mixed-citation>
</ref>
<ref id="B43">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sakamoto</surname> <given-names>T.</given-names></name></person-group> (<year>2020</year>). <article-title>Incorporating environmental variables into a MODIS-based crop yield estimation method for United States corn and soybeans through the use of a random forest regression algorithm</article-title>. <source>ISPRS J. Photogramm. Remote Sens</source>. <volume>160</volume>, <fpage>208</fpage>&#x02013;<lpage>228</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.isprsjprs.2019.12.012</pub-id></mixed-citation>
</ref>
<ref id="B44">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Schwalbert</surname> <given-names>R. A.</given-names></name> <name><surname>Amado</surname> <given-names>T.</given-names></name> <name><surname>Corassa</surname> <given-names>G.</given-names></name> <name><surname>Pott</surname> <given-names>L. P.</given-names></name> <name><surname>Prasad</surname> <given-names>P. V. V.</given-names></name> <name><surname>Ciampitti</surname> <given-names>I. A.</given-names></name></person-group> (<year>2020</year>). <article-title>Satellite-based soybean yield forecast: integrating machine learning and weather data for improving crop yield prediction in southern Brazil</article-title>. <source>Agric. For. Meteorol</source>. <volume>284</volume>:<fpage>107886</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.agrformet.2019.107886</pub-id></mixed-citation>
</ref>
<ref id="B45">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shi</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>S.</given-names></name> <name><surname>Qu</surname> <given-names>P.</given-names></name> <name><surname>Shao</surname> <given-names>J.</given-names></name></person-group> (<year>2024</year>). <article-title>Time series prediction model using LSTM-transformer neural network for mine water inflow</article-title>. <source>Sci. Rep</source>. <volume>14</volume>:<fpage>18284</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-024-69418-z</pub-id><pub-id pub-id-type="pmid">39112684</pub-id></mixed-citation>
</ref>
<ref id="B46">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shibles</surname> <given-names>R. M.</given-names></name> <name><surname>Weber</surname> <given-names>C. R.</given-names></name></person-group> (<year>1966</year>). <article-title>Interception of solar radiation and dry matter production by various soybean planting patterns<sup>1</sup></article-title>. <source>Crop Sci</source>. <volume>6</volume>, <fpage>55</fpage>&#x02013;<lpage>59</lpage>. doi: <pub-id pub-id-type="doi">10.2135/cropsci1966.0011183X000600010017x</pub-id></mixed-citation>
</ref>
<ref id="B47">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shook</surname> <given-names>J.</given-names></name> <name><surname>Gangopadhyay</surname> <given-names>T.</given-names></name> <name><surname>Wu</surname> <given-names>L.</given-names></name> <name><surname>Ganapathysubramanian</surname> <given-names>B.</given-names></name> <name><surname>Sarkar</surname> <given-names>S.</given-names></name> <name><surname>Singh</surname> <given-names>A. K.</given-names></name></person-group> (<year>2021</year>). <article-title>Crop yield prediction integrating genotype and weather variables using deep learning</article-title>. <source>PLoS ONE</source> <volume>16</volume>:<fpage>e0252402</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0252402</pub-id><pub-id pub-id-type="pmid">34138872</pub-id></mixed-citation>
</ref>
<ref id="B48">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Shrikumar</surname> <given-names>A.</given-names></name> <name><surname>Greenside</surname> <given-names>P.</given-names></name> <name><surname>Kundaje</surname> <given-names>A.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Learning important features through propagating activation differences,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>PMlR</publisher-loc>), <fpage>3145</fpage>&#x02013;<lpage>3153</lpage>.</mixed-citation>
</ref>
<ref id="B49">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>A. K.</given-names></name> <name><surname>Jones</surname> <given-names>S. E.</given-names></name> <name><surname>Van der Laan</surname> <given-names>L.</given-names></name> <name><surname>Ayanlade</surname> <given-names>T. T.</given-names></name> <name><surname>Raigne</surname> <given-names>J.</given-names></name> <name><surname>Saleem</surname> <given-names>N.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Use of artificial intelligence in soybean breeding and production</article-title>. <source>Adv. Agron</source>. <volume>190</volume>, <fpage>199</fpage>&#x02013;<lpage>273</lpage>. doi: <pub-id pub-id-type="doi">10.1016/bs.agron.2024.11.003</pub-id></mixed-citation>
</ref>
<ref id="B50">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Singh</surname> <given-names>D.</given-names></name> <name><surname>Singh</surname> <given-names>A.</given-names></name> <name><surname>Singh</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>Chapter 28-phenomics and machine learning in crop improvement</article-title>. <source>Plant Breed. Cult. Dev</source>. <volume>15</volume>, <fpage>563</fpage>&#x02013;<lpage>594</lpage>. doi: <pub-id pub-id-type="doi">10.1016/B978-0-12-817563-7.00011-8</pub-id></mixed-citation>
</ref>
<ref id="B51">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Sundararajan</surname> <given-names>M.</given-names></name> <name><surname>Taly</surname> <given-names>A.</given-names></name> <name><surname>Yan</surname> <given-names>Q.</given-names></name></person-group> (<year>2017</year>). <article-title>&#x0201C;Axiomatic attribution for deep networks,&#x0201D;</article-title> in <source>International Conference on Machine Learning</source> (<publisher-loc>PMLR</publisher-loc>), <fpage>3319</fpage>&#x02013;<lpage>3328</lpage>.</mixed-citation>
</ref>
<ref id="B52">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tan</surname> <given-names>S.</given-names></name> <name><surname>Caruana</surname> <given-names>R.</given-names></name> <name><surname>Hooker</surname> <given-names>G.</given-names></name> <name><surname>Koch</surname> <given-names>P.</given-names></name> <name><surname>Gordo</surname> <given-names>A.</given-names></name></person-group> (<year>2018</year>). <article-title>Learning global additive explanations for neural nets using model distillation</article-title>. <source>Stat</source> <volume>1050</volume>, <fpage>1518</fpage>&#x02013;<lpage>68323</lpage>.</mixed-citation>
</ref>
<ref id="B53">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Technow</surname> <given-names>F.</given-names></name> <name><surname>Messina</surname> <given-names>C. D.</given-names></name> <name><surname>Totir</surname> <given-names>L. R.</given-names></name> <name><surname>Cooper</surname> <given-names>M.</given-names></name></person-group> (<year>2015</year>). <article-title>Integrating crop growth models with whole genome prediction through approximate Bayesian computation</article-title>. <source>PLoS ONE</source> <volume>10</volume>:<fpage>e0130855</fpage>. doi: <pub-id pub-id-type="doi">10.1371/journal.pone.0130855</pub-id><pub-id pub-id-type="pmid">26121133</pub-id></mixed-citation>
</ref>
<ref id="B54">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Thenveettil</surname> <given-names>N.</given-names></name> <name><surname>Bheemanahalli</surname> <given-names>R.</given-names></name> <name><surname>Reddy</surname> <given-names>K. N.</given-names></name> <name><surname>Gao</surname> <given-names>W.</given-names></name> <name><surname>Reddy</surname> <given-names>K. R.</given-names></name></person-group> (<year>2024</year>). <article-title>Temperature and elevated co2 alter soybean seed yield and quality, exhibiting transgenerational effects on seedling emergence and vigor</article-title>. <source>Front. Plant Sci</source>. <volume>15</volume>:<fpage>1427086</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2024.1427086</pub-id><pub-id pub-id-type="pmid">39145187</pub-id></mixed-citation>
</ref>
<ref id="B55">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Tolhurst</surname> <given-names>D. J.</given-names></name> <name><surname>Gaynor</surname> <given-names>R. C.</given-names></name> <name><surname>Gardunia</surname> <given-names>B.</given-names></name> <name><surname>Hickey</surname> <given-names>J. M.</given-names></name> <name><surname>Gorjanc</surname> <given-names>G.</given-names></name></person-group> (<year>2022</year>). <article-title>Genomic selection using random regressions on known and latent environmental covariates</article-title>. <source>Theor. Appl.Genet</source>. <volume>135</volume>, <fpage>3393</fpage>&#x02013;<lpage>3415</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00122-022-04186-w</pub-id><pub-id pub-id-type="pmid">36066596</pub-id></mixed-citation>
</ref>
<ref id="B56">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Torsoni</surname> <given-names>G. B.</given-names></name> <name><surname>de Oliveira Aparecido</surname> <given-names>L. E.</given-names></name> <name><surname>dos Santos</surname> <given-names>G. M.</given-names></name> <name><surname>Chiquitto</surname> <given-names>A. G.</given-names></name> <name><surname>da Silva Cabral Moraes</surname> <given-names>J. R.</given-names></name> <name><surname>de Souza Rolim</surname> <given-names>G.</given-names></name></person-group> (<year>2023</year>). <article-title>Soybean yield prediction by machine learning and climate</article-title>. <source>Theor. Appl. Climatol</source>. <volume>151</volume>, <fpage>1709</fpage>&#x02013;<lpage>1725</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s00704-022-04341-9</pub-id></mixed-citation>
</ref>
<ref id="B57">
<mixed-citation publication-type="web"><collab>USDA-ARS</collab> (<year>2025a</year>). <source>Uniform soybean tests, northern region</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.ars.usda.gov/midwest-area/west-lafayette-in/crop-production-and-pest-control-research/docs/uniform-soybean-tests-northern-region/">https://www.ars.usda.gov/midwest-area/west-lafayette-in/crop-production-and-pest-control-research/docs/uniform-soybean-tests-northern-region/</ext-link> (Accessed May 27, 2025).</mixed-citation>
</ref>
<ref id="B58">
<mixed-citation publication-type="web"><collab>USDA-ARS</collab> (<year>2025b</year>). <source>Uniform soybean tests, southern region</source>. Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.ars.usda.gov/southeast-area/stoneville-ms/crop-genetics-research/docs/uniform-soybean-tests/">https://www.ars.usda.gov/southeast-area/stoneville-ms/crop-genetics-research/docs/uniform-soybean-tests/</ext-link> (Accessed May 27, 2025).</mixed-citation>
</ref>
<ref id="B59">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Van der Laan</surname> <given-names>L.</given-names></name> <name><surname>Parmley</surname> <given-names>K.</given-names></name> <name><surname>Saadati</surname> <given-names>M.</given-names></name> <name><surname>Pacin</surname> <given-names>H. T.</given-names></name> <name><surname>Panthulugiri</surname> <given-names>S.</given-names></name> <name><surname>Sarkar</surname> <given-names>S.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Genomic and phenomic prediction for soybean seed yield, protein, and oil</article-title>. <source>Plant Genome</source> <volume>18</volume>:<fpage>e70002</fpage>. doi: <pub-id pub-id-type="doi">10.1002/tpg2.70002</pub-id><pub-id pub-id-type="pmid">39972529</pub-id></mixed-citation>
</ref>
<ref id="B60">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Vaswani</surname> <given-names>A.</given-names></name> <name><surname>Shazeer</surname> <given-names>N.</given-names></name> <name><surname>Parmar</surname> <given-names>N.</given-names></name> <name><surname>Uszkoreit</surname> <given-names>J.</given-names></name> <name><surname>Jones</surname> <given-names>L.</given-names></name> <name><surname>Gomez</surname> <given-names>A. N.</given-names></name> <etal/></person-group>. (<year>2017</year>). <article-title>&#x0201C;Attention is all you need,&#x0201D;</article-title> in <source>Advances in Neural Information Processing Systems</source>, 30.</mixed-citation>
</ref>
<ref id="B61">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Feng</surname> <given-names>K.</given-names></name> <name><surname>Sun</surname> <given-names>L.</given-names></name> <name><surname>Xie</surname> <given-names>Y.</given-names></name> <name><surname>Song</surname> <given-names>X.-P.</given-names></name></person-group> (<year>2024</year>). <article-title>Satellite-based soybean yield prediction in argentina: a comparison between panel regression and deep learning methods</article-title>. <source>Comput. Electron. Agric</source>. <volume>221</volume>:<fpage>108978</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.compag.2024.108978</pub-id></mixed-citation>
</ref>
<ref id="B62">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wartha</surname> <given-names>C. A.</given-names></name> <name><surname>Campbell</surname> <given-names>B. W.</given-names></name> <name><surname>Ramasubramanian</surname> <given-names>V.</given-names></name> <name><surname>Nice</surname> <given-names>L.</given-names></name> <name><surname>Brock</surname> <given-names>A.</given-names></name> <name><surname>Cai</surname> <given-names>G.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Genomic analysis and predictive modeling in the northern uniform soybean tests</article-title>. <source>Crop Sci</source>. <volume>65</volume>:<fpage>e70138</fpage>. doi: <pub-id pub-id-type="doi">10.1002/csc2.70138</pub-id></mixed-citation>
</ref>
<ref id="B63">
<mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Winn</surname> <given-names>Z. J.</given-names></name> <name><surname>Amsberry</surname> <given-names>A. L.</given-names></name> <name><surname>Haley</surname> <given-names>S. D.</given-names></name> <name><surname>DeWitt</surname> <given-names>N. D.</given-names></name> <name><surname>Mason</surname> <given-names>R. E.</given-names></name></person-group> (<year>2023</year>). <article-title>Phenomic versus genomic prediction&#x02013;a comparison of prediction accuracies for grain yield in hard winter wheat lines</article-title>. <source>Plant Phenome J</source>. <volume>6</volume>:<fpage>e20084</fpage>. doi: <pub-id pub-id-type="doi">10.1002/ppj2.20084</pub-id></mixed-citation>
</ref>
<ref id="B64">
<mixed-citation publication-type="web"><person-group person-group-type="author"><name><surname>Woo</surname> <given-names>G.</given-names></name> <name><surname>Liu</surname> <given-names>C.</given-names></name> <name><surname>Kumar</surname> <given-names>A.</given-names></name> <name><surname>Xiong</surname> <given-names>C.</given-names></name> <name><surname>Savarese</surname> <given-names>S.</given-names></name> <name><surname>Sahoo</surname> <given-names>D.</given-names></name></person-group> (<year>2024</year>). <article-title>&#x0201C;Unified training of universal time series forecasting transformers,&#x0201D;</article-title> in <source>Proceedings of the 41st International Conference on Machine Learning, ICML&#x00027;24</source> (<publisher-loc><ext-link ext-link-type="uri" xlink:href="https://JMLR.org">JMLR.org</ext-link></publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>25</lpage>.</mixed-citation>
</ref>
<ref id="B65">
<mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Xiong</surname> <given-names>Z.</given-names></name> <name><surname>Guo</surname> <given-names>X.</given-names></name> <name><surname>Zhang</surname> <given-names>D.</given-names></name> <name><surname>Tseng</surname> <given-names>S.-P.</given-names></name> <name><surname>Ji</surname> <given-names>Y.</given-names></name> <name><surname>Sun</surname> <given-names>X.</given-names></name></person-group> (<year>2024</year>). <article-title>&#x0201C;Crop yield prediction based on attention-LSTM model,&#x0201D;</article-title> in <source>2024 International Conference on Orange Technology (ICOT)</source> (<publisher-loc>IEEE</publisher-loc>), <fpage>1</fpage>&#x02013;<lpage>4</lpage>. doi: <pub-id pub-id-type="doi">10.1109/ICOT64290.2024.10936959</pub-id></mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3167551/overview">Lei Song</ext-link>, Rutgers, The State University of New Jersey, United States</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1465428/overview">Banafsheh Rekabdar</ext-link>, Portland State University, United States</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2863460/overview">Ajay Kumar</ext-link>, University of Missouri, United States</p>
</fn>
</fn-group>
</back>
</article>