<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Therm. Eng.</journal-id>
<journal-title>Frontiers in Thermal Engineering</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Therm. Eng.</abbrev-journal-title>
<issn pub-type="epub">2813-0456</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1594443</article-id>
<article-id pub-id-type="doi">10.3389/fther.2025.1594443</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Thermal Engineering</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>
<italic>Phy-ChemNODE</italic>: an end-to-end physics-constrained autoencoder-NeuralODE framework for learning stiff chemical kinetics of hydrocarbon fuels</article-title>
<alt-title alt-title-type="left-running-head">Kumar et al.</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fther.2025.1594443">10.3389/fther.2025.1594443</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Kumar</surname>
<given-names>Tadbhagya</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/validation/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Kumar</surname>
<given-names>Anuj</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2819722/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/software/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Pal</surname>
<given-names>Pinaki</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1160335/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/supervision/"/>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/resources/"/>
<role content-type="https://credit.niso.org/contributor-roles/funding-acquisition/"/>
<role content-type="https://credit.niso.org/contributor-roles/project-administration/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Transportation and Power Systems Division, Argonne National Laboratory (DOE)</institution>, <addr-line>Lemont</addr-line>, <addr-line>IL</addr-line>, <country>United States</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Department of Mechanical and Aerospace Engineering, North Carolina State University</institution>, <addr-line>Raleigh</addr-line>, <addr-line>NC</addr-line>, <country>United States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1515900/overview">Fang Wang</ext-link>, Zhongyuan University of Technology, China</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2071721/overview">Yuanqing Wu</ext-link>, Dongguan University of Technology, China</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3028622/overview">Tommaso Grassi</ext-link>, Max Planck Institute for Extraterrestrial Physics, Germany</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3031222/overview">Ahmed Almeldein</ext-link>, Oak Ridge National Laboratory (DOE), United States</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Pinaki Pal, <email>pal@anl.gov</email>
</corresp>
</author-notes>
<pub-date pub-type="epub">
<day>15</day>
<month>08</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>5</volume>
<elocation-id>1594443</elocation-id>
<history>
<date date-type="received">
<day>16</day>
<month>03</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>30</day>
<month>07</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright This work is authored by Tadbhagya Kumar, Anuj Kumar and Pinaki Pal, &#xa9; 2025 UChicago Argonne, LLC.</copyright-statement>
<copyright-year>2025</copyright-year>
<copyright-holder>This work is authored by Tadbhagya Kumar, Anuj Kumar and Pinaki Pal, &#xa9; 2025 UChicago Argonne, LLC</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Predictive computational fluid dynamics (CFD) simulations of reacting flows in energy conversion systems are accompanied by a major computational bottleneck of solving a stiff system of coupled ordinary differential equations (ODEs) associated with detailed fuel chemistry. This issue is exacerbated with the complexity of fuel chemistry as the number of reactive scalars and chemical reactions increase. In this work, a physics-constrained Autoencoder (AE)-NeuralODE framework, termed as PhyChemNODE, is developed for data-driven modeling and temporal emulation of stiff chemical kinetics for complex hydrocarbon fuels, wherein a non-linear AE is employed for dimensionality reduction of the thermochemical state and the NODE learns temporal dynamics of the system in the low-dimensional latent space obtained from the AE. Both the AE and NODE are trained together in an end-to-end manner. We further enhance the approach by incorporating elemental mass conservation constraints directly into the loss function during model training. This ensures that total mass as well as individual elemental species masses are conserved in an <italic>a-posteriori</italic> manner. Demonstration studies are performed for methane combustion kinetics (32 species, 266 chemical reactions) over a wide thermodynamic and composition space at high pressure. Effects of various model hyperparameters, such as relative weighting of different terms in the loss function and dimensionality of the AE latent space, on the accuracy of Phy-ChemNODE are assessed. The physics-based constraints are shown to improve both training efficiency and physical consistency of the data-driven model. Further, <italic>a-posteriori</italic> autoregressive inference tests demonstrate that Phy-ChemNODE leads to reduced temporal stiffness in the latent space, and achieves 1-3 orders of magnitude speedup relative to the detailed kinetic mechanism depending on the type of ODE solver (implicit or explicit) used for numerical integration, while ensuring prediction fidelity.</p>
</abstract>
<kwd-group>
<kwd>stiff chemical kinetics</kwd>
<kwd>fuel combustion</kwd>
<kwd>autoencoders</kwd>
<kwd>neural ordinary differential equations</kwd>
<kwd>physics-informed neural network</kwd>
<kwd>chemistry acceleration</kwd>
<kwd>computational fluid dynamics</kwd>
<kwd>numerical modeling</kwd>
</kwd-group>
<contract-sponsor id="cn001">Office of Fossil Energy and Carbon Management<named-content content-type="fundref-id">10.13039/100020312</named-content>
</contract-sponsor>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Heat Engines</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>Computational fluid dynamics (CFD) modeling of reacting flows, such as those encountered in gas turbine and internal combustion engines, are computationally demanding owing to the complex interactions among multiple physico-chemical phenomena and the need to resolve a wide range of spatiotemporal scales governing the evolution of a large number of reactive scalars (chemical species). In particular, modeling of detailed chemical kinetics presents a major bottleneck, which is governed by a stiff system of coupled ordinary differential equations (ODEs) and characterized by high condition number of the corresponding chemical Jacobian matrix (<xref ref-type="bibr" rid="B27">Shampine, 1993</xref>). In addition, as the complexity of fuel chemistry increases, the dimensionality and stiffness of the ODE system also become more prohibitive (<xref ref-type="bibr" rid="B18">Lu and Law, 2009</xref>). To address these computational challenges, kinetic model reduction is typically performed (<xref ref-type="bibr" rid="B17">Lu and Law, 2005</xref>; <xref ref-type="bibr" rid="B10">Jones and Rigopoulos, 2005</xref>; <xref ref-type="bibr" rid="B30">Valorani et al., 2006</xref>; <xref ref-type="bibr" rid="B19">Maas and Pope, 1992</xref>), but it often leads to less reliable description of chemical kinetics.</p>
<p>In this context, data-driven machine learning (ML) techniques (<xref ref-type="bibr" rid="B6">Dana, 1987</xref>; <xref ref-type="bibr" rid="B34">Wold et al., 1987</xref>) have been extensively explored in an effort to emulate chemical kinetics and accelerate detailed finite-rate chemistry computations. Some of these approaches reduce the dimensionality of the reaction system via linear projection onto an appropriate basis typically identified through Principal Component Analysis (PCA). Artificial neural network (ANNs) are, then, employed for regression of the PC source terms and transport coefficients (<xref ref-type="bibr" rid="B24">Owoyele et al., 2017</xref>; <xref ref-type="bibr" rid="B13">Kumar et al., 2023</xref>). On the other hand, ANN-based approaches have also been used for predicting chemical source terms directly from the thermochemical state (<xref ref-type="bibr" rid="B5">Christo et al., 1996</xref>; <xref ref-type="bibr" rid="B3">Blasco et al., 1998</xref>; <xref ref-type="bibr" rid="B2">Sen et al., 2010</xref>; <xref ref-type="bibr" rid="B26">Ranade et al., 2019</xref>; <xref ref-type="bibr" rid="B32">Wan et al., 2020</xref>). More recently, novel deep learning architectures have been explored to capture the temporal evolution of chemical kinetics and combustion simulations, such as Deep Operator Network (DeepONet) (<xref ref-type="bibr" rid="B12">Kumar and Echekki, 2024</xref>) and Fourier Neural Operator (FNO) (<xref ref-type="bibr" rid="B35">Zhang et al., 2024</xref>).</p>
<p>Despite the potential benefits of employing ML methods, When they are trained in an <italic>a-priori</italic> or offline setting and then coupled with a numerical solver, predicted solutions may diverge or become unstable. Due to the non-linearity of the combustion process, even minor predictive errors can escalate into significant discrepancies in the temporal evolution of thermochemical state due to error accumulation. To handle this issue, an alternative and more robust data-driven technique for chemical source term computations based on neural ordinary differential equations (NeuralODEs, NODEs) (<xref ref-type="bibr" rid="B4">Chen et al., 2018</xref>) was developed by the authors, known as ChemNODE (<xref ref-type="bibr" rid="B23">Owoyele and Pal, 2021</xref>). It combines the chemical source term predictions with ODE integration in an <italic>a-posteriori</italic> learning paradigm, where the source terms predicted by the neural network are passed to the ODE solver for time integration, and the neural network weights are optimized to minimize the loss computed between the predicted and ground truth thermochemical states (comprised of species mass fractions and thermodynamic variables). A key advantage of this approach is that NODEs learn continuous-time dynamics which can be integrated using existing ODE solvers. This ensures that the predicted thermochemical state, even after a long-time horizon, remains adherent to the ground truth solution trajectory. For relatively large chemical kinetic mechanisms, the coupling of a non-linear autoencoder (AE) to perform dimensionality reduction and a NODE to evolve the dynamics in the lower-dimensional latent space has shown promise (<xref ref-type="bibr" rid="B31">Vijayarangan et al., 2024</xref>). A similar approach has also been pursued in the field of astro-chemistry (<xref ref-type="bibr" rid="B29">Sun Tang and Turk, 2022</xref>; <xref ref-type="bibr" rid="B20">Maes et al., 2024</xref>).</p>
<p>Another major limitation of traditional black-box ML techniques applied to the modeling of chemical kinetics is that they do not inherently incorporate conservation laws, which can adversely impact simulation accuracy and hinder the integration of ML surrogate models with multidimensional CFD solvers. With the advent of scientific machine learning, the practice of combining domain science-specific constraints to embed physics into neural network training has emerged lately (<xref ref-type="bibr" rid="B25">Raissi et al., 2019</xref>). This is achieved through regularization of the loss function, in which physical laws are embedded in the learning process as soft constraints (<xref ref-type="bibr" rid="B14">Kumar Tadbhagya et al., 2023</xref>; <xref ref-type="bibr" rid="B1">Almeldein and Van Dam, 2023</xref>; <xref ref-type="bibr" rid="B16">Kumar et al., 2025</xref>; <xref ref-type="bibr" rid="B33">Weng et al., 2025</xref>; <xref ref-type="bibr" rid="B11">Kercher and Votsmeier, 2025</xref>). This has been shown to improve model accuracy and generalization, especially in scenarios where traditional methods struggle with noisy data and high-dimensional problems governed by parameterized differential equations. Another way to enforce conservation laws is by enforcing a hard constraint in the neural network architecture through constraint layers (<xref ref-type="bibr" rid="B28">Sturm and Wexler, 2022</xref>; <xref ref-type="bibr" rid="B21">Mohan et al., 2023</xref>).</p>
<p>In light of the above discussion, the overarching goal of the present work is to demonstrate an end-to-end physics-constrained AE-NODE framework called Phy-ChemNODE for data-driven modeling and temporal emulation of stiff chemical kinetics for hydrocarbon fuels. Unlike previous work (<xref ref-type="bibr" rid="B31">Vijayarangan et al., 2024</xref>; <xref ref-type="bibr" rid="B29">Sun Tang and Turk, 2022</xref>; <xref ref-type="bibr" rid="B20">Maes et al., 2024</xref>), the present study demonstrates that this integrated approach improves physical consistency of the resulting data-driven model with mass conservation laws and accelerates training convergence, while simultaneously reducing temporal stiffness of the chemical kinetic system. The remainder of the paper is organized as follows: the Phy-ChemNODE framework is first outlined along with details of the physics-constrained formulation and training methodology in <xref ref-type="sec" rid="s2">Sections 2</xref>, <xref ref-type="sec" rid="s3">3</xref>, respectively. Subsequently, results from <italic>a-posteriori</italic> proof-of-concept studies are discussed in <xref ref-type="sec" rid="s4">Section 4</xref>. Finally, the Conclusions section summarizes the major findings and directions for future work.</p>
</sec>
<sec id="s2">
<title>2 Physics-constrained autoencoder-NODE (Phy-ChemNODE) framework for stiff chemical kinetics emulation</title>
<p>In combustion CFD simulations, it is a common numerical approach to decouple finite-rate chemistry from transport using operator splitting. Chemistry is solved (independently from advective and diffusive transport) within each computational grid cell considered as a homogeneous reactor, which is equivalent to solving a system of stiff ODEs. The temporal evolution of <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> reactive scalars (chemical species) can be defined by:<disp-formula id="e1">
<mml:math id="m2">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x307;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1,2,3</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the mass fraction of <inline-formula id="inf3">
<mml:math id="m4">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> species (<inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> being the total number of species), <inline-formula id="inf5">
<mml:math id="m6">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x307;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> is the corresponding chemical source term computed using law of mass action, and <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> refers to mixture density. The temporal evolution of temperature (<italic>T</italic>) is governed by a similar ODE defined as follows:<disp-formula id="e2">
<mml:math id="m8">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>&#x307;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c1;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:mspace width="1em"/>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1,2,3</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>where <inline-formula id="inf7">
<mml:math id="m9">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>h</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf8">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">c</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> refer to the enthalpy of <inline-formula id="inf9">
<mml:math id="m11">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> species and mixture-averaged constant pressure specific heat, respectively. To calculate the source terms, one needs to account for several elementary reactions involving production and consumption of multiple species. As the chemical mechanism becomes larger, the number of chemical species and elementary reactions also increase (<xref ref-type="bibr" rid="B18">Lu and Law, 2009</xref>). This leads to prohibitive computational costs since all chemical time scales must be fully resolved. In the NODE-based data-driven framework (<xref ref-type="bibr" rid="B23">Owoyele and Pal, 2021</xref>), the expensive physics-based computation of chemical source terms is replaced by a neural network, which can be described as <xref ref-type="disp-formula" rid="e3">Equation 3</xref> below:<disp-formula id="e3">
<mml:math id="m12">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>where <inline-formula id="inf10">
<mml:math id="m13">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mo>&#x2026;</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the vector of thermochemical state (temperature and species mass fractions), and <inline-formula id="inf11">
<mml:math id="m14">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is a feedforward neural network parameterized by weights <inline-formula id="inf12">
<mml:math id="m15">
<mml:mrow>
<mml:mi mathvariant="normal">&#x398;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. For larger chemical mechanisms, <inline-formula id="inf13">
<mml:math id="m16">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> increases in dimensionality and stiffness. To address the high dimensionality, a non-linear AE is coupled with the NODE for dimensionality reduction, so that the NODE learns the temporal evolution of the dynamical system in a reduced-order latent space obtained from the non-linear projection of the AE. A schematic of the coupled AE-NODE data-driven modeling framework is shown in <xref ref-type="fig" rid="F1">Figure 1</xref>. The model training process is posed as an optimization problem of determining the optimal parameters of the encoder <inline-formula id="inf14">
<mml:math id="m17">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, NODE <inline-formula id="inf15">
<mml:math id="m18">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>h</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and the decoder <inline-formula id="inf16">
<mml:math id="m19">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> networks in an end-to-end manner, that minimizes the loss function defined as:<disp-formula id="e4">
<mml:math id="m20">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">Phy&#x2212;ChemNODE</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">data</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where the reconstruction loss <inline-formula id="inf17">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>L</mml:mtext>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> measures the loss between ground truth <inline-formula id="inf18">
<mml:math id="m22">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and corresponding encoder-decoder mapping (<inline-formula id="inf19">
<mml:math id="m23">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo>&#x303;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c6;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, the data loss <inline-formula id="inf20">
<mml:math id="m24">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">data</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mtext>L</mml:mtext>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
<mml:mo>,</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> measures the loss between ground truth and encoder &#x2b; NODE &#x2b; decoder prediction <inline-formula id="inf21">
<mml:math id="m25">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and the latent loss <inline-formula id="inf22">
<mml:math id="m26">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> measures the loss between encoder mapping of ground truth (<inline-formula id="inf23">
<mml:math id="m27">
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo>&#x304;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; <inline-formula id="inf24">
<mml:math id="m28">
<mml:mrow>
<mml:mi>&#x3c6;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>) and encoder &#x2b; NODE prediction. Each loss term is chosen to be in mean absolute error (MAE) form. It is noted that the loss terms <inline-formula id="inf25">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf26">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> ensure that the encoder and decoder mappings are bijective or unique. The loss function also contains elemental mass conservation constraints (<xref ref-type="bibr" rid="B14">Kumar Tadbhagya et al., 2023</xref>; <xref ref-type="bibr" rid="B16">Kumar et al., 2025</xref>), defined in <xref ref-type="disp-formula" rid="e5">Equation 5</xref> below:<disp-formula id="e5">
<mml:math id="m31">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msubsup>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>where <inline-formula id="inf27">
<mml:math id="m32">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> refers to the loss associated with mass conservation of element <inline-formula id="inf28">
<mml:math id="m33">
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> (in the chemical system with a total of <inline-formula id="inf29">
<mml:math id="m34">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> elements). <inline-formula id="inf30">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mover accent="true">
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">&#x302;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf31">
<mml:math id="m36">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>Y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> correspond to the AE &#x2b; NODE predicted and ground truth mass fractions of <inline-formula id="inf32">
<mml:math id="m37">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> species, respectively. <inline-formula id="inf33">
<mml:math id="m38">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the atomic mass of element <inline-formula id="inf34">
<mml:math id="m39">
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf35">
<mml:math id="m40">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula> is the number of atoms of element <inline-formula id="inf36">
<mml:math id="m41">
<mml:mrow>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> in <inline-formula id="inf37">
<mml:math id="m42">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> species, <inline-formula id="inf38">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>W</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the molecular weight of <inline-formula id="inf39">
<mml:math id="m44">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mi>h</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> species, and <inline-formula id="inf40">
<mml:math id="m45">
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of training data points. Lastly, the weights <inline-formula id="inf41">
<mml:math id="m46">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf42">
<mml:math id="m47">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf43">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> in <xref ref-type="disp-formula" rid="e4">Equation 4</xref> balance the contributions from the different loss terms.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Schematic of the coupled AE-NODE framework.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g001.tif">
<alt-text content-type="machine-generated">Diagram illustrating a Neural ODE model. It shows data entering from &#x22;Physical Space&#x22; at time \( t = t^n \) into an Encoder \(\varphi(\Phi)\), transitioning into &#x22;Latent Space&#x22; represented by neural networks \( h(z) \), and then passing through a Decoder \(\psi(z)\) to reach &#x22;Physical Space&#x22; at time \( t = t^{n+1} \). Equations for transformation are depicted.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s3">
<title>3 Proof-of-concept study</title>
<p>For proof-of-concept demonstration of Phy-ChemNODE, an autoigniting methane-oxygen (<inline-formula id="inf44">
<mml:math id="m49">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>-<inline-formula id="inf45">
<mml:math id="m50">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>O</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) zero-dimensional (0D) homogeneous reactor is considered, at a constant pressure of 20&#xa0;atm that corresponds to practical high-pressure isobaric operating conditions for stationary gas turbine engines. The detailed chemical mechanism (<xref ref-type="bibr" rid="B8">Gregory et al., 2016</xref>) consists of 32 species and 266 chemical reactions. The ground truth data for model training is generated using Cantera (<xref ref-type="bibr" rid="B7">Goodwin et al., 2009</xref>), which solves the coupled ODE system (<xref ref-type="disp-formula" rid="e1">Equations 1</xref>, <xref ref-type="disp-formula" rid="e2">2</xref>) with detailed chemistry. The thermodynamic and composition space chosen for data generation comprises 9 equispaced initial temperatures in the range <inline-formula id="inf46">
<mml:math id="m51">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; [1600&#xa0;K, 2000 K] and 11 equivalence ratios within <inline-formula id="inf47">
<mml:math id="m52">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:mn>1.0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.5</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> resulting in a total of 99 initial conditions. Each of these initial conditions is integrated to chemical equilibrium. Ground truth data is generated by selecting time instants such that the change in any of the thermochemical scalars (species mass fractions or temperature) between two successive time instants is greater than 1% of their corresponding overall ranges of variation, and then downsampling 200 points from the selected time instants. A <inline-formula id="inf48">
<mml:math id="m53">
<mml:mrow>
<mml:mn>70</mml:mn>
<mml:mi>%</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>20</mml:mn>
<mml:mi>%</mml:mi>
<mml:mo>,</mml:mo>
<mml:mn>10</mml:mn>
<mml:mi>%</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> random split (based on initial conditions) is used to obtain the training, validation, and test datasets, respectively. The AE-NODE model is initialized with the same initial conditions (during training) as the physics-based simulations. The input to the encoder is the vector <inline-formula id="inf49">
<mml:math id="m54">
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> containing the temperature and species mass fractions, which is scaled using the maximum and minimum of the training data, respectively, and the output is a vector in the latent space <inline-formula id="inf52">
<mml:math id="m57">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. The decoder has the same dense architecture as the encoder, with an input size equal to the latent dimension <inline-formula id="inf53">
<mml:math id="m58">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> (chosen based on a sensitivity study discussed in <xref ref-type="sec" rid="s4">Section 4</xref>) and the output size equal to the physical space vector <inline-formula id="inf54">
<mml:math id="m59">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="normal">&#x3a6;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>33</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Both the encoder and decoder have 5 hidden layers with 64 neurons each and Exponential Linear Unit (ELU) activation function. The NODE has the same input and output dimensions as the latent space <inline-formula id="inf55">
<mml:math id="m60">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and a 4 hidden-layer dense network with 64 neurons in each hidden layer and ELU activation function, to model the chemical source terms in the lower-dimensional latent space. The output layers for the encoder, NODE, and the decoder are considered to be linear. The above discussed AE-NODE architecture was finalized based on multiple ablation studies, targeting a good balance between dimensionality reduction and overall model accuracy. Although more exhaustive hyperparameter optimization can be performed to further optimize the number of latent variables, it is out of scope of the current work.</p>
<p>The forward pass through the NODE requires time integration, for which a 4th order explicit Runge-Kutta (RK) solver is used. Once the time integration is completed, the temporal trajectories are mapped back to physical space and the loss is computed using <xref ref-type="disp-formula" rid="e4">Equation 4</xref>. The gradients for updating the neural network parameters are calculated using backward adjoint automatic differentiation and ADAM optimizer with exponential learning rate decay (every 200 epochs) is used. The model is trained for 10000 epochs. To ensure that all the loss terms are of similar magnitude, <inline-formula id="inf56">
<mml:math id="m61">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> is used. Further, <inline-formula id="inf57">
<mml:math id="m62">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf58">
<mml:math id="m63">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>H</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>O</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> are chosen based on a hyperparameter sweep (discussed in <xref ref-type="sec" rid="s4">Section 4</xref>). The training framework was implemented in Julia programming language using Flux.jl (<xref ref-type="bibr" rid="B9">Innes et al., 2018</xref>) library and the model was trained on 2 AMD EPYC 7713 64-core processors for a walltime of 96&#xa0;h.</p>
</sec>
<sec sec-type="results|discussion" id="s4">
<title>4 Results and discussion</title>
<p>To determine the optimal weighting of terms in the loss function and the latent space dimensionality, hyperparameter studies were carried out. <xref ref-type="fig" rid="F2">Figure 2a</xref> compares the loss terms computed on the validation set (post training) corresponding to different <inline-formula id="inf59">
<mml:math id="m64">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> values for <inline-formula id="inf60">
<mml:math id="m65">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and fixed size of the latent space (dim<inline-formula id="inf61">
<mml:math id="m66">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>4</mml:mn>
</mml:math>
</inline-formula>), and <xref ref-type="fig" rid="F2">Figure 2b</xref> shows a similar comparison for varying size of the latent space (dim<inline-formula id="inf62">
<mml:math id="m67">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>) with <inline-formula id="inf63">
<mml:math id="m68">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">rec</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf64">
<mml:math id="m69">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf65">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. Moreover, to assess the impact of adding elemental mass conservation constraints to the training loss function, <xref ref-type="fig" rid="F2">Figure 2c</xref> compares the decay of data loss (during training) between the cases trained with <inline-formula id="inf66">
<mml:math id="m71">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ele&#x2212;j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and without elemental mass constraints <inline-formula id="inf67">
<mml:math id="m72">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">ele&#x2212;j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> on the validation set. It can be clearly seen that incorporating the soft constraints in the loss function results in lower loss for the same number of epochs, thereby enabling more efficient model training and faster training convergence.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Comparison of loss terms (computed on validation set) across hyperparameter experiments: <bold>(a)</bold> varying <inline-formula id="inf68">
<mml:math id="m73">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <bold>(b)</bold> varying latent space size <inline-formula id="inf69">
<mml:math id="m74">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mi>i</mml:mi>
<mml:mi>m</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and <bold>(c)</bold> data loss <inline-formula id="inf70">
<mml:math id="m75">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="italic">data</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> evolution with <inline-formula id="inf71">
<mml:math id="m76">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.5</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and without <inline-formula id="inf72">
<mml:math id="m77">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>e</mml:mi>
<mml:mi>l</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> elemental mass constraints.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g002.tif">
<alt-text content-type="machine-generated">Three line graphs show different loss terms. Graph (a) depicts loss terms \(L_{rec}\), \(L_{data}\), and \(L_{latent}\) against \(\lambda_z\). Graph (b) shows \(10 \times L_{rec}\), \(L_{data}\), and \(L_{latent}\) against \(dim(z)\). Graph (c) illustrates \(L_{data}\) over epochs with \(\lambda_{el-j} = 0\) and \(\lambda_{el-j} = 0.5\).</alt-text>
</graphic>
</fig>
<p>The trained Phy-ChemNODE model is then deployed for predicting the temporal evolution of thermochemical scalars in <italic>a-posteriori</italic> autoregressive tests. <xref ref-type="fig" rid="F3">Figure 3</xref> plots the temporal evolution of temperature and a subset of species mass fractions (<inline-formula id="inf74">
<mml:math id="m79">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, CO, <inline-formula id="inf75">
<mml:math id="m80">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CO</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, OH, and <inline-formula id="inf76">
<mml:math id="m81">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>O</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) for an initial condition (corresponding to <inline-formula id="inf77">
<mml:math id="m82">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1600</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K) in the training set <inline-formula id="inf78">
<mml:math id="m83">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and test set <inline-formula id="inf79">
<mml:math id="m84">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, where ground truth data is indicated by solid lines and the predicted Phy-ChemNODE solutions are shown in markers. Overall, great agreement can be observed between the predictions and ground truth data. <xref ref-type="fig" rid="F4">Figure 4</xref> shows the temporal evolution of a few intermediate species for another set of initial conditions in the training set (<inline-formula id="inf80">
<mml:math id="m85">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1650</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf81">
<mml:math id="m86">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) and the test set (<inline-formula id="inf82">
<mml:math id="m87">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1700</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K<inline-formula id="inf83">
<mml:math id="m88">
<mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>), again demonstrating high accuracy. As further quantification of the accuracy of Phy-ChemNODE framework, <xref ref-type="fig" rid="F5">Figure 5</xref> shows the test set MAEs for a subset of the thermochemical scalars including both major and minor species, scaled by their corresponding data ranges, indicating that Phy-ChemNODE performs extremely well in capturing the temporal dynamics.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Temporal evolution of temperature <inline-formula id="inf84">
<mml:math id="m89">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and mass fractions of <inline-formula id="inf85">
<mml:math id="m90">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, CO, <inline-formula id="inf86">
<mml:math id="m91">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CO</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, OH, and <inline-formula id="inf87">
<mml:math id="m92">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>O</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> corresponding to initial conditions in <bold>(a)</bold> training set (<inline-formula id="inf88">
<mml:math id="m93">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1600</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf89">
<mml:math id="m94">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) and <bold>(b)</bold> test set (<inline-formula id="inf90">
<mml:math id="m95">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1600</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf91">
<mml:math id="m96">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). The mass fractions of <inline-formula id="inf92">
<mml:math id="m97">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf93">
<mml:math id="m98">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CO</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and OH are scaled by 4, and that of CO by 3 for ease of plotting. Solid lines denote ground truth and markers denote Phy-ChemNODE predictions.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g003.tif">
<alt-text content-type="machine-generated">Two line graphs labeled (a) and (b) show molecule fractions and temperature over time in milliseconds. Graphs track O2, OH, CO, CO2, CH4, and temperature (T) with distinct colors. Both graphs display similar trends, with sharp changes around 0.01 milliseconds. Temperature is measured on the right y-axis, reaching approximately 3500 Kelvin, while molecule fractions are plotted on the left, ranging from 0 to 0.8.</alt-text>
</graphic>
</fig>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Temporal evolution of intermediate species (<inline-formula id="inf94">
<mml:math id="m99">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf95">
<mml:math id="m100">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>CH</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>O and <inline-formula id="inf96">
<mml:math id="m101">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mtext>C</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mtext>H</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) corresponding to initial conditions in <bold>(a)</bold> training set <inline-formula id="inf97">
<mml:math id="m102">
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1650</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf98">
<mml:math id="m103">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) and <bold>(b)</bold> test set <inline-formula id="inf99">
<mml:math id="m104">
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1700</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf100">
<mml:math id="m105">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.05</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>). Solid lines denote ground truth and markers denote Phy-ChemNODE predictions.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g004.tif">
<alt-text content-type="machine-generated">Two line graphs (a) and (b) show the concentration of CH3, CH2O, and C2H4 over time in milliseconds on the x-axis, with Y_i on the y-axis. Both graphs have similar trend lines with CH3 in red, CH2O in green, and C2H4 in brown. The lines increase sharply, peak, and drop quickly.</alt-text>
</graphic>
</fig>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Scaled test set MAEs for prediction of temperature <inline-formula id="inf101">
<mml:math id="m106">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and a few species mass fractions.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g005.tif">
<alt-text content-type="machine-generated">Bar chart displaying the normalized error ratio \(|y - \hat{y}| / (y_{max} - y_{min})\) for various substances: \(T\), \(H_2\), \(H\), \(O\), \(O_2\), \(OH\), \(H_2O\), \(HO_2\), \(H_2O_2\), \(CO\), \(CO_2\), \(C\), \(CH\), \(CH_2\), \(CH_3\), and \(CH_4\). Heights vary, with prominent peaks for \(H\) and \(H_2O_2\).</alt-text>
</graphic>
</fig>
<p>Based on inference on an Intel i7-1165G7 workstation with 16 cores, Phy-ChemNODE yields speedups of 6x and 860x over the full chemical mechanism in terms of overall simulation walltime, when deployed with implicit (backward differentiation formula (BDF)) and explicit (RK45) solvers, respectively. <xref ref-type="fig" rid="F6">Figure 6</xref> plots the latent space temporal dynamics corresponding to the initial conditions of <inline-formula id="inf102">
<mml:math id="m107">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1650</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf103">
<mml:math id="m108">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. It is evident that the evolution of latent variables is much smoother than that of thermochemical scalars in the physical space earlier shown in <xref ref-type="fig" rid="F4">Figure 4a</xref>. This indicates that the coupled AE-NODE approach significantly reduces the temporal stiffness of the chemical kinetic system. Similar stiffness reduction in the latent space was also observed by <xref ref-type="bibr" rid="B22">Nair et al. (2025)</xref> in case of AE-NODE models of advection-dominated dynamical systems. Lastly, <xref ref-type="fig" rid="F7">Figure 7</xref> shows the predicted temporal evolution of C, H, and O mass fractions from autoregressive tests corresponding to certain initial conditions from the training and test sets. Evidently, the model trained with elemental mass constraints in the loss function (Phy-ChemNODE/PCNODE) conserves the elemental mass fractions during deployment much better than the MAE-trained case without constraints.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Temporal evoluiton of latent space variables corresponding to <inline-formula id="inf104">
<mml:math id="m109">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1650</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf105">
<mml:math id="m110">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g006.tif">
<alt-text content-type="machine-generated">Line graph showing latent variables over time in milliseconds. Latent-1 (blue) decreases steadily, Latent-2 (black) quickly rises then stabilizes, Latent-3 (red) peaks then drops to zero, and Latent-4 (green) remains stable.</alt-text>
</graphic>
</fig>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Temporal evolution of mass fractions of: <bold>(a)</bold> C, <bold>(b)</bold> H, and <bold>(c)</bold> O elements corresponding to initial conditions in the training (<inline-formula id="inf106">
<mml:math id="m111">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1600</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf107">
<mml:math id="m112">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) and test (<inline-formula id="inf108">
<mml:math id="m113">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1600</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>K, <inline-formula id="inf109">
<mml:math id="m114">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>) sets.</p>
</caption>
<graphic xlink:href="fther-05-1594443-g007.tif">
<alt-text content-type="machine-generated">Three graphs show elemental mass fractions over time in milliseconds for carbon (C), hydrogen (H), and oxygen (O). Each graph compares Ground Truth (black line), MAE (blue circles), and PCNODE (red diamonds). The top graph (a) illustrates carbon with slight fluctuations; the middle graph (b) shows hydrogen with minimal variation; the bottom graph (c) depicts oxygen maintaining a stable pattern. Each graph has a different equivalence ratio (\(\phi\)) around 1.0 and 1.1.</alt-text>
</graphic>
</fig>
<p>In future studies, multiple avenues for further extension of the Phy-ChemNODE framework will be pursued. These include: (a) efficient scaling of the training workflow to wider ranges of thermodynamic conditions (including multiple pressures), larger kinetic mechanisms, and multicomponent fuels; (b) integration of Phy-ChemNODE with CFD solvers and demonstration of accelerated multidimensional reacting flow simulations on modern high-performance computing (HPC) platforms; (c) demonstration for constant-volume combustion; (d) incorporation of uncertainty quantification; and (e) exploration of training methodologies to enhance model out-of-distribution generalizability.</p>
</sec>
<sec sec-type="conclusion" id="s5">
<title>5 Conclusion</title>
<p>In this work, an end-to-end physics-constrained AE-NODE framework (PhyChemNODE) was introduced for accelerated temporal emulation of stiff chemical kinetics targeting complex hydrocarbon fuel combustion. The deep learning approach employed a non-linear AE for dimensionality reduction of the thermochemical state and utilized a NODE to learn temporal dynamics of the kinetic system in the low-dimensional latent space obtained from the AE. In addition, elemental mass conservation constraints were included in the loss function during training of the data-driven model to ensure that total mass and mass of each elemental species are conserved. Proof-of-concept studies were performed for homogeneous autoignition of a methane-oxygen mixture over a range of composition and thermodynamic conditions at high pressure. The results showed that the physics-based constraints not only improve physical consistency of the resulting data-driven model, but also enhance training efficiency. In addition, Phy-ChemNODE achieved 6&#x2013;860<inline-formula id="inf110">
<mml:math id="m115">
<mml:mrow>
<mml:mo>&#xd7;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> speedup relative to the detailed chemical mechanism depending on the type of ODE solver (implicit or explicit) used for numerical integration during autoregressive inference tests. Temporal evolution of the latent variables was visualized and it was found that the coupled AE-NODE approach leads to reduced temporal stiffness of the chemical kinetic system. In future work, Phy-ChemNODE will be further scaled to larger kinetic mechanisms and wider ranges of thermodynamic conditions (including pressure), and demonstrated for multidimensional combustion CFD simulations. In addition, uncertainty quantification techniques will be incorporated and training methodologies will be explored to enhance out-of-distribution generalizability.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s6">
<title>Data availability statement</title>
<p>The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.</p>
</sec>
<sec sec-type="author-contributions" id="s7">
<title>Author contributions</title>
<p>TK: Data curation, Methodology, Conceptualization, Validation, Investigation, Software, Formal Analysis, Writing &#x2013; original draft. AK: Software, Investigation, Methodology, Writing &#x2013; original draft, Formal Analysis, Data curation. PP: Methodology, Supervision, Conceptualization, Investigation, Writing &#x2013; original draft, Resources, Funding acquisition, Project administration.</p>
</sec>
<sec sec-type="funding-information" id="s8">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research and/or publication of this article. The research work was funded by the U.S. DOE Fossil Energy and Carbon Management (FECM) office through the Technology Commercialization Fund (TCF) program.</p>
</sec>
<ack>
<p>The authors would like to acknowledge the computing core hours available through the Improv cluster provided by the Laboratory Computing Resource Center (LCRC) at Argonne National Laboratory. The contents of this manuscript have been presented at the Machine Learning and the Physical Sciences Workshop organized as part of the 38th conference on Neural Information Processing Systems (NeurIPS) in 2024 (<xref ref-type="bibr" rid="B15">Kumar et al., 2024</xref>).</p>
</ack>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
<p>The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.</p>
</sec>
<sec sec-type="ai-statement" id="s10">
<title>Generative AI statement</title>
<p>The author(s) declare that no Generative AI was used in the creation of this manuscript.</p>
</sec>
<sec sec-type="disclaimer" id="s11">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="disclaimer" id="s12">
<title>Licenses and Permissions</title>
<p>The submitted manuscript has been created by UChicago Argonne, LLC, Operator of Argonne National Laboratory (Argonne). Argonne, a U.S. Department of Energy (DOE) Office of Science laboratory, is operated under Contract No. DEAC02-06CH11357. The U.S. Government retains for itself, and others acting on its behalf, a paid-up nonexclusive, irrevocable worldwide license in said article to reproduce, prepare derivative works, distribute copies to the public, and perform publicly and display publicly, by or on behalf of the Government.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Almeldein</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Van Dam</surname>
<given-names>N.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Accelerating chemical kinetics calculations with physics informed neural networks</article-title>. <source>J. Eng. Gas Turbines Power</source> <volume>145</volume> (<issue>9</issue>), <fpage>091008</fpage>. <pub-id pub-id-type="doi">10.1115/1.4062654</pub-id>
</citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Blasco</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Fueyo</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Dopazo</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Ballester</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>1998</year>). <article-title>Modelling the temporal evolution of a reduced combustion chemical system with an artificial neural network</article-title>. <source>Combust. Flame</source> <volume>113</volume> (<issue>1</issue>), <fpage>38</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1016/s0010-2180(97)00211-3</pub-id>
</citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>R. T. Q.</given-names>
</name>
<name>
<surname>Rubanova</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bettencourt</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Duvenaud</surname>
<given-names>D. K.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Neural ordinary differential equations</article-title>. <source>Adv. neural Inf. Process. Syst.</source> <volume>31</volume>.</citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Christo</surname>
<given-names>F. C.</given-names>
</name>
<name>
<surname>Masri</surname>
<given-names>A. R.</given-names>
</name>
<name>
<surname>Nebot</surname>
<given-names>E. M.</given-names>
</name>
<name>
<surname>Pope</surname>
<given-names>S. B.</given-names>
</name>
</person-group> (<year>1996</year>). <article-title>An integrated PDF/neural network approach for simulating turbulent reacting systems</article-title>. <source>Symposium Int. Combust.</source> <volume>26</volume> (<issue>1</issue>), <fpage>43</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1016/s0082-0784(96)80198-6</pub-id>
</citation>
</ref>
<ref id="B6">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dana</surname>
<given-names>H. B.</given-names>
</name>
</person-group> (<year>1987</year>). &#x201c;<article-title>Modular learning in neural networks</article-title>,&#x201d; in <source>
<italic>Proceedings of the sixth national conference on artificial intelligence - volume 1</italic>, AAAI&#x2019;87</source> (<publisher-name>AAAI Press</publisher-name>), <fpage>279</fpage>&#x2013;<lpage>284</lpage>.</citation>
</ref>
<ref id="B7">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodwin</surname>
<given-names>D. G.</given-names>
</name>
<name>
<surname>Moffat</surname>
<given-names>H. K.</given-names>
</name>
<name>
<surname>Speth</surname>
<given-names>R. L.</given-names>
</name>
</person-group> (<year>2009</year>). <source>Cantera: an object-oriented software toolkit for chemical kinetics, thermodynamics, and transport processes caltech</source>, <volume>124</volume>. <publisher-loc>CA</publisher-loc>: <publisher-name>Pasadena</publisher-name>.</citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gregory</surname>
<given-names>P. S.</given-names>
</name>
<name>
<surname>Tao</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Foundational fuel chemistry model version 1.0 (ffcm-1)</article-title>.</citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Innes</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Saba</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Gandhi</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Rudilosso</surname>
<given-names>M. C.</given-names>
</name>
<name>
<surname>Joy</surname>
<given-names>N. M.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Fashionable modelling with flux</article-title>. <source>CoRR</source>, <fpage>01457</fpage>. <comment>abs/1811</comment>.</citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jones</surname>
<given-names>W. P.</given-names>
</name>
<name>
<surname>Rigopoulos</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>Rate-controlled constrained equilibrium: formulation and application to nonpremixed laminar flames</article-title>. <source>Combust. Flame</source> <volume>142</volume> (<issue>3</issue>), <fpage>223</fpage>&#x2013;<lpage>234</lpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2005.03.008</pub-id>
</citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kercher</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Votsmeier</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Machine learning surrogate models for mechanistic kinetics: embedding atomic balance and positivity</article-title>. <source>J. Phys. Chem. Lett.</source> <volume>16</volume> (<issue>19</issue>). <pub-id pub-id-type="doi">10.1021/acs.jpclett.5c00602</pub-id>
</citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Echekki</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Combustion chemistry acceleration with DeepONets</article-title>. <source>Fuel</source> <volume>365</volume>, <fpage>131212</fpage>. <pub-id pub-id-type="doi">10.1016/j.fuel.2024.131212</pub-id>
</citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Rieth</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Owoyele</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>J. H.</given-names>
</name>
<name>
<surname>Echekki</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2023a</year>). <article-title>Acceleration of turbulent combustion dns <italic>via</italic> principal component transport</article-title>. <source>Combust. Flame</source> <volume>255</volume>, <fpage>112903</fpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2023.112903</pub-id>
</citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2023b</year>). &#x201c;<article-title>A physics-constrained neuralODE approach for robust learning of stiff chemical kinetics</article-title>,&#x201d; in <source>NeurIPS machine learning and the physical sciences workshop</source>.</citation>
</ref>
<ref id="B15">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2024</year>). &#x201c;<article-title>A physics-informed autoencoder-neuralODE framework (Phy-ChemNODE) for learning complex fuel combustion kinetics</article-title>,&#x201d; in <source>NeurIPS machine learning and the physical sciences workshop</source>.</citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>A physics-constrained neural ordinary differential equations approach for robust learning of stiff chemical kinetics</article-title>. <source>Combust. Theory Model.</source> <volume>29</volume> (<issue>2</issue>), <fpage>332</fpage>&#x2013;<lpage>347</lpage>. <pub-id pub-id-type="doi">10.1080/13647830.2025.2478266</pub-id>
</citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Law</surname>
<given-names>C. K.</given-names>
</name>
</person-group> (<year>2005</year>). <article-title>A directed relation graph method for mechanism reduction</article-title>. <source>Proc. Combust. Inst.</source> <volume>30</volume> (<issue>1</issue>), <fpage>1333</fpage>&#x2013;<lpage>1341</lpage>. <pub-id pub-id-type="doi">10.1016/j.proci.2004.08.145</pub-id>
</citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lu</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Law</surname>
<given-names>C. K.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Toward accommodating realistic fuel chemistry in large-scale computations</article-title>. <source>Prog. Energy Combust. Sci.</source> <volume>35</volume> (<issue>2</issue>), <fpage>192</fpage>&#x2013;<lpage>215</lpage>. <pub-id pub-id-type="doi">10.1016/j.pecs.2008.10.002</pub-id>
</citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Maas</surname>
<given-names>U.</given-names>
</name>
<name>
<surname>Pope</surname>
<given-names>S. B.</given-names>
</name>
</person-group> (<year>1992</year>). <article-title>Simplifying chemical kinetics: intrinsic low-dimensional manifolds in composition space</article-title>. <source>Combust. Flame</source> <volume>88</volume> (<issue>3</issue>), <fpage>239</fpage>&#x2013;<lpage>264</lpage>. <pub-id pub-id-type="doi">10.1016/0010-2180(92)90034-m</pub-id>
</citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Maes</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>De Ceuster</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Sande</surname>
<given-names>M. V. de</given-names>
</name>
<name>
<surname>Decin</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>MACE: a machine-learning approach to chemistry emulation</article-title>. <source>Astrophysical J.</source> <volume>969</volume> (<issue>79</issue>), <fpage>79</fpage>. <pub-id pub-id-type="doi">10.3847/1538-4357/ad47a1</pub-id>
</citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mohan</surname>
<given-names>A. T.</given-names>
</name>
<name>
<surname>Lubbers</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Chertkov</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Livescu</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2023</year>). <article-title>Embedding hard physical constraints in neural network coarse-graining of three-dimensional turbulence</article-title>. <source>Phys. Rev. Fluids</source> <volume>8</volume>, <fpage>014604</fpage>. <pub-id pub-id-type="doi">10.1103/physrevfluids.8.014604</pub-id>
</citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nair</surname>
<given-names>A. S.</given-names>
</name>
<name>
<surname>Barwey</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>MacArt</surname>
<given-names>J. F.</given-names>
</name>
<name>
<surname>Arcomano</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Maulik</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Understanding latent timescales in neural ordinary differential equation models of advection-dominated dynamical systems</article-title>. <source>Phys. D: Nonlinear Phenom.</source> <volume>476</volume>, <fpage>134650</fpage>. <pub-id pub-id-type="doi">10.1016/j.physd.2025.134650</pub-id>
</citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Owoyele</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>ChemNODE: a neural ordinary differential equations framework for efficient chemical kinetic solvers</article-title>. <source>Energy AI</source> <volume>7</volume>, <fpage>100118</fpage>. <pub-id pub-id-type="doi">10.1016/j.egyai.2021.100118</pub-id>
</citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Owoyele</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Echekki</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Toward computationally efficient combustion DNS with complex fuels via principal component transport</article-title>. <source>Combust. Theory Model.</source> <volume>21</volume> (<issue>4</issue>), <fpage>770</fpage>&#x2013;<lpage>798</lpage>. <pub-id pub-id-type="doi">10.1080/13647830.2017.1296976</pub-id>
</citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Raissi</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Perdikaris</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Karniadakis</surname>
<given-names>G. E.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations</article-title>. <source>J. Comput. Phys.</source> <volume>378</volume>, <fpage>686</fpage>&#x2013;<lpage>707</lpage>. <pub-id pub-id-type="doi">10.1016/j.jcp.2018.10.045</pub-id>
</citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ranade</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Alqahtani</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Farooq</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Echekki</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>An ann based hybrid chemistry framework for complex fuels</article-title>. <source>Fuel</source> <volume>241</volume>, <fpage>625</fpage>&#x2013;<lpage>636</lpage>. <pub-id pub-id-type="doi">10.1016/j.fuel.2018.12.082</pub-id>
</citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sen</surname>
<given-names>B. A</given-names>
</name>
<name>
<surname>Hawkes</surname>
<given-names>E. R.</given-names>
</name>
<name>
<surname>Menon</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Large eddy simulation of extinction and reignition with artificial neural networks based chemical kinetics</article-title>. <source>Combust. Flame</source> <volume>157</volume> (<issue>3</issue>), <fpage>566</fpage>&#x2013;<lpage>578</lpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2009.11.006</pub-id>
</citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shampine</surname>
<given-names>L. F.</given-names>
</name>
</person-group> (<year>1993</year>). <article-title>Ill-conditioned matrices and the integration of stiff ODEs</article-title>. <source>J. Comput. Appl. Math.</source> <volume>48</volume> (<issue>3</issue>), <fpage>279</fpage>&#x2013;<lpage>292</lpage>. <pub-id pub-id-type="doi">10.1016/0377-0427(93)90025-7</pub-id>
</citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sturm</surname>
<given-names>P. O.</given-names>
</name>
<name>
<surname>Wexler</surname>
<given-names>A. S.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Conservation laws in a neural network architecture: enforcing the atom balance of a julia-based photochemical model (v0.2.0)</article-title>. <source>Geosci. Model Dev.</source> <volume>15</volume> (<issue>8</issue>), <fpage>3417</fpage>&#x2013;<lpage>3431</lpage>. <pub-id pub-id-type="doi">10.5194/gmd-15-3417-2022</pub-id>
</citation>
</ref>
<ref id="B29">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sun Tang</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Turk</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2022</year>). &#x201c;<article-title>Reduced order model for chemical kinetics: a case study with primordial chemical network</article-title>,&#x201d; in <source>ICML workshop on machine learning for astrophysics</source>.</citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Valorani</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Creta</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Goussis</surname>
<given-names>D. A.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Habib</surname>
<given-names>N. N.</given-names>
</name>
</person-group> (<year>2006</year>). <article-title>An automatic procedure for the simplification of chemical kinetic mechanisms based on CSP</article-title>. <source>Combust. Flame</source> <volume>146</volume> (<issue>1</issue>), <fpage>29</fpage>&#x2013;<lpage>51</lpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2006.03.011</pub-id>
</citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vijayarangan</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Uranakara</surname>
<given-names>H. A.</given-names>
</name>
<name>
<surname>Barwey</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Malpica Galassi</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Rafi Malik</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Valorani</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2024</year>). <article-title>A data-driven reduced-order model for stiff chemical kinetics using dynamics-informed training</article-title>. <source>Energy AI</source> <volume>15</volume>, <fpage>100325</fpage>. <pub-id pub-id-type="doi">10.1016/j.egyai.2023.100325</pub-id>
</citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Barnaud</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Vervisch</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Domingo</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Chemistry reduction using machine learning trained from non-premixed micro-mixing modeling: application to dns of a syngas turbulent oxy-flame with side-wall effects</article-title>. <source>Combust. Flame</source> <volume>220</volume>, <fpage>119</fpage>&#x2013;<lpage>129</lpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2020.06.008</pub-id>
</citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>Li</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>Z. X.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2025</year>). <article-title>Extended fourier neural operators to learn stiff kinetics under unseen conditions</article-title>. <source>Combust. Flame</source> <volume>272</volume>, <fpage>113847</fpage>. <pub-id pub-id-type="doi">10.1016/j.combustflame.2024.113847</pub-id>
</citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wold</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Esbensen</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Paul</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>1987</year>). <article-title>Principal component analysis</article-title>. <source>Chemom. Intelligent Laboratory Syst.</source> <volume>2</volume> (<issue>1</issue>), <fpage>37</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1016/0169-7439(87)80084-9</pub-id>
</citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Weng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2024</year>). <article-title>Learning transient evolution of multidimensional reacting flows by multiscale fourier neural operators</article-title>. <source>Proc. Combust. Inst.</source> <volume>40</volume>, <fpage>105714</fpage>. <pub-id pub-id-type="doi">10.1016/j.proci.2024.105714</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>