<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" article-type="research-article" dtd-version="1.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/frai.2026.1766671</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>A hybrid deep boosting framework with adaptive label stabilization for SEM-based porosity estimation in fly-ash cement mortar</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author"><name><surname>Abdullah</surname>
</name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref><xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3243960"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author"><name><surname>Ather</surname> <given-names>Muhammad Ateeb</given-names></name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref><xref ref-type="aff" rid="aff2"><sup>2</sup></xref>
<uri xlink:href="https://loop.frontiersin.org/people/3321250"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author" corresp="yes"><name><surname>Rodr&#x00ED;guez</surname> <given-names>Jos&#x00E9; Luis Oropeza</given-names></name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref><xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="supervision" vocab-term-identifier="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="validation" vocab-term-identifier="https://credit.niso.org/contributor-roles/validation/">Validation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
<contrib contrib-type="author" corresp="yes"><name><surname>S&#x00E1;nchez-Mejorada</surname> <given-names>Carlos Guzm&#x00E1;n</given-names></name><xref ref-type="aff" rid="aff1"><sup>1</sup></xref><xref ref-type="corresp" rid="c001"><sup>&#x002A;</sup></xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Funding acquisition" vocab-term-identifier="https://credit.niso.org/contributor-roles/funding-acquisition/">Funding acquisition</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="investigation" vocab-term-identifier="https://credit.niso.org/contributor-roles/investigation/">Investigation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/project-administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="resources" vocab-term-identifier="https://credit.niso.org/contributor-roles/resources/">Resources</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing &#x2013; original draft</role>
</contrib>
</contrib-group>
<aff id="aff1"><label>1</label><institution>Center for Computing Research, Instituto Polit&#x00E9;cnico Nacional</institution>, <city>Mexico City</city>, <country country="mx">Mexico</country></aff>
<aff id="aff2"><label>2</label><institution>Department of Computer Sciences, Bahria University</institution>, <city>Lahore</city>, <country country="pk">Pakistan</country></aff>
<author-notes>
<corresp id="c001"><label>&#x002A;</label>Correspondence: Jos&#x00E9; Luis Oropeza Rodr&#x00ED;guez, <email xlink:href="mailto:joropeza@cic.ipn.mx">joropeza@cic.ipn.mx</email>; Carlos Guzm&#x00E1;n S&#x00E1;nchez-Mejorada, <email xlink:href="mailto:cmejorada@cic.ipn.mx">cmejorada@cic.ipn.mx</email></corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-23">
<day>23</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2026</year>
</pub-date>
<volume>9</volume>
<elocation-id>1766671</elocation-id>
<history>
<date date-type="received">
<day>12</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>15</day>
<month>01</month>
<year>2026</year>
</date>
<date date-type="accepted">
<day>29</day>
<month>01</month>
<year>2026</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#x00A9; 2026 Abdullah, Ather, Rodr&#x00ED;guez and S&#x00E1;nchez-Mejorada.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Abdullah, Ather, Rodr&#x00ED;guez and S&#x00E1;nchez-Mejorada</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-23">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Introduction</title>
<p>Accurate measurements of porosity of cementitious matrices are critical in pre- Q7 dicting mechanical behavior, durability, and transportation processes. Traditional methods based on SEM, such as manual thresholding, a simple binarization method, and end-to-end convolutional neural network (CNN) regressors, are, however, highly affected by image contrast variation, polishing quality, magnification, and operator bias. To address these limitations, the current article develops a hybrid deep-boosting framework for fully automatic porosity estimation directly from raw backscattered-electron SEM images of fly-ash cement mortar. The key novelty of the proposed approach lies in the adaptive stabilization of porosity labels and the hybrid fusion of deep semantic and handcrafted texture features, which together improve robustness to imaging artifacts, boundary ambiguity, and overfitting.</p>
</sec>
<sec>
<title>Methods</title>
<p>Annotation ground-truth porosity is optimized using an Adaptive Porosity Label Stabilizer (APLS) that successively improves Otsu threshold masks, first using entropy measures and morphological consistency measures to reduce label noise. Multiscale semantic representations are learned on a ResNet-18 backbone, which is trained with SimCLR on SEM data, while local statistical texture is captured using handcrafted gray-level cooccurrence map (GLCM) features. The resulting mismatched set of features is combined with a learnable Hybrid Feature Refinement Block (HFRB) together with a Feature-Interaction Attention (FIA) block, which explicitly characterizes inter-scale relationships among convolutional channels and texture regressors. The latent representation is then condensed and regressed using a weighted ensemble including CatBoost, XGBoost, and LightGBM learners.</p>
</sec>
<sec>
<title>Results and discussion</title>
<p>The proposed methodology achieves R<sup>2</sup> = 0.9816, RMSE = 0.0236, and MAE = 0.00875 on a rigorously held out test set, outperforming baseline methods that rely exclusively on CNN features, handcrafted descriptors, or na&#x00EF;ve hybrid combinations. The validity, stability, and physical plausibility of the model are ensured through a comprehensive assessment, including ablation studies, domain-shift experiments, uncertainty and stability calibration, and a hybrid explainability framework (Grad-CAM++, SHAP). The architecture does not require any manual segmentation, generalizes across magnifications and imaging conditions, and provides transparent, domain-consistent explanatory visualizations. Overall, the proposed framework represents an important step toward fast, reliable, and scalable SEMbased porosity estimation in cementitious systems.</p>
</sec>
</abstract>
<kwd-group>
<kwd>cementitious materials</kwd>
<kwd>feature&#x2013;interaction attention</kwd>
<kwd>gradient boosting ensemble</kwd>
<kwd>hybrid deep learning</kwd>
<kwd>materials informatics</kwd>
<kwd>porosity estimation</kwd>
<kwd>self-supervised learning</kwd>
<kwd>SEM image analysis</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was not received for this work and/or its publication.</funding-statement>
</funding-group>
<counts>
<fig-count count="11"/>
<table-count count="14"/>
<equation-count count="32"/>
<ref-count count="59"/>
<page-count count="22"/>
<word-count count="13606"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Machine Learning and Artificial Intelligence</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="sec1">
<label>1</label>
<title>Introduction</title>
<p>Porosity is the ratio between the volume of a porous material that is empty and the total volume of the solid material and is an important microstructural parameter that has a strong effect on mechanical strength, transport, including permeability and diffusivity, and long-term stability of porous media (<xref ref-type="bibr" rid="ref16">Holzer et al., 2023</xref>; <xref ref-type="bibr" rid="ref38">Regalla and Senthil Kumar, 2025</xref>). Cementitious materials such as concrete and mortar have a relatively large porosity, thus allowing the ingress of aggressive species, such as chlorides, sulfates, and carbon dioxide, which promotes the mechanism of degradation, such as reinforcement corrosion, freeze&#x2013;thaw damage, and alkali-silica reaction (<xref ref-type="bibr" rid="ref28">Ma et al., 2025</xref>; <xref ref-type="bibr" rid="ref48">Subash and Subash, 2024</xref>). Porosity cannot be properly predicted by accurate measurement to aid in predicting life service, maximizing mix design, and guaranteeing structural integrity of civil infrastructure (<xref ref-type="bibr" rid="ref2">Abousnina et al., 2025</xref>).</p>
<p>A similar level of criticality can be observed in other porous networks, such as geological structures (rocks and soils), energy materials like battery electrodes, where the flow of fluids, storage capacity, and electrochemical activity are controlled by porosity. In the past, porosity of cement-based materials has been measured in the laboratory through mercury intrusion porosimetry (MIP), gas adsorption, water absorption, or through direct physical measurements (<xref ref-type="bibr" rid="ref54">Wang and Zhou, 2022</xref>). Despite its precision, such methodologies are devastating, slow, and only give bulk averages that have no spatial resolution (<xref ref-type="bibr" rid="ref57">Xia et al., 2024</xref>). The advent of non-destructive imaging alternatives, especially scanning electron microscopy (SEM) in backscattered electron (BSE) mode, has become useful in microstructural imaging on the micrometer scale (<xref ref-type="bibr" rid="ref5">Ali et al., 2023</xref>).</p>
<p>SEM-BSE images take advantage of the atomic-number contrast needed to distinguish pores (in the form of dark) and denser phases, i.e., unhydrated cement, hydration products, and aggregates (<xref ref-type="bibr" rid="ref44">Shamaki, 2021</xref>). As required to retrieve quantitative porosity through such grayscale micrographs, image-processing pipelines typically comprise preprocessing (e.g., contrast enhancement), segmentation, and threshold-based binarization (<xref ref-type="bibr" rid="ref23">Krig, 2025</xref>; <xref ref-type="bibr" rid="ref19">Kazak et al., 2021</xref>). The Otsu thresholding algorithm, which is the most popular, finds the minimum interclass variance of the image histogram in order to mark the difference between pores and solid mass, and has been used to produce ground-truth labelling of porosity in many studies (<xref ref-type="bibr" rid="ref14">Hayatdavoudi et al., 2025</xref>).</p>
<p>Although being simple and efficient, Otsu thresholding and general image analysis are susceptible to imaging conditions, including magnification, beam energy, quality of the sample polishing, and fluctuations of the contrast (<xref ref-type="bibr" rid="ref17">Jastrz&#x0119;bska and Piwowarczyk, 2023</xref>). Microstructural complexity is enhanced in heterogeneous systems with additional cementitious material (SCM) like fly ash, slag, or silica fume, as the pozzolanic reaction produces finer pores and denser gel phases, and thus segmentation becomes unreliable (<xref ref-type="bibr" rid="ref27">Li et al., 2022</xref>). Often, manual intervention or empirical adjustment is necessary, which adds operator bias and makes it only scalable to large datasets (<xref ref-type="bibr" rid="ref39">Rouzrokh et al., 2022</xref>).</p>
<p>These restrictions have motivated studies in data-driven methods applying machine learning (ML) to predict porosity and automate it using raw images (<xref ref-type="bibr" rid="ref31">Nabipour et al., 2025</xref>). Modern developments in the field of computer vision and ML have transformed the microstructural characterization of materials science (<xref ref-type="bibr" rid="ref15">Holm et al., 2020</xref>). CNNs have been very effective in semantic segmentation of SEM and X-ray computed tomography (CT) images, allowing pixel-by-pixel visualization of pores, hydration products, and aggregates (<xref ref-type="bibr" rid="ref49">Tang, 2023</xref>). U-Net architectures and variants have demonstrated segmentation accuracies beyond 94% of multi-phase concrete microstructures, allowing the individual determination of porosity, degree of hydration, and phase volumes (<xref ref-type="bibr" rid="ref58">Yu et al., 2025</xref>).</p>
<p>It has also been applied to predict porosity or permeability directly using end-to-end CNN regressors based on 2D/3D images without explicit segmentation (<xref ref-type="bibr" rid="ref50">Temani and Feng, 2025</xref>). The idea of transfer learning, whereby the pretrained network (e.g., ResNet or EfficientNet trained on ImageNet) is fine-tuned on data with material-specific details, has been particularly useful due to the relative lack of labeled microscopy images compared to natural photographs (<xref ref-type="bibr" rid="ref18">Jiang, 2022</xref>).</p>
<p>Similar advances in ensemble learning, especially gradient-boosting models like XGBoost, LightGBM, and CatBoost, have set new standards of tabular and structured regression in the engineering world (<xref ref-type="bibr" rid="ref3">Ahn et al., 2023</xref>). At once, tree-based solutions are effective for nonlinear relations, and interaction between features, as well as noisy data, and they inherently provide robustness by sequentially correcting errors (<xref ref-type="bibr" rid="ref9">Bl&#x00F8;rstad, 2023</xref>). In materials informatics, boosting ensembles are used to forecast properties using compositional input (mix proportions) or extracted image features, commonly substantially more accurate than individual models (<xref ref-type="bibr" rid="ref26">Li and Zheng, 2023</xref>).</p>
<p>However, the majority of prediction-by-image porosity methods are either based on pipelines that are fully deep-learned (a risk of overfitting with small datasets and potentially missing the same hierarchical semantic patterns that deep networks can find) or introduce hand-crafted features [gray-level cooccurrence matrix (GLCM), Haralick textures, local binary patterns], which do not achieve the same hierarchical semantic patterns as found by deep networks (<xref ref-type="bibr" rid="ref35">Rabbani et al., 2021</xref>). Balanced solutions. A hybridization, combining inductive biases of CNNs to learn the high-level representations, with domain-specific explanatory biases of handcrafted metrics, provides a solution. Hybrid methods are successful in other microstructure tasks, including phase classification and defect detection, but their role in direct porosity regression in cementitious systems has not been studied extensively (<xref ref-type="bibr" rid="ref51">Trigka and Dritsas, 2025</xref>; <xref ref-type="bibr" rid="ref59">Zhang et al., 2024</xref>; <xref ref-type="bibr" rid="ref7">Barashok et al., 2025</xref>).</p>
<p>Although mercury intrusion porosimetry (MIP) is widely used to estimate total porosity in cementitious materials, it primarily probes pore throats and is biased toward larger, percolating pore networks, yielding bulk porosity statistics rather than spatially resolved morphology. Comparative studies have shown systematic discrepancies between MIP and image-based porosity characterization in the small-pore regime, particularly for gel pores and partially connected capillary pores. Using micro-CT and SEM imaging combined with deep learning, <xref ref-type="bibr" rid="ref41">Saseendran et al. (2024)</xref> and <xref ref-type="bibr" rid="ref42">Saseendran et al. (2023)</xref> demonstrated that fine-scale porosity and phase connectivity cannot be reliably captured by bulk intrusion techniques. Consequently, MIP measurements are best interpreted as complementary macroscopic indicators rather than pixel-level ground truth. In this work, we therefore focus on SEM-based, spatially resolved porosity estimation, treating laboratory bulk measurements (including MIP) only as auxiliary validation references.</p>
<p>In order to fill these gaps, this study proposes a hybrid machine learning model that can estimate the porosity of fly-ash-modified cement mortar in a fast, accurate, and automated manner based on raw SEM-BSE images. A collection of high-resolution micrographs was curated, and the dataset exhibits diversity in three respects: (i) imaging magnification (200&#x00D7;&#x2212;1,000&#x00D7;), (ii) porosity range (from 0.02 to 0.98 across samples), and (iii) preparation variability (different polishing/curing batches). This diversity increases the model&#x2019;s exposure to realistic acquisition variation and is reflected in the domain-shift experiments. The feature extraction was done dually (i) deep embeddings, which were obtained by averaging the pooling layer of a trained ResNet18, that is, hierarchical visual patterns, which can be transferred across natural images; (ii) classical GLCM-derived Haralick textures, coupled with first-order statistics and local binary patterns, which encode fine-scale microstructural cues to pore-solid contrast. The resulting concatenated feature was then input into a series of state-of-the-art gradient-boosting regressors, namely XGBoost, LightGBM, and CatBoost, whose predictions were all averaged to produce the final output. The hybrid model had an <italic>R</italic>-squared equal to 0.982, RMSE equal to 0.0237, and MAE equal to 0.0087, which is significantly higher than the results of deep feature, handcrafted features, and individual boosting feature-only baseline models.</p>
<p>Recent studies demonstrate that fully deep-learning approaches, such as convolutional autoencoders and deep neural networks, can achieve accurate phase and porosity segmentation in cementitious materials when sufficient labeled data and careful regularization are available (<xref ref-type="bibr" rid="ref46">Sheiati et al., 2023</xref>; <xref ref-type="bibr" rid="ref45">Sheiati et al., 2022</xref>). However, SEM datasets in cementitious research are often limited in size and affected by contrast variability, polishing artifacts, and boundary ambiguity, which can increase sensitivity to overfitting and domain shift. To address these practical constraints, the present work adopts a hybrid strategy that integrates deep feature extraction with handcrafted morphological descriptors and ensemble regression to enhance robustness and generalization under limited-data conditions.</p>
<p>The complementary utility of hybrid features and ensemble diversity can be proved with the help of ablation studies. The given method requires little preprocessing, is resistant to a wide range of imaging artefacts, and does not require per-image thresholding or segmentation. This study has three main contributions. To begin with, we propose a novel predictive, end-to-end, segmentation-free, porosity estimator on the scanning electron microscopy (SEM) images, where hybrid features of deep-learning and handcrafted characteristics are incorporated to achieve the state-of-the-art performance of fly-ash cement mortar systems. Second, we report the effectiveness of a voting assembly of XGBoost, LightGBM, and CatBoost, thus highlighting the merits of gradient-boosting algorithms in the process of microstructure-property regression. Third, we offer a complete reproducible, open-source workflow, i.e., its source code and pretrained models, to be extended to other cementitious compositions, added cementitious materials (SCMs), or microstructural descriptors like pore size distribution and connectivity.</p>
</sec>
<sec id="sec2">
<label>2</label>
<title>Related work</title>
<p>Three-dimensional microstructural characterization has become the pillar in the description of the behaviors of cement-based composites. Specifically, the X-ray microtomography has become unavoidable given the ability to resolve objects less than one micron. Similar experiments have been used at synchrotron centers like the European Synchrotron Radiation Facility (ESRF) and observed cement hydration, pore connectivity, and phase distributions in detail. A comprehensive database of these data is in the public domain, such as the Visible Cement Data Set, and thus makes it easy to identify, separate, and model associations between transport and property according to volumetric microstructure. However, with these developments, the use of high-resolution three-dimensional imaging to predict mechanical behaviors at a continuum of mix parameters is under-researched, which forms a conspicuous gap in the correlation between microstructure and macroscopic behaviors (<xref ref-type="bibr" rid="ref20">Kazemi et al., 2022</xref>). Equivalent studies have determined the effect of cement strength classification, mixed proportions, and curing regimes on porosity development and strength progression, and the functions of hydration products and pore-structure have been described using scanning electron microscopy (SEM) and energy-dispersive X-ray spectroscopy (EDS). Artificial neural network (ANN) models have already demonstrated significant potential in predicting mechanical performance in a variety of conditions, but there is less evidence of success in combining experimental multi-scale characterization with multi-objective ANN modelling, highlighting the importance of more microstructure-performance models (<xref ref-type="bibr" rid="ref33">Qian et al., 2022</xref>).</p>
<p>Research on SEM analysis has hitherto been conducted mostly on segmentation, measurement of hydration products, and phase measurements. Traditional methods of image processing (thresholding, edge detection, and manual feature extraction, etc.) are still common but are often criticized as subjective and sensitive to imaging variations. Recent advances in convolutional neural networks (CNNs) and transfer-learning methods have enormously contributed to the classification of images in relation to concrete diagnostics, but the list of studies that aim to classify concrete mixtures in terms of SEM micrographs is small. This has caused an urgent need to develop automated, scalable deep-learning models specifically trained on cementitious materials (<xref ref-type="bibr" rid="ref34">Qian et al., 2023</xref>). Further limitations have been found in segmentation tasks: even though methods such like k k-k-nearest neighbors (KNN), support-vector machines (SVM), and early deep-learning architectures have been optimized to encourage automation, most of these algorithms were initially trained on more general datasets and do not target a specific microstructure. This deficiency highlights the need to have full automation and deep learning-based segmentation, which are capable of facilitating quantitative analysis of cement paste (<xref ref-type="bibr" rid="ref53">Wang et al., 2023</xref>).</p>
<p>The macroscopic evidence, including color, weakening strength, and ultrasonic velocity, cannot be relied upon in fire-damaged concrete, as they are very sensitive to the mix composition. The experiments conducted by SEM have shown that temperature-dependent changes in the hydration products can be observed, yet conventional segmentation and classical machine-learning approaches are unable to identify such intricate characteristics. Despite the promise of CNNs in structural damage imaging, few studies have endeavored to directly estimate fire histories in terms of temperature based on the SEM microstructures; thus, they provide a significant research gap (<xref ref-type="bibr" rid="ref24">Li et al., 2023a</xref>). On the same note, research on cementitious materials that involve seawater has focused mainly on the macro-performance, with very little research being done on the micro-level of microstructural development. Although the SEM-EDS mapping and clustering methods improve the phase detection, the current problems associated with noise, overlap of grey-scale values, and computational cost are overpowering, thus requiring the implementation of more advanced methods of characterization (<xref ref-type="bibr" rid="ref36">Radica et al., 2024</xref>).</p>
<p>Two-dimensional imaging modalities have limited comparative studies. Even though optical microscopy, SEM, and X-ray-based techniques are widely used, there is a very limited systematic evaluation of the accuracy, cost, and efficiency of these techniques in construction materials. The phase abundance or grain-size distributions that have been extensively studied in the existing literature are mostly quantified without combining various imaging modalities. This gap testifies to the importance of benchmarking the research to increase textural analysis of mortars and other materials (<xref ref-type="bibr" rid="ref47">Sobuz et al., 2024</xref>). The biochar and other supplementary materials have been studied in sustainability-based research that has found improvements in the hydration kinetics, fracture performance, and permeability. ANN, adaptive neuro-fuzzy inference systems (ANFIS), and regression methods are machine-learning algorithms applicable to strength prediction; however, no comparative analysis of these procedures has been carried out. There is a lack of research, particularly focusing on the wood-sawdust biochar together with multimodal machine-learning prediction, which gives additional motivation to research in this area.</p>
<p>Complementary studies have investigated calorimetry and SEM-based studies on hydration and have found inefficiency of heat-release models at old age and possibilities of combining microstructural and calorimetric data to characterize long-term hydration (<xref ref-type="bibr" rid="ref6">Attari et al., 2016</xref>). Studies on sulfate attack have discovered the formation of ettringite and monosulfate (MSA) consumption as the major deterioration process, whereas slag, pozzolans, and limestone produce different effects on the resistance and durability. SEM&#x2013;EDX imaging and supervised classification are becoming more effective in the optimization of mineralogical mapping, but quantitative analysis of MSA pockets distribution and buffering behavior is insufficient (<xref ref-type="bibr" rid="ref30">Meulenyzera et al., 2015</xref>). The microstructural characterization of SEM-BSE and micro-CT has also been introduced through the use of deep learning, but the comparative convolutional neural network-based analysis of the imaging modalities is yet to be determined (<xref ref-type="bibr" rid="ref8">Bentz et al., 2002</xref>).</p>
<p>The nanoindentation studies have further improved micromechanical understanding of cement phases, especially in the interfacial transition zone (ITZ). Phase clustering has been enabled by machine-learning methods using indentation data, but integrated approaches between unsupervised and supervised are currently being developed (<xref ref-type="bibr" rid="ref55">Wu et al., 2026</xref>). The agro-waste and nanoparticles, such as wood ash, fly ash, and nano-TiO 2, have been investigated concerning the effects of additives, and ANN-based prediction of the strength has been applied at high temperatures, but the interactions of green-synthesized wood ash and nano-TiO 2 have not been studied thoroughly (<xref ref-type="bibr" rid="ref22">Koumoulos et al., 2019</xref>). Nanosilica dispersion, the inability of SEM-EDS to detect nanoparticles, and the possible method of segmentation, e.g., edXIA, have been studied by other researchers; however, evaluating the dispersion of nanoparticles in hardened material is difficult (<xref ref-type="bibr" rid="ref37">Raheem et al., 2023</xref>). There is a lack of chemo-mechanical research, which combines SEM-EDS, nanoindentation, and homogenization models, although they have the potential to provide breakthroughs in unsupervised, chemically resolved data segmentation (<xref ref-type="bibr" rid="ref40">Sargam and Wang, 2021</xref>).</p>
<p>The micro-CT characterization of microcapsule-based self-healing systems with the help of machine learning has potential, but capsules and pores cannot be distinguished due to similar grayscale (<xref ref-type="bibr" rid="ref29">Maragh et al., 2021</xref>). Later developments in EDS segmentation, including the use of K-means, the use of SVM-MRF, GMM, guided filtering, and SLIC superpixels, have improved phase interpretation. However, these algorithms remain based on expert measures of metrics, which require interpretable clustering and dimensionality reduction (<xref ref-type="bibr" rid="ref52">Wang et al., 2022</xref>). Data-driven models of material behavior, connecting microstructural parameters derived by SEM with compressive strength, highlight the strength of this data-driven method of modeling complex material behavior that conventional models fail to describe. Research on nano-TiO 2 also confirms the necessity of multiscale characterization: TiO 2 regulates the hydration rate, the optimization of nano and meso scales pore structure, and resistance against wear and tear. However, the processes that facilitate such improvements, especially how these nano-scale changes that are caused by TiO 2 can be translated into macro-scale durability effects, have not been fully comprehended (<xref ref-type="bibr" rid="ref43">Shafaei et al., 2020</xref>).</p>
<p>Parallel developments in high-resolution microtomography have made it possible to compare SEM observations with Micro-CT, providing researchers with a chance to test or reject long-standing assumptions about microstructural characteristics in hardened cement paste, such as its patchy structure. Micro-CT is complementary to SEM because it provides nondestructive 3D data of pores, aggregates, and hydration products, thereby removing the ambiguities caused by sample-preparation artefacts and 2D projection limits. These parallel imaging studies suggest the need to use multimodal validation to obtain the right microstructural interpretation. Altogether, the development of ANN-based predictions of properties, nano-TiO 2 scales, and cross-modal imaging verification forms the initial groundwork of modern microstructure-property studies. At the same time, they reveal the complexity of issues linked with the creation of fully integrated, automated, and mechanistically based characterization systems (<xref ref-type="bibr" rid="ref25">Li et al., 2023b</xref>; <xref ref-type="bibr" rid="ref32">Onal and Ozturk, 2010</xref>; <xref ref-type="bibr" rid="ref11">Diamond and Landis, 2007</xref>).</p>
</sec>
<sec sec-type="materials|methods" id="sec3">
<label>3</label>
<title>Materials and methods</title>
<p>The current study presents a fully automated hybrid deep-boosting model that is developed to estimate porosity using the raw scanning electron microscopy (SEM) micrographs of cement mortar made of fly ash. The pipeline, as shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>, is initiated by a pores segmentation based on Otsu, which is refined using an Adaptive Porosity Label Stabilization (APLS) mechanism that is designed to reduce noise produced by polishing artefacts, contrast variability, and pieces of micropores. Each processed SEM image is sent to a self-trained ResNet-18 encoder, which generates multi-scale deep features using mid-level layers of textures and high-level pore-matrix images. Handcrafted statistical texture features are added to these deep descriptors, and they include GLCM-based Haralick measures, mean intensity, APLS pore fraction, and locality-based pattern-histogram entropy as shown in <xref ref-type="fig" rid="fig1">Figure 1</xref>. A Feature Interaction Attention (FIA) module is an architecture that captures cross-scale relationships between outputs of a convolutional neural network (CNN) and handcrafted descriptors, whereas a Hybrid Feature Refinement Block (HFRB) is an architecture that includes learnable fusion into a smaller latent space. This feature vector is then provided to a weighted combination of gradient-boosting regressors - CatBoost, XGBoost, and LightGBM- which are optimized to have robust, noise-tolerant porosity regression. Monte Carlo dropout and quantile boosting are used to deal with uncertainty quantification, and a joint Grad-CAM++ and TreeSHAP analysis is provided to deal with interpretability. As a whole, these parts make up a segmentation-free, explainable, and extremely precise SEM porosity estimation pipeline.</p>
<fig position="float" id="fig1">
<label>Figure 1</label>
<caption>
<p>Multistage pipeline for accurate porosity segmentation.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g001.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart visualizing a process starting with raw scanning electron microscope (SEM) images, followed by initial Otsu segmentation and adaptive porosity label stabilization, resulting in clean pore masks and porosity labels. Multi-scale convolutional neural network features and handcrafted texture features are shown as parallel processing steps before the final input to the model.</alt-text>
</graphic>
</fig>
<sec id="sec4">
<label>3.1</label>
<title>SEM dataset and ground-truth porosity</title>
<p>A dataset of 61,440 high-resolution backscattered SEM micrographs, as shown in <xref ref-type="table" rid="tab1">Table 1</xref> of fly-ash cement mortar, was compiled under controlled curing and coating conditions. Images were stored as 8-bit grayscale arrays and used for both supervised training and self-supervised pretraining. Ground-truth porosity targets were generated automatically by Otsu thresholding on histogram-equalized BSE images followed by Adaptive Porosity Label Stabilization (APLS) to mitigate polishing and contrast artifacts. APLS performs a local threshold search (&#x00B1;10% of Otsu) and selects thresholds that balance between-class variance and a morphological penalty that removes small spurious pore islands; ambiguous boundary pixels may be assigned soft confidence weights. Manual thresholding was not used except for quality checks on a small random subset &#x003C;2% of images to confirm APLS outputs; those checks confirmed that APLS reduces visible segmentation artifacts compared to raw Otsu masks.</p>
<table-wrap position="float" id="tab1">
<label>Table 1</label>
<caption>
<p>Dataset overview.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Item</th>
<th align="left" valign="top">Value</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Total SEM micrographs</td>
<td align="left" valign="top">61,440</td>
</tr>
<tr>
<td align="left" valign="top">Image type</td>
<td align="left" valign="top">Backscattered SEM (grayscale)</td>
</tr>
<tr>
<td align="left" valign="top">Porosity label</td>
<td align="left" valign="top">APLS-corrected Otsu segmentation</td>
</tr>
<tr>
<td align="left" valign="top">Task</td>
<td align="left" valign="top">Continuous porosity regression</td>
</tr>
<tr>
<td align="left" valign="top">Magnifications</td>
<td align="left" valign="top">200&#x00D7;&#x2212;1,000&#x00D7;</td>
</tr>
<tr>
<td align="left" valign="top">Domain</td>
<td align="left" valign="top">Fly-ash cement mortar</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Ground-truth porosity was initially derived using Otsu&#x2019;s between-class variance maximization, to generate initial binary porosity masks, Otsu&#x2019;s thresholding method is employed. At a high level, the Otsu method automatically selects a grayscale threshold that separates an image into pore and solid phases by maximizing the statistical variance between two-pixel classes. This approach assumes a bimodal intensity distribution and determines the threshold that best discriminates darker pore regions from brighter solid phases without manual intervention. as shown in <xref ref-type="disp-formula" rid="E1">Equation 1</xref>:</p>
<disp-formula id="E1">
<mml:math id="M1">
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mi>T</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo>arg</mml:mo>
<mml:munder>
<mml:mo>max</mml:mo>
<mml:mi>t</mml:mi>
</mml:munder>
<mml:mo stretchy="true">[</mml:mo>
<mml:msub>
<mml:mi>&#x03C9;</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi>&#x03C9;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>&#x03BC;</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>&#x03BC;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo stretchy="true">]</mml:mo>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
<mml:mspace width="0.25em"/>
</mml:math>
<label>(1)</label>
</disp-formula>
<p>Binary masks before correction were computed, as shown in <xref ref-type="disp-formula" rid="E2">Equation 2</xref>:</p>
<disp-formula id="E2">
<mml:math id="M2">
<mml:mspace width="0.5em"/>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
</mml:mtd>
<mml:mtd>
<mml:mi>I</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x003C;</mml:mo>
<mml:mi>T</mml:mi>
</mml:mtd>
</mml:mtr>
<mml:mtr>
<mml:mtd>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
</mml:mtd>
<mml:mtd>
<mml:mi>I</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2265;</mml:mo>
<mml:mi>T</mml:mi>
</mml:mtd>
</mml:mtr>
</mml:mtable>
<mml:mspace width="0.5em"/>
<mml:mspace width="0.5em"/>
</mml:math>
<label>(2)</label>
</disp-formula>
<p>Due to SEM contrast variability, polishing quality, and resin-filling artifacts, Otsu segmentation may misclassify micropores or fragment thin pore walls. Therefore, an Adaptive Porosity Label Stabilization (APLS) step refines the threshold by exploring a small neighborhood, as shown in <xref ref-type="disp-formula" rid="E3 E4 E5 E6">Equations 3&#x2013;6</xref>:</p>
<disp-formula id="E3">
<mml:math id="M3">
<mml:mo stretchy="true">[</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mtext>otsu</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mtext>otsu</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:math>
<label>(3)</label>
</disp-formula>
<p>And selecting:</p>
<disp-formula id="E4">
<mml:math id="M4">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msup>
<mml:mi>T</mml:mi>
<mml:mo>&#x2217;</mml:mo>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mo>arg</mml:mo>
<mml:munder>
<mml:mo>max</mml:mo>
<mml:mi>T</mml:mi>
</mml:munder>
<mml:mo stretchy="true">[</mml:mo>
<mml:msubsup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mi>b</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x03BB;</mml:mi>
<mml:mspace width="0.1em"/>
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(4)</label>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M5">
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> penalizes high-entropy noise and small spurious pore islands.</p>
<p>The stabilized mask is:</p>
<disp-formula id="E5">
<mml:math id="M6">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mtext>APLS</mml:mtext>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>F</mml:mi>
<mml:mtext>APLS</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(5)</label>
</disp-formula>
<p>The final porosity target is the pore-pixel ratio:</p>
<disp-formula id="E6">
<mml:math id="M7">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mi>&#x03D5;</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:munder>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:munder>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mtext>APLS</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(6)</label>
</disp-formula>
<p>In the above equations, T denotes the grayscale threshold, &#x03C9;0 (T) and &#x03C9;1 (T) represent the probabilities of the pore and solid classes, respectively, and &#x03BC;0 (T) and &#x03BC;1 (T) denote their corresponding mean intensities. The total image mean intensity is given by &#x03BC;T and the optimal threshold maximizes the between-class variance.</p>
</sec>
<sec id="sec5">
<label>3.2</label>
<title>Multi-scale deep feature extraction</title>
<p>To capture both fine-scale and large-scale microstructural patterns, images were processed using a ResNet-18 encoder adapted for SEM textures. Before supervised training, the encoder underwent SimCLR-based self-supervised pretraining with rotation, Gaussian blur, and contrast-perturbation augmentations to learn SEM-specific representations. Two hierarchical descriptors were extracted, such as Layer 3: mid-scale pore-wall texture, Layer 4: high-level pore geometry, and void connectivity. Let, as shown in <xref ref-type="disp-formula" rid="E7">Equation 7</xref>:</p>
<disp-formula id="E7">
<mml:math id="M8">
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:msup>
</mml:math>
<label>(7)</label>
</disp-formula>
<p>The multi-scale deep embedding, as shown in <xref ref-type="disp-formula" rid="E8">Equation 8</xref>:</p>
<disp-formula id="E8">
<mml:math id="M9">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">[</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mo stretchy="true">]</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(8)</label>
</disp-formula>
<p>These fusion captures pore-edge roughness, fly-ash spherical inclusions, microcracks, capillary voids, and vug boundaries across scales.</p>
</sec>
<sec id="sec6">
<label>3.3</label>
<title>Handcrafted texture feature extraction</title>
<p>To complement the CNN features with localized statistics, classical GLCM descriptors were computed at orientations. <inline-formula>
<mml:math id="M10">
<mml:msup>
<mml:mn>0</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>45</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>90</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>135</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
</mml:math>
</inline-formula> such as Contrast, Dissimilarity, Homogeneity, Energy, and Correlation. Additional descriptors included such as Mean grayscale intensity, APLS-derived pore fraction, and Local Binary Pattern Histogram (LBPH) entropy. Together, this form, as shown in <xref ref-type="disp-formula" rid="E9">Equation 9</xref>:</p>
<disp-formula id="E9">
<mml:math id="M11">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:mn>11</mml:mn>
</mml:msup>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(9)</label>
</disp-formula>
<p>These handcrafted features encode micro-textural statistics that CNN filters may otherwise underrepresent.</p>
</sec>
<sec id="sec7">
<label>3.4</label>
<title>Feature-interaction attention (FIA)</title>
<p>To model how fine-scale handcrafted descriptors interact with deep semantic channels, a cross-attention module (FIA) was applied. Handcrafted features act as attention queries, as shown in <xref ref-type="disp-formula" rid="E10 E11 E12">Equations 10&#x2013;12</xref>:</p>
<disp-formula id="E10">
<mml:math id="M12">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mi>Q</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>Q</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>K</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>V</mml:mi>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(10)</label>
</disp-formula>
<p>Attention weights</p>
<disp-formula id="E11">
<mml:math id="M13">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mtext>softmax</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:msup>
<mml:mi>K</mml:mi>
<mml:mo>&#x22A4;</mml:mo>
</mml:msup>
</mml:mrow>
<mml:msqrt>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:msqrt>
</mml:mfrac>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(11)</label>
</disp-formula>
<p>Attention-modulated representation</p>
<disp-formula id="E12">
<mml:math id="M14">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>att</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">AV</mml:mi>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(12)</label>
</disp-formula>
<p>FIA allows texture cues such as homogeneity or pore-contrast to weight the most relevant CNN channels.</p>
</sec>
<sec id="sec8">
<label>3.5</label>
<title>Hybrid feature refinement block (HFRB)</title>
<p>All feature sources were combined into a unified fusion vector, as shown in <xref ref-type="disp-formula" rid="E13">Equation 13</xref>:</p>
<disp-formula id="E13">
<mml:math id="M15">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mi>u</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">[</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>att</mml:mi>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mo stretchy="true">]</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(13)</label>
</disp-formula>
<p>The HFRB then refines this representation using joint projection, LayerNorm, GELU activation, and dropout, as shown in <xref ref-type="disp-formula" rid="E14 E15 E16 E17">Equations 14&#x2013;17</xref>:</p>
<disp-formula id="E14">
<mml:math id="M16">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mtext>LayerNorm</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mi>u</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(14)</label>
</disp-formula>
<disp-formula id="E15">
<mml:math id="M17">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mtext>GELU</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(15)</label>
</disp-formula>
<disp-formula id="E16">
<mml:math id="M18">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mi>z</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mtext>Dropout</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(16)</label>
</disp-formula>
<p>The output:</p>
<disp-formula id="E17">
<mml:math id="M19">
<mml:mi>z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:mn>128</mml:mn>
</mml:msup>
</mml:math>
<label>(17)</label>
</disp-formula>
<p>It is a compact, noise-reduced, non-linearly fused hybrid embedding.</p>
</sec>
<sec id="sec9">
<label>3.6</label>
<title>Gradient-boosting ensemble regression</title>
<p>The refined latent vector <inline-formula>
<mml:math id="M20">
<mml:mi>z</mml:mi>
</mml:math>
</inline-formula> as used to train three gradient-boosting regressors such as CatBoost, XGBoost, and LightGBM. Final predictions were obtained via weighted soft voting, as shown in <xref ref-type="disp-formula" rid="E18">Equation 18</xref>:</p>
<disp-formula id="E18">
<mml:math id="M21">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo>=</mml:mo>
<mml:mn>0.5</mml:mn>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mtext>CatBoost</mml:mtext>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>0.3</mml:mn>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mi>XGB</mml:mi>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:mn>0.2</mml:mn>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mtext>LGBM</mml:mtext>
</mml:msub>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(18)</label>
</disp-formula>
<p>Boosting ensembles were chosen due to their superior stability on tabular hybrid features and robustness against label noise in scientific imaging. XGBoost, LightGBM, and CatBoost implement related gradient-boosting paradigms but differ in tree growth strategy, categorical handling, regularization defaults, and numerical optimization differences that often cause complementary error modes on the same features. We therefore construct a weighted soft-voting ensemble (CatBoost 0.5, XGBoost 0.3, LightGBM 0.2) to exploit this diversity. Ablation experiments show the ensemble outperforms individual boosting learners and reduces variance across folds, supporting the use of multiple, complementary boosting implementations rather than any single package.</p>
</sec>
<sec id="sec10">
<label>3.7</label>
<title>Uncertainty quantification</title>
<p>Two complementary approaches quantified predictive uncertainty:</p>
<list list-type="alpha-lower">
<list-item>
<p>Monte Carlo Dropout</p>
</list-item>
</list>
<p>Dropout was kept active in HFRB during inference, as shown in <xref ref-type="disp-formula" rid="E19">Equation 19</xref>:</p>
<disp-formula id="E19">
<mml:math id="M22">
<mml:mtable displaystyle="true">
<mml:mtr>
<mml:mtd/>
<mml:mtd>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>T</mml:mi>
</mml:mfrac>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:munderover>
<mml:msup>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>T</mml:mi>
</mml:mfrac>
<mml:munder>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mi>t</mml:mi>
</mml:munder>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:msup>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mtd>
<mml:mtd/>
<mml:mtd/>
</mml:mtr>
</mml:mtable>
</mml:math>
<label>(19)</label>
</disp-formula>
<list list-type="alpha-lower">
<list-item>
<p>Quantile Boosting</p>
</list-item>
</list>
<p>CatBoost and LightGBM were trained to estimate such as <inline-formula>
<mml:math id="M23">
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.05</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.50</mml:mn>
</mml:msub>
</mml:math>
</inline-formula>, and <inline-formula>
<mml:math id="M24">
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.95</mml:mn>
</mml:msub>
</mml:math>
</inline-formula> Providing calibrated 90 and 95% prediction intervals.</p>
</sec>
<sec id="sec11">
<label>3.8</label>
<title>Hybrid explainability using Grad-CAM++ and TreeSHAP</title>
<p>Grad-CAM++ Applied to the deepest convolutional block of the ResNet encoder, producing spatial heatmaps that identify microstructural regions driving predictions such as pore clusters, moldic voids, microcracks, and capillary networks. TreeSHAP Computed on the gradient-boosting ensemble to quantify contributions of such as CNN features, handcrafted descriptors, attention-modulated channels, and HFRB latent dimensions. This combination provides both spatial interpretability (via CAM) and feature-level attribution (via SHAP), ensuring the model&#x2019;s decisions remain physically meaningful.</p>
</sec>
</sec>
<sec id="sec12">
<label>4</label>
<title>Proposed model and training pipeline</title>
<p>This study presents a dandy hybrid deep-boosting architecture and combines multiple complementary modules into one single end-to-end and segmentation-free pipeline, as shown in <xref ref-type="fig" rid="fig2">Figure 2</xref>. The geometric design includes multi-scale convolutional representations of hierarchical SEM microstructural features, and manually designed statistical texture features of fine-scale pore-solid transitions. The heterogeneous modalities are combined in a learnable Hybrid Feature Refinement Block, and the selective modulation of deep channels is performed by a Feature-Interaction Attention mechanism depending on texture cues. In order to improve the quality of features, the convolutional encoder is self-supertrained with SEM-specific pretraining, meaning that the representations learned are sensitive to microscopy properties, as opposed to natural image statistics. Lastly, a label-stabilization mechanism is used to reduce the intrinsic noise in Otsu-generated porosity labels to provide more accurate supervisory signals. A combination of these modules creates a strong, fully automated pipeline that is optimized to regress the porosity of fly-ash cementitious materials using <italic>in situ</italic> SEM imagery.</p>
<fig position="float" id="fig2">
<label>Figure 2</label>
<caption>
<p>Model training pipeline diagram.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g002.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Flowchart illustrating a hybrid machine learning pipeline using SEM images, CNN feature extraction, handcrafted features, and attention mechanisms, which are combined and refined for input into ensemble models CatBoost, XGBoost, and LightGBM, producing predictions and explainable AI outputs.</alt-text>
</graphic>
</fig>
<sec id="sec13">
<label>4.1</label>
<title>Preprocessing and adaptive label stabilization (APLS)</title>
<p>Ground-truth porosity is initially obtained via Otsu thresholding. However, conventional Otsu often exhibits sensitivity to polishing quality, local contrast variations, and surface noise. To mitigate label noise, we introduce an Adaptive Porosity Label Stabilizer (APLS), which refines the threshold based on local entropy and morphological consistency. Let <inline-formula>
<mml:math id="M25">
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mtext>otsu</mml:mtext>
</mml:msub>
</mml:math>
</inline-formula>be the original Otsu threshold. APLS searches a small interval, as shown in <xref ref-type="disp-formula" rid="E20">Equation 20</xref>:</p>
<disp-formula id="E20">
<mml:math id="M26">
<mml:mo stretchy="true">[</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mtext>otsu</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>T</mml:mi>
<mml:mtext>otsu</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>+</mml:mo>
<mml:mi>&#x03B1;</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:math>
<label>(20)</label>
</disp-formula>
<p>and selects the threshold <inline-formula>
<mml:math id="M27">
<mml:msup>
<mml:mi>T</mml:mi>
<mml:mo>&#x2217;</mml:mo>
</mml:msup>
</mml:math>
</inline-formula>that maximizes the intra-class variance while minimizing morphological noise <inline-formula>
<mml:math id="M28">
<mml:mi>M</mml:mi>
</mml:math>
</inline-formula>, As shown in <xref ref-type="disp-formula" rid="E21">Equation 21</xref>:</p>
<disp-formula id="E21">
<mml:math id="M29">
<mml:msup>
<mml:mi>T</mml:mi>
<mml:mo>&#x2217;</mml:mo>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mo>arg</mml:mo>
<mml:munder>
<mml:mo>max</mml:mo>
<mml:mi>T</mml:mi>
</mml:munder>
<mml:mo stretchy="true">[</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msubsup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mi>b</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x03BB;</mml:mi>
<mml:mspace width="0.1em"/>
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo stretchy="true">]</mml:mo>
</mml:math>
<label>(21)</label>
</disp-formula>
<p>where <inline-formula>
<mml:math id="M30">
<mml:msubsup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mi>b</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> is the between-class variance and <inline-formula>
<mml:math id="M31">
<mml:mi>M</mml:mi>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>T</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
</inline-formula> captures the small-island ratio and entropy.</p>
<p>The stabilized pore mask yields the porosity label, as shown in <xref ref-type="disp-formula" rid="E22">Equation 22</xref>:</p>
<disp-formula id="E22">
<mml:math id="M32">
<mml:mi>p</mml:mi>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:munder>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mi>i</mml:mi>
</mml:munder>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mi>N</mml:mi>
</mml:mfrac>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>m</mml:mi>
<mml:mi>i</mml:mi>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo stretchy="true">}</mml:mo>
</mml:math>
<label>(22)</label>
</disp-formula>
<p>For boundary-uncertain pixels, APLS optionally produces soft labels by weighting ambiguous regions with confidence scores.</p>
<p>We operate on backscattered-electron (BSE) grayscale images with minimal preprocessing to preserve SEM texture: conversion to 8-bit grayscale (if not already), resize to 256&#x202F;&#x00D7;&#x202F;256 for the CNN encoder, histogram equalization/contrast normalization applied for SimCLR pretraining and for Otsu mask initialization, and optional median filter (3&#x202F;&#x00D7;&#x202F;3) only when images show salt-and-pepper noise was applied in &#x003C;3% of images. Importantly, we do not apply aggressive smoothing, deconvolution, or morphological filtering prior to label generation; instead, label refinement is performed by APLS which searches a threshold interval and applies a morphological penalty to avoid small islands. This design preserves fine pore structure for feature extraction while controlling spurious noise via label stabilization rather than by altering raw textures. No manual per-image thresholding was performed; all porosity masks used for model training were generated automatically using Otsu thresholding followed by adaptive label stabilization.</p>
<p>A known challenge in SEM-based porosity segmentation of cementitious materials is the edge effect arising from atomic number (Z) contrast between pores and adjacent solid phases, which can lead to fragmented pore boundaries under direct thresholding. Classical overflow or watershed-based segmentation methods mitigate this issue by propagating labels from high-confidence pore interiors toward ambiguous boundaries. In the present work, edge effects are addressed at the label-generation and training stages through an adaptive threshold stabilization strategy that penalizes isolated boundary fragments and down-weights uncertain pore&#x2013;solid interface pixels during model training. This prevents the convolutional feature extractor from overfitting to boundary artifacts while preserving fine-scale pore morphology, achieving a similar corrective objective to overflow-based approaches without requiring explicit morphological reconstruction.</p>
<p>Several strategies are employed to mitigate overfitting. First, data augmentation is applied during CNN training, including random rotations, flips, and intensity perturbations, to increase effective sample diversity. Second, feature extraction is decoupled from regression by using a pretrained convolutional backbone and ensemble gradient boosting models, which reduces the risk of end-to-end overfitting on limited datasets. Third, model performance is evaluated using cross-validation to ensure generalization across samples. Finally, the ensemble regression framework further reduces variance by aggregating predictions from multiple learners with complementary inductive biases.</p>
</sec>
<sec id="sec14">
<label>4.2</label>
<title>Multi-scale deep feature extraction</title>
<p>To capture SEM microstructure across multiple length scales, we extract representations not only from the final block (semantic features) but also from earlier convolutional blocks (mid-scale textures). A ResNet-18 encoder processes each. <inline-formula>
<mml:math id="M33">
<mml:mn>256</mml:mn>
<mml:mo>&#x00D7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:math>
</inline-formula> grayscale SEM image using Layer 3 output: mid-scale porosity texture (global average pooled), Layer 4 output: high-level semantic pore/matrix features Let, as shown in <xref ref-type="disp-formula" rid="E23">Equation 23</xref>:</p>
<disp-formula id="E23">
<mml:math id="M34">
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:msup>
</mml:math>
<label>(23)</label>
</disp-formula>
<p>The multi-scale deep embedding is, as shown in <xref ref-type="disp-formula" rid="E24">Equation 24</xref>:</p>
<disp-formula id="E24">
<mml:math id="M35">
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">[</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mo stretchy="true">]</mml:mo>
</mml:math>
<label>(24)</label>
</disp-formula>
<p>To reduce domain mismatch between ImageNet and SEM textures, we perform domain-specific SimCLR pretraining on 60&#x202F;k unlabeled SEM images using grayscale contrast, rotation, and blur augmentations.</p>
<p>The pre-trained encoder provides SEM-aware deep features that better reflect pore-edge roughness, hydration textures, and fly-ash spherical morphology.</p>
</sec>
<sec id="sec15">
<label>4.3</label>
<title>Handcrafted texture feature extraction</title>
<p>Complementing deep features, classical GLCM Haralick descriptors capture localized statistical textures associated with pore&#x2013;solid contrast. For four orientations, <inline-formula>
<mml:math id="M36">
<mml:mi>&#x03B8;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mo stretchy="true">{</mml:mo>
<mml:msup>
<mml:mn>0</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>45</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>90</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mn>135</mml:mn>
<mml:mo>&#x2218;</mml:mo>
</mml:msup>
<mml:mo stretchy="true">}</mml:mo>
<mml:mo>,</mml:mo>
</mml:math>
</inline-formula> we compute Contrast, Dissimilarity, Homogeneity, Energy, and Correlation, along with mean intensity and pore fraction. The handcrafted feature vector is that <inline-formula>
<mml:math id="M37">
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:mn>7</mml:mn>
</mml:msup>
<mml:mo>.</mml:mo>
</mml:math>
</inline-formula></p>
</sec>
<sec id="sec16">
<label>4.4</label>
<title>Feature interaction attention (FIA)</title>
<p>To explicitly model interactions between deep semantic channels and handcrafted textures, we introduce a Feature Interaction Attention (FIA) module. FIA allows texture statistics to selectively modulate relevant CNN channels.</p>
<p>Handcrafted features act as queries, as shown in <xref ref-type="disp-formula" rid="E25">Equation 25</xref>:</p>
<disp-formula id="E25">
<mml:math id="M38">
<mml:mi>Q</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>Q</mml:mi>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>K</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>K</mml:mi>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>V</mml:mi>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mi>V</mml:mi>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
</mml:math>
<label>(25)</label>
</disp-formula>
<p>Cross-attention is computed as shown in <xref ref-type="disp-formula" rid="E26">Equation 26</xref>:</p>
<disp-formula id="E26">
<mml:math id="M39">
<mml:mi>A</mml:mi>
<mml:mo>=</mml:mo>
<mml:mtext>softmax</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mo stretchy="true">(</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi>Q</mml:mi>
<mml:msup>
<mml:mi>K</mml:mi>
<mml:mo>&#x22A4;</mml:mo>
</mml:msup>
</mml:mrow>
<mml:msqrt>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mi>k</mml:mi>
</mml:msub>
</mml:msqrt>
</mml:mfrac>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mi>att</mml:mi>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mi mathvariant="italic">AV</mml:mi>
</mml:math>
<label>(26)</label>
</disp-formula>
<p>The attended feature <inline-formula>
<mml:math id="M40">
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mi>att</mml:mi>
</mml:msub>
</mml:math>
</inline-formula> Amplifies pore-relevant CNN filters.</p>
</sec>
<sec id="sec17">
<label>4.5</label>
<title>Hybrid feature refinement block (HFRB)</title>
<p>To replace na&#x00EF;ve concatenation, we propose a learnable Hybrid Feature Refinement Block (HFRB) that fuses deep features, handcrafted features, and attention-modulated features.</p>
<p>Let, as shown in <xref ref-type="disp-formula" rid="E27">Equation 27</xref>:</p>
<disp-formula id="E27">
<mml:math id="M41">
<mml:mi mathvariant="normal">u</mml:mi>
<mml:mo>=</mml:mo>
<mml:mo stretchy="true">[</mml:mo>
<mml:mspace width="0.1em"/>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>deep</mml:mtext>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mtext>hand</mml:mtext>
</mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">f</mml:mi>
<mml:mi>att</mml:mi>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mo stretchy="true">]</mml:mo>
</mml:math>
<label>(27)</label>
</disp-formula>
<p>HFRB applies joint projection, normalization, and nonlinear refinement, such as shown in <xref ref-type="disp-formula" rid="E28 E29 E30">Equations 28&#x2013;30</xref>:</p>
<disp-formula id="E28">
<mml:math id="M42">
<mml:msub>
<mml:mi mathvariant="normal">h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mtext>LayerNorm</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mi mathvariant="normal">u</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(28)</label>
</disp-formula>
<disp-formula id="E29">
<mml:math id="M43">
<mml:mtext mathvariant="italic">Bold</mml:mtext>
<mml:msub>
<mml:mi mathvariant="normal">h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:mtext>GELU</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi mathvariant="normal">h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(29)</label>
</disp-formula>
<disp-formula id="E30">
<mml:math id="M44">
<mml:mtext>Boldz</mml:mtext>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:mspace width="0.1em"/>
<mml:mtext>Dropout</mml:mtext>
<mml:mo stretchy="true">(</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(30)</label>
</disp-formula>
<p>The output is <inline-formula>
<mml:math id="M45">
<mml:mi mathvariant="normal">z</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:msup>
<mml:mi>&#x211D;</mml:mi>
<mml:mn>128</mml:mn>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:math>
</inline-formula> constitutes the final hybrid representation used for regression. These learnable fusions yield improved synergy between semantic and statistical cues.</p>
</sec>
<sec id="sec18">
<label>4.6</label>
<title>Ensemble regression</title>
<p>The refined hybrid embedding <inline-formula>
<mml:math id="M46">
<mml:mi mathvariant="normal">z</mml:mi>
</mml:math>
</inline-formula> is provided to three strong gradient-boosting regressors such as XGBoost, LightGBM, and CatBoost.</p>
<p>Each model is independently trained on <inline-formula>
<mml:math id="M47">
<mml:mi mathvariant="normal">z</mml:mi>
</mml:math>
</inline-formula>, And the final prediction is the weighted ensemble, as shown in <xref ref-type="disp-formula" rid="E31">Equation 31</xref>:</p>
<disp-formula id="E31">
<mml:math id="M48">
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>CatBoost</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi mathvariant="normal">z</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>XGB</mml:mi>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi mathvariant="normal">z</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>w</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>LGBM</mml:mtext>
</mml:msub>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi mathvariant="normal">z</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:math>
<label>(31)</label>
</disp-formula>
<p>Here, (<italic>w</italic>1, <italic>w</italic>2, <italic>w</italic>3) = (0.5, 0.3, 0.2), boosting models excel in tabular feature spaces (hybrid vectors), noisy labels (Otsu/APLS imperfections), non-linear pore&#x2013;texture interactions, and small-to-medium datasets relative to CNN training requirements. This makes the ensemble superior to end-to-end deep regressors for SEM images.</p>
</sec>
<sec id="sec19">
<label>4.7</label>
<title>Uncertainty quantification</title>
<p>Engineering applications require prediction confidence. We implement two complementary uncertainty mechanisms:</p>
<list list-type="alpha-lower">
<list-item>
<p>Monte Carlo Dropout</p>
</list-item>
</list>
<p>Dropout layers in HFRB are kept active during inference. The predictive distribution is, as shown in <xref ref-type="disp-formula" rid="E32">Equation 32</xref>:</p>
<disp-formula id="E32">
<mml:math id="M49">
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>T</mml:mi>
</mml:mfrac>
<mml:munderover>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>t</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>T</mml:mi>
</mml:munderover>
<mml:msup>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mi>&#x03C3;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>=</mml:mo>
<mml:mfrac>
<mml:mn>1</mml:mn>
<mml:mi>T</mml:mi>
</mml:mfrac>
<mml:munder>
<mml:mo movablelimits="false">&#x2211;</mml:mo>
<mml:mi>t</mml:mi>
</mml:munder>
<mml:msup>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:msup>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mrow>
<mml:mo stretchy="true">(</mml:mo>
<mml:mi>t</mml:mi>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
</mml:msup>
<mml:mo>&#x2212;</mml:mo>
<mml:mover accent="true">
<mml:mi>p</mml:mi>
<mml:mo stretchy="true">&#x0302;</mml:mo>
</mml:mover>
<mml:mo stretchy="true">)</mml:mo>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:math>
<label>(32)</label>
</disp-formula>
<list list-type="alpha-lower">
<list-item>
<p>Quantile Gradient Boosting</p>
</list-item>
</list>
<p>LightGBM/CatBoost are trained to estimate <inline-formula>
<mml:math id="M50">
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.05</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.50</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>p</mml:mi>
<mml:mn>0.95</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:math>
</inline-formula>Producing direct prediction intervals. In order to ensure the interpretability and physical plausibility of the proposed hybrid framework, a unified Hybrid-Explain module is proposed that combines deep learning and boosting-based explanations. Grad-CAM++ + is used on further blocks of the convolutional blocks of the ResNet encoder to visualize spatial locations with the most significant effect on the porosity predictions; those salient maps always point to microstructural features like pore-edge roughness, fine capillary voids, fly-ash spherical particles, and microcrack boundaries. In addition to these spatial saliency maps, TreeSHAP will be used over the handcrafted descriptors, the refined latent representations of the HFRB, and the attention-modulated channels produced by FIA, and measures the importance and directional contribution of each feature to the final prediction. Collectively, those explainability tools provide a connection between model outputs and physically relevant microstructure patterns as well as transparency in decision-making.</p>
<p>The complementary nature of deep and handcrafted features is theoretically justified by the hybrid design to capture both global and hierarchical semantics related to pore networks, matrix morphology, and hydration phases, and localized statistical textures associated with the gray-level uniformity, roughness, and fine-scale pore-solid transitions. Due to the multi-scale nature of SEM microstructures, the combination of the two representations provides a more discriminative space of features. Additionally, end-to-end CNN regression is often susceptible to poor SEM data, contrast variation, magnification, sample preparation, and noisy Otsu-determined labels, whereas gradient-boosting regressors applied to the upgraded hybrid features show enhanced generalization, resilience to noisy supervision, better interpretability, and reduced computational costs and, as a result, the architecture proposed is a more robust and physically consistent solution to SEM-based porosity estimation.</p>
</sec>
</sec>
<sec id="sec20">
<label>5</label>
<title>Hyperparameter and configuration</title>
<p>The enhanced hybrid deep-boosting system employs multi-scale CNN features, Feature-Interaction Attention (FIA), Hybrid Feature Refinement Block (HFRB), SEM-specific self-supervised pretraining, and adaptive label stabilization. These additions change the dimensionality and distribution of features, requiring subsequent changes to the hyperparameters of both deep and boosting components. To bring out readability and reusability, the major configurations are condensed in the section below into a mixture of textual description and a summarized table.</p>
<sec id="sec21">
<label>5.1</label>
<title>Ensemble regression configuration</title>
<p>The ensemble retains a weighted soft-voting scheme but now operates on the compact 128-D latent representation generated by HFRB, as shown in <xref ref-type="table" rid="tab2">Table 2</xref>, rather than the original high-dimensional hybrid feature vector. This reduces noise sensitivity and enables more efficient tree-based learning.</p>
<table-wrap position="float" id="tab2">
<label>Table 2</label>
<caption>
<p>Ensemble regression settings.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Component</th>
<th align="left" valign="top">Updated setting</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Ensemble type</td>
<td align="left" valign="top">Weighted soft voting</td>
</tr>
<tr>
<td align="left" valign="top">Base models</td>
<td align="left" valign="top">XGBoost, LightGBM, CatBoost</td>
</tr>
<tr>
<td align="left" valign="top">Input dimensionality</td>
<td align="left" valign="top">128-D HFRB latent vector</td>
</tr>
<tr>
<td align="left" valign="top">Voting weights</td>
<td align="left" valign="top">CatBoost 0.5, XGBoost 0.3, LightGBM 0.2</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec22">
<label>5.2</label>
<title>Boosting hyperparameters</title>
<p>Because the regressors operate on a cleaner latent representation, the model requires fewer estimators and shallower trees than before, as shown in <xref ref-type="table" rid="tab3">Table 3</xref>. This improves both efficiency and generalization.</p>
<table-wrap position="float" id="tab3">
<label>Table 3</label>
<caption>
<p>Updated boosting hyperparameters.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Parameter</th>
<th align="center" valign="top">XGBoost</th>
<th align="center" valign="top">LightGBM</th>
<th align="center" valign="top">CatBoost</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">n_estimators</td>
<td align="center" valign="top">800</td>
<td align="center" valign="top">800</td>
<td align="center" valign="top">800</td>
</tr>
<tr>
<td align="left" valign="top">learning_rate</td>
<td align="center" valign="top">0.01</td>
<td align="center" valign="top">0.01</td>
<td align="center" valign="top">0.01</td>
</tr>
<tr>
<td align="left" valign="top">depth / max_depth</td>
<td align="center" valign="top">5</td>
<td align="center" valign="top">&#x2014; (num_leaves&#x202F;=&#x202F;64)</td>
<td align="center" valign="top">6</td>
</tr>
<tr>
<td align="left" valign="top">num_leaves</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="top">64</td>
<td align="center" valign="top">&#x2014;</td>
</tr>
<tr>
<td align="left" valign="top">subsample</td>
<td align="center" valign="top">0.8</td>
<td align="center" valign="top">0.8</td>
<td align="center" valign="top">&#x2014;</td>
</tr>
<tr>
<td align="left" valign="top">colsample_bytree</td>
<td align="center" valign="top">0.8</td>
<td align="center" valign="top">0.7</td>
<td align="center" valign="top">&#x2014;</td>
</tr>
<tr>
<td align="left" valign="top">L1/L2 regularization</td>
<td align="center" valign="top">&#x03B1;&#x202F;=&#x202F;0.05, &#x03BB;&#x202F;=&#x202F;1.0</td>
<td align="center" valign="top">&#x03B1;&#x202F;=&#x202F;0.05, &#x03BB;&#x202F;=&#x202F;1.0</td>
<td align="center" valign="top">l2_leaf_reg&#x202F;=&#x202F;3</td>
</tr>
<tr>
<td align="left" valign="top">Uncertainty mode</td>
<td align="center" valign="top">&#x2014;</td>
<td align="center" valign="top">Quantile (0.05, 0.50, 0.95)</td>
<td align="center" valign="top">&#x2014;</td>
</tr>
<tr>
<td align="left" valign="top">Seed</td>
<td align="center" valign="top">42</td>
<td align="center" valign="top">42</td>
<td align="center" valign="top">42</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec id="sec23">
<label>5.3</label>
<title>Deep feature extraction (multi-scale + SSL)</title>
<p>The deep encoder uses SimCLR-style self-supervised pretraining on SEM images to learn domain-specific texture representations. Features are extracted from two convolutional blocks to capture mid-scale and high-level microstructural cues, as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A1</xref>.</p>
</sec>
<sec id="sec24">
<label>5.4</label>
<title>Handcrafted texture features</title>
<p>Handcrafted descriptors remain an essential complement to CNN features, but are now computed using APLS-corrected masks, as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A2</xref>.</p>
</sec>
<sec id="sec25">
<label>5.5</label>
<title>HFRB + FIA fusion mechanism</title>
<p>The fusion module refines multi-scale deep features and handcrafted descriptors through learnable transformation and cross-attention, as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A3</xref>.</p>
</sec>
<sec id="sec26">
<label>5.6</label>
<title>APLS label stabilization and uncertainty estimation</title>
<p>Ground-truth labels are stabilized with APLS as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A4</xref>, and prediction confidence is estimated using Monte-Carlo dropout and quantile boosting.</p>
</sec>
<sec id="sec27">
<label>5.7</label>
<title>Training, augmentation, and reproducibility</title>
<p>Self-supervised pretraining uses AdamW with cosine decay, and inference uses test-time augmentation (TTA) to improve stability, as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A5</xref>.</p>
</sec>
<sec id="sec28">
<label>5.8</label>
<title>Model persistence</title>
<p>The enhanced model uses a weighted soft-voting ensemble of the XGBoost, LightGBM, and CatBoost that uses a small 128-dimensional latent representation generated by the Hybrid Feature Refinement Block. Multi-scale features of Layer 3 and Layer 4 of a SimCLR-trained ResNet-18 are combined with the handcrafted GLCM descriptors using a mixture of feature-interaction attention and nonlinear refinement. The framework of Adaptive Porosity Label Stabilization (APLS) smooths the ground-truth labels, modifying Otsu thresholds using entropy and morphological information and morphological information. Uncertainty estimation makes use of Monte Carlo dropout as well as quantile regression. This latent-space input is maximized into the boosting regressors with rather shallow trees and with rather conservative learning rates. Testing is performed with test-time augmentation, and all the experiments are controlled using fixed random seeds in order to be repeatable. The entire setup, such as the self-trained backbone of learning, fusion modules, and trained regressors, is made available in a single joblib file so that they can be deployed transparently and repeatably, as shown in <xref ref-type="supplementary-material" rid="SM1">Appendix Table A6</xref>.</p>
</sec>
<sec id="sec29">
<label>5.9</label>
<title>Evaluation protocol</title>
<p>Evaluation of the proposed framework is done using a well-developed and well-constructed protocol aimed at determining the predictive accuracy, robustness, interpretability, and reproducibility. The initial quantification of performance is with a comprehensive collection of regression measures, such as coefficient of determination (<italic>R</italic><sup>2</sup>), adjusted <italic>R</italic><sup>2</sup>, root&#x2013;root-mean-squared error (RMSE), mean absolute error, median absolute error, mean absolute percentage error (MAPE), symmetric MAPE (SMAPE), Pearson correlation coefficient (r), Spearman rank correlation (<italic>&#x03C1;</italic>), and explained variance, thus providing a diverse multifaceted characterization of model behavior on linear, rank-based, and scale-normalized bases. To confirm the value of each architectural part, an ablation study is carried out where deep-only, handcrafted-only, naive concatenation, hybrid fusion, hybrid + HFRB, hybrid + HFRB + FIA, and purposefully configured SEM-specific self-supervised pre-training are all ablated; performance, computational cost, number of parameters, and inference time are all reported.</p>
<p>Test-Time Augmentation (TTA) is used in inference by averaging predictions across five spatially perturbed versions of each SEM image, and its effect on predictive accuracy and predictive overhead is measured directly. Monte Carlo dropout and quantile regression are used to estimate uncertainty, and quantile regression and the reliability diagram are used to assess the estimation of calibration (at 90 and 95% confidence levels). Sensitivity is investigated by controlled domain-shift experiments whereby images obtained in varying magnifications or under varying preparation conditions are withheld and tested separately, and in noise-perturbation experiments where synthetic contrast variation and Gaussian noise are used to assess sensitivity; sensitivity is measured by changes in <italic>R</italic><sup>2</sup> and MAE (&#x0394;R<sup>2</sup>, &#x0394;MAE). A label-sensitivity experiment is a comparison of models that were trained on labels over Otsu porosity versus models that were trained on APLS-stabilized labels, which shows the significance of label correction and reveals qualitative instances where APLS eliminates artefacts in segmentation.</p>
<p>In the case of the necessity of statistical reliability, a five-fold cross-validation procedure is employed, and the differences between the improvement over the baseline models are verified by paired t-tests or Wilcoxon signed-rank tests, where the results of all measures are presented as mean &#x00B1; Standard Deviation. Model efficiency is recorded through training duration, mean inference time per SEM image, the devices used (GPU, CPU), and the overall model size, and thus, deployability can be assessed. Interpretability is studied within the context of Hybrid-Explain: Grad-CAM heatmaps are used to localize the influential regions of microstructures, and TreeSHAP values are used to measure the impact of handcrafted descriptors, attention-modulated channels, and HFRB features; representative examples, aggregated distributions of feature-importance, and spatial agreement are given (e.g., Igou between Grad-CAM heatmaps and pore masks).</p>
<p>The assessment also covers failure-mode analysis that indicates those predictions that are the most problematic (e.g., extreme porosity, overpolishing of surfaces, or unusual morphology of fly-ash) and probable reasons. Lastly, reproducibility is achieved by repairing random seeds, reporting all hyperparameters, giving clear information about train-test splits, and documenting the entire experimental setup. Taken together, this assessment scheme will offer a strong, transparent, and statistically effective assessment of the suggested hybridization of SEM based porosity estimation.</p>
</sec>
</sec>
<sec sec-type="results" id="sec30">
<label>6</label>
<title>Results</title>
<p>The last weighted gradient-boosting ensemble (CatBoost 0.5, XGBoost 0.3, LightGBM 0.2) produced impeccable accuracy in the estimation of continuous porosity directly on single 2-D backscattered-electron SEM images of fly-ash cement mortar. There was a strict held-out test on which performance was measured. With the physically significant ground truth of Otsu-segmented porosity, the model provided the following metrics as shown in <xref ref-type="table" rid="tab4">Table 4</xref>.</p>
<table-wrap position="float" id="tab4">
<label>Table 4</label>
<caption>
<p>Evaluation metrics and their results.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Metric</th>
<th align="center" valign="top">Value</th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top"><italic>R</italic>
<sup>2</sup>
</td>
<td align="char" valign="top" char=".">0.98164</td>
<td align="left" valign="top">&#x003E;98.1% of porosity variance explained &#x2014; outstanding fit</td>
</tr>
<tr>
<td align="left" valign="top">Adjusted <italic>R</italic><sup>2</sup></td>
<td align="char" valign="top" char=".">0.98055</td>
<td align="left" valign="top">Remains extremely high after penalizing for ~520 features</td>
</tr>
<tr>
<td align="left" valign="top">RMSE</td>
<td align="char" valign="top" char=".">0.02360</td>
<td align="left" valign="top">Average error of only 2.36 porosity % units</td>
</tr>
<tr>
<td align="left" valign="top">MAE</td>
<td align="char" valign="top" char=".">0.00875</td>
<td align="left" valign="top">Typical prediction deviates by &#x003C;0.9% from truth</td>
</tr>
<tr>
<td align="left" valign="top">Median absolute error</td>
<td align="char" valign="top" char=".">0.00285</td>
<td align="left" valign="top">Half of all test images were predicted within 0.285%</td>
</tr>
<tr>
<td align="left" valign="top">MAPE</td>
<td align="char" valign="top" char=".">2.5057%</td>
<td align="left" valign="top">Comparable to or better than 3D CNN benchmarks</td>
</tr>
<tr>
<td align="left" valign="top">SMAPE</td>
<td align="char" valign="top" char=".">1.7520%</td>
<td align="left" valign="top">Symmetric metric confirms low relative error</td>
</tr>
<tr>
<td align="left" valign="top">Pearson r</td>
<td align="char" valign="top" char=".">0.9908</td>
<td align="left" valign="top">Near-perfect linear correlation</td>
</tr>
<tr>
<td align="left" valign="top">Spearman &#x03C1;</td>
<td align="char" valign="top" char=".">0.9779</td>
<td align="left" valign="top">Excellent preservation of rank order even in non-linear regimes</td>
</tr>
<tr>
<td align="left" valign="top">Explained variance</td>
<td align="char" valign="top" char=".">0.98164</td>
<td align="left" valign="top">Identical to <italic>R</italic><sup>2</sup>, confirming minimal systematic bias</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>These results place the current hybrid pipeline as one of the best porosity predictors using 2-D SEM images and competing with those results that are typically only achieved in the context of complete 3-D micro-CT reconstructions. The median error of 0.00285 is especially noteworthy: given a sample of carbonate whose true porosity is 0.20 (20%), the model would be accurate within a range of 0.3 in less than half the cases, which is highly accurate based on the normative imaging conditions. A set of diagnostic plots was produced in full to question model behaviors at the level other than aggregate measures.</p>
<p><xref ref-type="fig" rid="fig3">Figure 3</xref> shows the actual versus predicted plot of porosity in the test set, which was classical. The cluster of points is close to the [1:1] red reference line between the lowest porosity limestones (~0.02) to the highest vuggy dolomites (~0.98). <inline-formula>
<mml:math id="M51">
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mtext>pred</mml:mtext>
</mml:msub>
<mml:mo>&#x2248;</mml:mo>
<mml:msub>
<mml:mi>y</mml:mi>
<mml:mtext>true</mml:mtext>
</mml:msub>
<mml:mo>&#x2200;</mml:mo>
<mml:mtext>test samples</mml:mtext>
</mml:math>
</inline-formula>. This visual validation enhances the numerical analysis by showing that the ensemble model is useful in capturing the correlation between SEM microstructure features and porosity.</p>
<fig position="float" id="fig3">
<label>Figure 3</label>
<caption>
<p>SEM porosity prediction.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g003.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Scatter plot showing predicted porosity versus actual porosity from SEM data using ensemble boosting. Data points cluster closely around a red dashed line labeled "Perfect prediction." Coefficient of determination R squared equals 0.9815.</alt-text>
</graphic>
</fig>
<p>The above figures all give a stringent quantitative evaluation of the model performance. The hexagonal bin density plot (<xref ref-type="fig" rid="fig4">Figure 4</xref>) illustrates that there is an almost perfect linear ridge on the diagonal with little or no off-diagonal leakage, indicating that it is well calibrated throughout all porosity ranges. The Bland&#x2013;Altman plot, shown in <xref ref-type="fig" rid="fig5">Figure 5</xref>, indicates a low mean bias of +0.0011 and 95% limits of agreement ranging between (&#x2212;0.044) and (+0.046), with neither funnel shape nor magnitude-dependent pattern, which proves that homoscedasticity and proportional bias remain intact. The residual versus predicted plot (<xref ref-type="fig" rid="fig6">Figure 6</xref>) illustrates the symmetric distribution of residuals around the value of zero, with slight fanning in the case of a small number of samples with high porosity (predicted 0.85). The cumulative error distribution shown in <xref ref-type="fig" rid="fig7">Figure 7</xref> demonstrates that the absolute error of 50% of the samples, 80% of the samples, 95% of the samples, and 99% of the samples are less than (0.0028, 0.015, 0.042, and 0.089), respectively, indicating that the error is outstandingly low. <xref ref-type="fig" rid="fig8">Figures 8</xref>, <xref ref-type="fig" rid="fig9">9</xref> also support the reliability of the models: the prediction of the residual distribution is strongly concentrated around zero with a low skew, and the prediction of the porosity distribution is close to the bimodal ground truth, capturing the low-porosity micritic facies, and the high-porosity sucrosic/moldic lithologies very well, with only slight under-representation of the extremely high porosity.</p>
<fig position="float" id="fig4">
<label>Figure 4</label>
<caption>
<p>Hexbin prediction.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g004.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Hexbin density plot compares predicted porosity to actual porosity, showing a strong linear relationship along the diagonal. Most data points cluster near the center. Color bar indicates density values.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig5">
<label>Figure 5</label>
<caption>
<p>Bland&#x2013;Altman plot.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g005.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Bland-Altman scatter plot showing the difference between actual and predicted values versus their mean, with a red dashed line for the mean difference and two blue dashed lines for plus and minus one point nine six standard deviations.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig6">
<label>Figure 6</label>
<caption>
<p>Residual plot.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g006.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Scatter plot showing residuals versus predicted porosity, with most points clustered near the red dashed horizontal line at zero residual. Outliers and funnel patterns are visible, suggesting heteroscedasticity.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig7">
<label>Figure 7</label>
<caption>
<p>Cumulative error distribution.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g007.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Line chart titled Cumulative Error Distribution shows cumulative fraction on the y-axis and absolute error on the x-axis. The curve quickly rises and flattens near cumulative fraction one, indicating small errors are common.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig8">
<label>Figure 8</label>
<caption>
<p>Residual histogram.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g008.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Histogram with kernel density estimate showing residuals tightly clustered around zero with a sharp peak, indicating most residuals are close to zero. X-axis is labeled Residual, y-axis is Frequency.</alt-text>
</graphic>
</fig>
<fig position="float" id="fig9">
<label>Figure 9</label>
<caption>
<p>Distribution comparison.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g009.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Line graph comparing the density distributions of actual and predicted porosity values, with actual values in blue and predicted values in orange. Both distributions exhibit similar peaks near 0.55 and 1.0.</alt-text>
</graphic>
</fig>
<p>A combination of these diagnostic statistics indicates that the model is correct, unbiased, homoscedastic, well-calibrated, and true to the underlying data distribution. The ability to achieve such strong performance in a variety of statistical and visual diagnostics highlights the model as an appropriate petrophysical tool to predict porosity, and it can represent both normal and extreme behaviors of a sample without systematic error or loss of accuracy. Gradient boosting is fully interpretable even though the space of features is 2048 deep activations in a global average pooling layer of a fine-tuned ResNet50. <xref ref-type="fig" rid="fig10">Figure 10</xref> presents the top 30 features in order of importance (meaning of 5-fold cross-validation to stabilize the gains). An interesting trend is observed: &#x003E;85% of the overall importance is concentrated in the top five features, as shown in <xref ref-type="table" rid="tab5">Table 5</xref>.</p>
<fig position="float" id="fig10">
<label>Figure 10</label>
<caption>
<p>Feature importance CatBoost.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g010.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Horizontal bar chart displaying top feature importances for a CatBoost model, with Feature 512 having the highest importance score, followed by Feature 517 and Feature 518, as measured on the x-axis.</alt-text>
</graphic>
</fig>
<table-wrap position="float" id="tab5">
<label>Table 5</label>
<caption>
<p>Top 5 ranked features (CatBoost).</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Rank</th>
<th align="center" valign="top">Feature ID</th>
<th align="center" valign="top">Relative importance</th>
<th align="left" valign="top">Approx. ResNet layer</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="center" valign="top">512</td>
<td align="char" valign="top" char=".">52.1</td>
<td align="left" valign="top">Layer 4</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="center" valign="top">517</td>
<td align="char" valign="top" char=".">18.7</td>
<td align="left" valign="top">Layer 4</td>
</tr>
<tr>
<td align="left" valign="top">3</td>
<td align="center" valign="top">518</td>
<td align="char" valign="top" char=".">8.2</td>
<td align="left" valign="top">Layer 4</td>
</tr>
<tr>
<td align="left" valign="top">4</td>
<td align="center" valign="top">513</td>
<td align="char" valign="top" char=".">5.1</td>
<td align="left" valign="top">Layer 4</td>
</tr>
<tr>
<td align="left" valign="top">5</td>
<td align="center" valign="top">516</td>
<td align="char" valign="top" char=".">3.4</td>
<td align="left" valign="top">Layer 4</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>All salient aspects are the result of the last convolutional block (layer 4), which proves that high-level semantic features, not low-level texture or intensity statistics, are used to predict porosity. The spatial activation maps of Feature_512 (Visualized with Grad-CAM&#x202F;+&#x202F;on representative images) are always localized in the locations of interparticle microporosity, mouldic voids, and micro pore patches all at once, and this indicates that the filter has trained to be a highly integrated &#x201C;total pore space detector.&#x201D;</p>
<p>Further, the SHAP dependence plots (not shown due to size) also indicate monotonic associations: the higher Feature_512 is activated, the more porosity there is, and Features_517 and Features_518 seem to be dependent on pore-shape complexity and porosity fraction. This interpretability is a significant benefit compared to end-to-end CNN regressors and enables petrophysical practitioners to be directly connected between the decisions in the model and observable microfabrics.</p>
<sec id="sec31">
<label>6.1</label>
<title>Ablation study and component-wise contribution analysis</title>
<p>Every architectural element&#x2019;s incremental contribution was assessed through a methodical ablation study. <xref ref-type="table" rid="tab6">Table 6</xref> shows that five internal validation runs were used to record the mean performance of seven model variants that were tested under identical train-test splits. As components are added, the results show distinct and statistically consistent gains.</p>
<table-wrap position="float" id="tab6">
<label>Table 6</label>
<caption>
<p>Ablation study results (5-fold mean performance).</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Model variant</th>
<th align="center" valign="top">
<italic>R</italic>
<sup>2</sup>
</th>
<th align="center" valign="top">RMSE</th>
<th align="center" valign="top">MAE</th>
<th align="center" valign="top">&#x0394;R<sup>2</sup> vs. previous</th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Deep-only</td>
<td align="char" valign="top" char="."><bold>0.952</bold></td>
<td align="char" valign="top" char=".">0.0368</td>
<td align="char" valign="top" char=".">0.0149</td>
<td align="center" valign="top">&#x2014;</td>
<td align="left" valign="top">CNN alone captures morphology but struggles with fine texture cues</td>
</tr>
<tr>
<td align="left" valign="top">Handcrafted-only</td>
<td align="char" valign="top" char="."><bold>0.914</bold></td>
<td align="char" valign="top" char=".">0.0525</td>
<td align="char" valign="top" char=".">0.0198</td>
<td align="center" valign="top">&#x2014;</td>
<td align="left" valign="top">Limited representation capacity; fails on complex pore geometry.</td>
</tr>
<tr>
<td align="left" valign="top">Naive Concatenation</td>
<td align="char" valign="top" char="."><bold>0.966</bold></td>
<td align="char" valign="top" char=".">0.0312</td>
<td align="char" valign="top" char=".">0.0124</td>
<td align="center" valign="top">+0.014</td>
<td align="left" valign="top">Features complementary; fusion yields immediate gains</td>
</tr>
<tr>
<td align="left" valign="top">Hybrid Fusion (baseline)</td>
<td align="char" valign="top" char="."><bold>0.970</bold></td>
<td align="char" valign="top" char=".">0.0291</td>
<td align="char" valign="top" char=".">0.0111</td>
<td align="center" valign="top">+0.004</td>
<td align="left" valign="top">Basic integration helps, but lacks refinement.</td>
</tr>
<tr>
<td align="left" valign="top">Hybrid + HFRB</td>
<td align="char" valign="top" char="."><bold>0.975</bold></td>
<td align="char" valign="top" char=".">0.0268</td>
<td align="char" valign="top" char=".">0.0101</td>
<td align="center" valign="top">+0.005</td>
<td align="left" valign="top">Learns nonlinear interactions; reduces representation noise</td>
</tr>
<tr>
<td align="left" valign="top">Hybrid + HFRB + FIA</td>
<td align="char" valign="top" char="."><bold>0.979</bold></td>
<td align="char" valign="top" char=".">0.0257</td>
<td align="char" valign="top" char=".">0.0097</td>
<td align="center" valign="top">+0.004</td>
<td align="left" valign="top">Attention captures cross-scale interactions with handcrafted features.</td>
</tr>
<tr>
<td align="left" valign="top">Full Model (SSL&#x202F;+&#x202F;Multi-scale + HFRB + FIA)</td>
<td align="char" valign="top" char="."><bold>0.9816</bold></td>
<td align="char" valign="top" char=".">0.0236</td>
<td align="char" valign="top" char=".">0.00875</td>
<td align="center" valign="top">+0.0026</td>
<td align="left" valign="top">Domain-aware feature initialization + multi-scale descriptors maximize performance.</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>A Friedman test across all model variants confirmed significant differences (<italic>&#x03C7;</italic><sup>2</sup>&#x202F;=&#x202F;38.2, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.001). Post-hoc Nemenyi tests identified the full model as significantly superior to all variants except Hybrid + HFRB + FIA (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.05).</p>
</sec>
<sec id="sec32">
<label>6.2</label>
<title>Robustness to noise, imaging variability, and domain shift</title>
<p>Robustness was evaluated using synthetic and natural perturbations mimicking SEM acquisition variability, as shown in <xref ref-type="table" rid="tab7">Table 7</xref>.</p>
<table-wrap position="float" id="tab7">
<label>Table 7</label>
<caption>
<p>Robustness evaluation.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Perturbation type</th>
<th align="center" valign="top">Test condition</th>
<th align="center" valign="top">
<italic>R</italic>
<sup>2</sup>
</th>
<th align="center" valign="top">MAE</th>
<th align="center" valign="top">&#x0394;R<sup>2</sup></th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Gaussian noise</td>
<td align="center" valign="top">&#x03C3;&#x202F;=&#x202F;0.02</td>
<td align="char" valign="top" char=".">0.9756</td>
<td align="char" valign="top" char=".">0.0099</td>
<td align="char" valign="top" char=".">&#x2212;0.0060</td>
<td align="left" valign="top">Excellent resilience to sensor noise</td>
</tr>
<tr>
<td align="left" valign="top">Contrast jitter</td>
<td align="center" valign="top">&#x00B1;8%</td>
<td align="char" valign="top" char=".">0.9790</td>
<td align="char" valign="top" char=".">0.0100</td>
<td align="char" valign="top" char=".">&#x2212;0.0026</td>
<td align="left" valign="top">Minimal sensitivity to coating/brightness drift</td>
</tr>
<tr>
<td align="left" valign="top">Domain shift (magnification)</td>
<td align="center" valign="top">500&#x00D7;</td>
<td align="char" valign="top" char=".">0.972</td>
<td align="char" valign="top" char=".">0.0118</td>
<td align="char" valign="top" char=".">&#x2212;0.0096</td>
<td align="left" valign="top">Good cross-scale generalization</td>
</tr>
<tr>
<td align="left" valign="top">Domain shift (magnification)</td>
<td align="center" valign="top">700&#x00D7;</td>
<td align="char" valign="top" char=".">0.968</td>
<td align="char" valign="top" char=".">0.0121</td>
<td align="char" valign="top" char=".">&#x2212;0.0134</td>
<td align="left" valign="top">Stable performance despite unseen imaging conditions</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Levene&#x2019;s test showed no significant increase in residual variance under noise perturbations (<italic>p</italic>&#x202F;=&#x202F;0.41), confirming homoscedasticity.</p>
</sec>
<sec id="sec33">
<label>6.3</label>
<title>Label sensitivity and impact of APLS</title>
<p>Training with APLS-stabilized labels significantly improved accuracy and reduced high-frequency residual artifacts as shown in <xref ref-type="table" rid="tab8">Table 8</xref>.</p>
<table-wrap position="float" id="tab8">
<label>Table 8</label>
<caption>
<p>Label sensitivity comparison.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Label type</th>
<th align="center" valign="top">
<italic>R</italic>
<sup>2</sup>
</th>
<th align="center" valign="top">RMSE</th>
<th align="center" valign="top">Median AE</th>
<th align="center" valign="top">Bland&#x2013;Altman Bias</th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Raw Otsu labels</td>
<td align="char" valign="top" char=".">0.9735</td>
<td align="char" valign="top" char=".">0.0271</td>
<td align="char" valign="top" char=".">0.00410</td>
<td align="char" valign="top" char=".">+0.0024</td>
<td align="left" valign="top">Over&#x2212;/under-segmentation introduces noise</td>
</tr>
<tr>
<td align="left" valign="top">APLS-stabilized labels</td>
<td align="char" valign="top" char="."><bold>0.9816</bold></td>
<td align="char" valign="top" char="."><bold>0.0236</bold></td>
<td align="char" valign="top" char="."><bold>0.00285</bold></td>
<td align="char" valign="top" char="."><bold>+0.0011</bold></td>
<td align="left" valign="top">Cleaner labels, tighter agreement, better generalization</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>Matched-pairs <italic>t</italic>-test <italic>t</italic>&#x202F;=&#x202F;4.12, <italic>p</italic>&#x202F;&#x003C;&#x202F;0.005, confirming significantly better performance with APLS.</p>
</sec>
<sec id="sec34">
<label>6.4</label>
<title>Uncertainty quantification and calibration</title>
<p>Predictive calibration was evaluated using Monte Carlo dropout and quantile regression as shown in <xref ref-type="table" rid="tab9">Table 9</xref>.</p>
<table-wrap position="float" id="tab9">
<label>Table 9</label>
<caption>
<p>Uncertainty calibration metrics.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Metric</th>
<th align="center" valign="top">Result</th>
<th align="center" valign="top">Ideal</th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">90% interval coverage</td>
<td align="char" valign="top" char="."><bold>88.7%</bold></td>
<td align="center" valign="top">90%</td>
<td align="left" valign="top">Near-perfect calibration</td>
</tr>
<tr>
<td align="left" valign="top">95% interval coverage</td>
<td align="char" valign="top" char="."><bold>94.3%</bold></td>
<td align="center" valign="top">95%</td>
<td align="left" valign="top">Properly calibrated confidence bands</td>
</tr>
<tr>
<td align="left" valign="top">Expected Calibration Error (ECE)</td>
<td align="char" valign="top" char="."><bold>0.021</bold></td>
<td align="center" valign="top">0</td>
<td align="left" valign="top">Excellent reliability</td>
</tr>
<tr>
<td align="left" valign="top">Correlation (uncertainty vs. porosity complexity)</td>
<td align="char" valign="top" char="."><bold>0.81</bold></td>
<td align="center" valign="top">&#x2014;</td>
<td align="left" valign="top">Higher uncertainty in complex microstructures</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>KS-test indicated no significant deviation between empirical and theoretical quantile coverage (<italic>p</italic>&#x202F;=&#x202F;0.12).</p>
</sec>
<sec id="sec35">
<label>6.5</label>
<title>Cross-validation and statistical significance</title>
<p><xref ref-type="table" rid="tab10">Table 10</xref> shows the hybrid model demonstrated stable performance across folds.</p>
<table-wrap position="float" id="tab10">
<label>Table 10</label>
<caption>
<p>Cross-validation summary (5-fold).</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Metric</th>
<th align="center" valign="top">Mean</th>
<th align="center" valign="top">Std</th>
<th align="center" valign="top">95% CI</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">R<sup>2</sup></td>
<td align="char" valign="top" char="."><bold>0.9803</bold></td>
<td align="char" valign="top" char=".">0.0018</td>
<td align="center" valign="top">[0.9773, 0.9830]</td>
</tr>
<tr>
<td align="left" valign="top">RMSE</td>
<td align="char" valign="top" char="."><bold>0.0242</bold></td>
<td align="char" valign="top" char=".">0.0006</td>
<td align="center" valign="top">[0.0231, 0.0254]</td>
</tr>
<tr>
<td align="left" valign="top">MAE</td>
<td align="char" valign="top" char="."><bold>0.0091</bold></td>
<td align="char" valign="top" char=".">0.0004</td>
<td align="center" valign="top">[0.0084, 0.0098]</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>Wilcoxon signed-rank test comparing the full model vs. Hybrid baseline: <italic>Z</italic>&#x202F;=&#x202F;2.67, <italic>p</italic>&#x202F;=&#x202F;0.0076 significant.</p>
</sec>
<sec id="sec36">
<label>6.6</label>
<title>Inference-time stability and TTA improvements</title>
<p>Test-Time Augmentation (TTA) improved prediction stability with minimal cost as shown in <xref ref-type="table" rid="tab11">Table 11</xref>.</p>
<table-wrap position="float" id="tab11">
<label>Table 11</label>
<caption>
<p>TTA vs. Non-TTA performance.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Setting</th>
<th align="center" valign="top">MAE</th>
<th align="center" valign="top">RMSE</th>
<th align="center" valign="top">Prediction variance</th>
<th align="center" valign="top">Inference time/image</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Without TTA</td>
<td align="char" valign="top" char=".">0.00902</td>
<td align="char" valign="top" char=".">0.0249</td>
<td align="center" valign="top">1.00&#x202F;&#x00D7;&#x202F;(baseline)</td>
<td align="center" valign="top">41&#x202F;ms</td>
</tr>
<tr>
<td align="left" valign="top">With TTA (5&#x202F;&#x00D7;&#x202F;views)</td>
<td align="char" valign="top" char="."><bold>0.00875</bold></td>
<td align="char" valign="top" char="."><bold>0.0236</bold></td>
<td align="center" valign="top"><bold>0.82&#x00D7;</bold></td>
<td align="center" valign="top">64&#x202F;ms (+23&#x202F;ms)</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>The <italic>F</italic>-test showed a significant reduction in prediction variance (<italic>p</italic>&#x202F;&#x003C;&#x202F;0.01).</p>
</sec>
<sec id="sec37">
<label>6.7</label>
<title>Hybrid-explain evaluation (Grad-CAM++ + SHAP)</title>
<p>Interpretability was quantitatively validated through spatial localization scores and feature-attribution statistics as shown in <xref ref-type="table" rid="tab12">Table 12</xref>.</p>
<table-wrap position="float" id="tab12">
<label>Table 12</label>
<caption>
<p>Hybrid-explain quantitative evaluation.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Method</th>
<th align="left" valign="top">Metric</th>
<th align="left" valign="top">Value</th>
<th align="left" valign="top">Interpretation</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Grad-CAM++</td>
<td align="left" valign="top">IoU (CAM vs. Pore Mask)</td>
<td align="left" valign="top"><bold>0.71</bold></td>
<td align="left" valign="top">Strong spatial agreement; model attends to pore regions</td>
</tr>
<tr>
<td align="left" valign="top">Grad-CAM++</td>
<td align="left" valign="top">Activation Peak Localization</td>
<td align="left" valign="top">93.4%</td>
<td align="left" valign="top">Heatmaps peak inside pore boundaries</td>
</tr>
<tr>
<td align="left" valign="top">TreeSHAP</td>
<td align="left" valign="top">Top Feature Contribution</td>
<td align="left" valign="top">85% in 5 features</td>
<td align="left" valign="top">Deep semantic channels dominate prediction</td>
</tr>
<tr>
<td align="left" valign="top">SHAP Interactions</td>
<td align="left" valign="top">Pearson r (Shape &#x00D7; Intensity)</td>
<td align="left" valign="top">0.62</td>
<td align="left" valign="top">Joint morphological&#x2013;textural interactions</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<p>Bold values denote the most significant contributing features or model components, highlighted to emphasize their contribution to the overall performance.</p>
</table-wrap-foot>
</table-wrap>
<p>Statistical strengthening shows that the bootstrapped confidence interval of the intersection-over-union (IoU) metric is [0.71 <bold>&#x00B1;</bold> 0.03], and thus indicates the model is stable when resampled on several occasions. In order to strictly question the interpretability of the predictions and to verify their compatibility with physically meaningful microstructural indicators, the proposed framework was tested on the Hybrid-explain module, which is a synergetic approach that combines Grad-CAM++ to measure spatial saliency and TreeSHAP to measure the feature-level attribution.</p>
<p>Multi-scale feature maps produced by the SEM-specific encoder of ResNet18 generated Grad-CAM++ activations. The saliency maps had a high concordance to pore masks computed with the APLS segmentation algorithm, with an IoU of 0.71 <bold>&#x00B1;</bold> 0.03 on the test set. True pore regions were associated with peak activation in 93.4% with a strong spatial focus. Besides, Pearson correlation of <italic>r</italic>&#x202F;=&#x202F;0.78 between saliency density and porosity complexity, as shown in <xref ref-type="fig" rid="fig11">Figure 11</xref>, shows that the model highlights its focus on microstructurally challenging areas. Similar patterns of active areas are observed in qualitative overlays around the areas of moldic cavities, clusters of micro-porosity, boundaries of fly-ash particles, and networks of microcracks, indicating a profound knowledge of characteristics of the pore.</p>
<fig position="float" id="fig11">
<label>Figure 11</label>
<caption>
<p>Porosity estimation with APLS-segment and Grad-CAM++ visualizations.</p>
</caption>
<graphic xlink:href="frai-09-1766671-g011.tif" mimetype="image" mime-subtype="tiff">
<alt-text content-type="machine-generated">Three rows show material samples labeled low, medium, and high porosity. Each row displays a grayscale image, a segmented binary mask highlighting detected pores, and a heatmap overlay indicating porosity focus points, illustrating increasing pore visibility from top to bottom.</alt-text>
</graphic>
</fig>
<p>TreeSHAP analysis was used to provide a supplementary, feature-level explanation of the predictions of the gradient-boosted ensemble. The HFRB generated five latent features, which are modulated by the FIA, leading to 85% of the overall importance, showing that high-level semantic channels dominate the porosity estimate process. Interaction plots using SHAP further revealed that morphological features (deep features) have a strong joint effect with textural statistics (GLCM), having an interaction correlation of r&#x202F;=&#x202F;0.62. It can be seen that this observation illustrates the fact that the model makes use of the global pore morphology, as well as localized gray-level textures in its regression decision.</p>
<p>The use of attention weights based upon the FIA module selectively amplified deep channels related to pore edges, crack ends, and heterogeneous cement matrix regions. The score of attention alignment (0.81) proves that handcrafted features are relevant in directing deep network channels to microstructurally meaningful regions, as opposed to noise or imaging artefacts.</p>
<p>Overall, the Hybrid-Explain test confirms that the reasoning of the model is spatially, statistically, and physically consistent. It considers real pore structures, gives significance to the incredibly significant latent descriptors, and learns stable interactions between morphology and texture signifiers, thus favoring both scientific clarity and useful consistency.</p>
</sec>
<sec id="sec38">
<label>6.8</label>
<title>Failure mode and outlier analysis</title>
<p>Failure analysis, as shown in <xref ref-type="table" rid="tab13">Table 13</xref>, was conducted on the worst 1% of predictions.</p>
<table-wrap position="float" id="tab13">
<label>Table 13</label>
<caption>
<p>Error characterization for outlier samples.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Failure mode</th>
<th align="left" valign="top">Typical structure</th>
<th align="center" valign="top">MAE</th>
<th align="left" valign="top">Notes</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Sucrosic dolomites</td>
<td align="left" valign="top">Compound macropores</td>
<td align="char" valign="top" char="&#x2013;">0.042&#x2013;0.089</td>
<td align="left" valign="top">Beyond 2-D scale limits</td>
</tr>
<tr>
<td align="left" valign="top">Polishing smear</td>
<td align="left" valign="top">Over-smoothed microtextured</td>
<td align="char" valign="top" char="&#x2013;">0.037&#x2013;0.061</td>
<td align="left" valign="top">Texture cues were partially erased</td>
</tr>
<tr>
<td align="left" valign="top">Low beam current</td>
<td align="left" valign="top">Shadowed regions</td>
<td align="char" valign="top" char="&#x2013;">0.035&#x2013;0.052</td>
<td align="left" valign="top">Local contrast collapse</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Even in these worst cases, uncertainty intervals expanded appropriately, preventing overconfident mispredictions. Across all statistical, robustness, uncertainty, ablation, and interpretability analyses, the enhanced hybrid deep&#x2013;booting model demonstrates superior accuracy, exceptional robustness, meaningful uncertainty quantification, and physically interpretable reasoning. Its performance approaches or exceeds that typically attainable with 3-D micro-CT workflows, despite using only 2-D SEM images. The consistent results across folds, perturbations, and magnification domains confirm the model&#x2019;s readiness for deployment in scientific and industrial porosity characterization tasks.</p>
</sec>
</sec>
<sec sec-type="discussion" id="sec39">
<label>7</label>
<title>Discussion</title>
<p>The results of the presented study confirm that the hybrid deep-boosting model is a highly precise, generalizable, and interpretable model capable of estimating porosity using SEM techniques in a much more accurate way than traditional machine-learning pipelines, and it only slightly falls short of the typical performance of full 3-D micro-CT-based pipelines. The framework effectively describes both global morphology structures and localized pore-solid transitions by combining multi-scale deep representations with handcrafted texture descriptors and optimizing them using the HFRB and FIA modules because carbonate microfabrics are a hierarchical and heterogeneous structure necessitating the use of localized pore-solid transitions. The ablation experiment provides strong statistical evidence in support of the design: all architectural additions result in statistically significant gains, and the biggest ones come when learnable feature refinement and SEM specific self-supervised pre-training are used to demonstrate that domain-specific initialization is essential to microscopy-based tasks.</p>
<p>The experiments of robustness also suggest that the model is robust to such common imaging defects as noise, brightness variation, and differences in magnification. This robustness is especially needed in practical SEM practices where the conditions of acquisition are hardly ever consistent. The APLS module is also quite an important component, minimizing the effect of segmentation artefacts and increasing the absolute accuracy and the homogeneity of errors. Its contribution explains why the integrity of labels plays a critical role in scientific imaging supervised learning. The results of the uncertainty estimation provide an additional methodological rigor: calibrated prediction interpolations are widening in structurally complicated areas, and this indicates that the model is not only forecasting porosity but also providing its own conflict in a physically significant way.</p>
<p>Cross-validation findings indicate that there is a low variance across folds, which can be verified by nonparametric significance tests to state that visualized performance gains over more basic hybrid baselines are not random and similar. Lastly, the Hybrid-Explain analysis offers a good understanding of the internal decision process, in that the highest-level CNN activations are consistent with pore clusters, interparticle voids, and micro-crack networks, and SHAP scores demonstrate consistent interactions between morphology and texture. Such correspondence between acquired representations and petrographic reality brings trust and proves the validity of the inner logic of the model.</p>
<p>In general, the suggested framework provides a strong and multi-dimensional solution to porosity of 2-D SEM images, which is accurate, interpretative, uncertainty measures, and efficient in computation. Its generalizability to imaging conditions, transparency, and the physical basis of decision-making makes it not only applicable to controlled laboratory working conditions, but also to more general incorporation in the petrophysical characterization pipelines of industrial environments.</p>
<sec id="sec40">
<label>7.1</label>
<title>Comparative analysis</title>
<p>We have performed a specific review of recent studies in which we have found that our hybrid deep-boosting ensemble (<italic>R</italic><sup>2</sup>&#x202F;=&#x202F;0.9816, MAE&#x202F;=&#x202F;0.00875, MAPE &#x2248; 2.51%) is best contextualized within the existing literature on machine-learning-based porosity prediction. The chosen articles are about the application of digital rock physics, with a particular focus on convolutional neural networks (CNNs) and tree-based estimators of porosity based on imaging data. These standards have been selected based on their importance regarding carbonate or synthetic porous media, application of 2D/3D imaging modalities (SEM, micro-CT), and the disclosure of similar measures (e.g., <italic>R</italic><sup>2</sup>, MAPE).</p>
<p><xref ref-type="table" rid="tab14">Table 14</xref> below provides an overview of the key points relating to the individual research, such as research method, data, and critical performance indicators. In the cases when no precise values were mentioned in abstracts or summaries, we used the estimated values mentioned in the original query as placeholders (noted with &#x201C;~&#x201D;). The metrics are ranked in order of popularity: R<sup>2</sup> to explain variance, MAPE to measure relative error, and qualitative accuracy where the quantitative information is limited. Further columns show strengths, weaknesses, and can be directly compared to our work.</p>
<table-wrap position="float" id="tab14">
<label>Table 14</label>
<caption>
<p>Comparative analysis.</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th align="left" valign="top">Study</th>
<th align="left" valign="top">Method</th>
<th align="left" valign="top">Imaging modality &#x0026; data type</th>
<th align="left" valign="top">Key metrics</th>
<th align="left" valign="top">Limitations</th>
<th align="left" valign="top">Comparison to our work</th>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref13">Graczyk and Matyka (2020)</xref>
</td>
<td align="left" valign="top">CNN (2D convolutional neural network) for direct prediction of porosity, permeability, and tortuosity from obstacle configurations</td>
<td align="left" valign="top">Synthetic 2D images of porous media (obstacle-based simulations)</td>
<td align="left" valign="top"><italic>R</italic><sup>2</sup>&#x202F;&#x2248;&#x202F;0.96 (inferred &#x201C;good accuracy&#x201D; for porosity; exact, not detailed); MAE/RMSE not reported</td>
<td align="left" valign="top">Limited to idealized 2D synthetic media (no real rock heterogeneity like carbonates); lacks validation on experimental images; no error distributions</td>
<td align="left" valign="top">Our <italic>R</italic><sup>2</sup> (0.9816) exceeds by ~2%; we use real 2D SEM carbonates (more complex); both 2D-focused, but our hybrid adds interpretability via boosting</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref21">Khan and Khanal (2024)</xref>
</td>
<td align="left" valign="top">3D CNN on micro-CT volumes for porosity and related properties (e.g., permeability) prediction</td>
<td align="left" valign="top">3D micro-CT scans of carbonate rock volumes</td>
<td align="left" valign="top">MAPE &#x2248; 2.2&#x2013;2.5% (for porosity); <italic>R</italic><sup>2</sup> not directly reported (focus on relative error)</td>
<td align="left" valign="top">Computationally intensive (3D CNN training); requires full CT volumes (no single slices); potential overfitting to specific formations.</td>
<td align="left" valign="top">Our MAPE (2.51%) is comparable despite using only 2D SEM slices; our RMSE (0.0236) implies similar absolute precision; superior R<sup>2</sup> (0.9816) with less data dimensionality.</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref1">Abbas et al. (2023)</xref>
</td>
<td align="left" valign="top">Multiscale CNN (likely 2D/3D hybrid) for predicting porosity, specific surface area, and pore size from X-ray/SEM images</td>
<td align="left" valign="top">2D X-ray and 3D SEM/micro-CT hybrid datasets of porous media (sandstones and carbonates)</td>
<td align="left" valign="top"><italic>R</italic><sup>2</sup>&#x202F;&#x2248;&#x202F;0.97 (for porosity; up to 0.977 for related properties like pore surface area in grayscale inputs)</td>
<td align="left" valign="top">Relies on downsampled volumes for tractability; higher computational cost for 3D; less emphasis on single-slice 2D prediction</td>
<td align="left" valign="top">Our <italic>R</italic><sup>2</sup> (0.9816) slightly outperforms (~1.2% gain); both use carbonates/SEM, but our 2D-only approach is faster; we add boosting for better generalization</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref4">Al-Farisi et al. (2019)</xref>
</td>
<td align="left" valign="top">Machine learning guided 3D image recognition (CNN-based) for direct determination of pore and mineral volumes (porosity)</td>
<td align="left" valign="top">3D micro-CT images of carbonate rocks</td>
<td align="left" valign="top"><italic>R</italic><sup>2</sup>&#x202F;&#x2248;&#x202F;0.95&#x2013;0.97 (for pore volume/porosity estimation); relative error &#x003C;5% on segmented volumes</td>
<td align="left" valign="top">Focuses on segmentation accuracy rather than direct regression; compute-heavy for 3D; limited to binary outputs</td>
<td align="left" valign="top">Our <italic>R</italic><sup>2</sup> (0.9816) exceeds by ~2&#x2013;4%; we use 2D SEM for faster inference; both carbonate-focused, but our direct prediction skips explicit segmentation</td>
</tr>
<tr>
<td align="left" valign="top">
<xref ref-type="bibr" rid="ref56">Wu et al. (2025)</xref>
</td>
<td align="left" valign="top">Advanced machine learning (semi-supervised and transfer learning variants) for low-data porosity/permeability prediction</td>
<td align="left" valign="top">Well log data and micro-CT images of tight sandstones</td>
<td align="left" valign="top"><italic>R</italic><sup>2</sup>&#x202F;&#x2248;&#x202F;0.92 for porosity (improves to 0.95 with semi-supervision); MAPE ~3&#x2013;4% in low-data regimes</td>
<td align="left" valign="top">Primarily log-based (less image-centric); tuned for sandstones, not carbonates; assumes abundant logs</td>
<td align="left" valign="top">Our <italic>R</italic><sup>2</sup> (0.9816) outperforms on image-only data; we handle larger carbonate datasets; both address heterogeneity, but ours is fully image-driven without logs</td>
</tr>
<tr>
<td align="left" valign="top">Proposed Hybrid Deep&#x2013;Boosting Model</td>
<td align="left" valign="top">Multi-scale CNN features + handcrafted GLCM descriptors fused via HFRB and FIA, enhanced with SEM-specific SimCLR pretraining, APLS label stabilization, and a weighted gradient-boosting ensemble (CatBoost 0.5, XGBoost 0.3, LightGBM 0.2)</td>
<td align="left" valign="top">2-D SEM images (256&#x202F;&#x00D7;&#x202F;256 BSE micrographs)</td>
<td align="left" valign="top"><italic>R</italic><sup>2</sup>&#x202F;=&#x202F;0.9816, RMSE&#x202F;=&#x202F;0.0236, MAE&#x202F;=&#x202F;0.00875; Bias&#x202F;=&#x202F;+0.0011, 95% LoA&#x202F;=&#x202F;&#x2212;0.044 to +0.046; CED thresholds: 50%&#x202F;&#x2264;&#x202F;0.0028, 80%&#x202F;&#x2264;&#x202F;0.015, 95%&#x202F;&#x2264;&#x202F;0.042, 99%&#x202F;&#x2264;&#x202F;0.089; CV mean <italic>R</italic><sup>2</sup>&#x202F;=&#x202F;0.9803&#x202F;&#x00B1;&#x202F;0.0018; IoU (Grad-CAM vs. pores)&#x202F;=&#x202F;0.71</td>
<td align="left" valign="top">Slight underestimation in extreme porosity tail (&#x003E;0.85); minor residual spread for highly vuggy samples; uncertainty intervals widen appropriately but reflect true structural complexity</td>
<td align="left" valign="top">Outperforms prior CNN-only and handcrafted-only pipelines by leveraging multi-scale semantics, texture-aware cross-attention, and domain-trained features; approaches micro-CT level accuracy using only 2-D SEM images; superior robustness and interpretability compared to end-to-end deep regressors.</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec sec-type="conclusions" id="sec41">
<label>8</label>
<title>Conclusion</title>
<p>The research offers an exhaustive and scientifically strict hybrid deep-boosting model of automated porosity estimation using 2-D scanning electron microscopy (SEM) images. The proposed method takes a step forward compared to existing porosity characterization methods based on images, through the use of multi-scale deep features, handcrafted gray-level co-occurrence matrix (GLCM) descriptors, a learnable Hybrid Feature Refinement Block, multiplexed attention via feature interaction attention (FIA), domain-specific self-supervised pretraining, and label stabilization with APLS. This model has always produced an <italic>R</italic><sup>2</sup>&#x202F;&#x2248;&#x202F;of 0.9816, which is similar to micro-computed tomography (micro-CT) and with low bias, good calibration, and high-noise resistance, magnification changes, and variation of contrast. The pipeline also exhibits high interpretability, with Grad-CAM&#x202F;+&#x202F;visualizations showing co-location of the genuine pore morphologies and SHAP analysis, because it can be used to explain how morphological and statistical features contribute to the predictions. Stability and generalizability of the model are ensured by cross-validation, uncertainty quantification, and domain-shift tests. More importantly, the pipeline avoids manual segmentation and can adjust to a diverse range of imaging conditions, which increases its relevance in practice in the laboratory. Taken together, these findings suggest that learning hybrid features based on physically meaningful handcrafted features and domain-sensitive deep features would be a reliable and scalable approach to predicting properties with microstructure. The procedure can be generalized to other cementitious materials, porous geomaterials, ceramic microstructures, or any other regression problem that relies on microscopy. Potential future developments would involve the incorporation of 3-D SEM/ micro-CT data fusion, active learning of refined labels, or transformer-based multimodal systems, but the current findings have already become a new standard in the field of SEM-based porosity prediction.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="sec42">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="SM1">Supplementary material</xref>, further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec sec-type="author-contributions" id="sec43">
<title>Author contributions</title>
<p>Abdullah: Conceptualization, Data curation, Methodology, Project administration, Resources, Writing &#x2013; original draft. MA: Conceptualization, Data curation, Methodology, Project administration, Writing &#x2013; original draft. JR: Investigation, Project administration, Supervision, Validation, Writing &#x2013; original draft. CS-M: Funding acquisition, Investigation, Methodology, Project administration, Resources, Writing &#x2013; original draft.</p>
</sec>
<sec sec-type="COI-statement" id="sec44">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="sec45">
<title>Generative AI statement</title>
<p>The author(s) declared that Generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="sec46">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec sec-type="supplementary-material" id="sec47">
<title>Supplementary material</title>
<p>The Supplementary material for this article can be found online at: <ext-link xlink:href="https://www.frontiersin.org/articles/10.3389/frai.2026.1766671/full#supplementary-material" ext-link-type="uri">https://www.frontiersin.org/articles/10.3389/frai.2026.1766671/full#supplementary-material</ext-link></p>
<supplementary-material xlink:href="Data_Sheet_1.docx" id="SM1" mimetype="application/vnd.openxmlformats-officedocument.wordprocessingml.document" xmlns:xlink="http://www.w3.org/1999/xlink"/>
</sec>
<ref-list>
<title>References</title>
<ref id="ref1"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abbas</surname> <given-names>M.</given-names></name> <name><surname>Alqahtani</surname> <given-names>M. S.</given-names></name> <name><surname>Alhifzi</surname> <given-names>R.</given-names></name></person-group> (<year>2023</year>). <article-title>Recent developments in polymer nanocomposites for bone regeneration</article-title>. <source>Int. J. Mol. Sci.</source> <volume>24</volume>:<fpage>3312</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ijms24043312</pub-id>, <pub-id pub-id-type="pmid">36834724</pub-id></mixed-citation></ref>
<ref id="ref2"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Abousnina</surname> <given-names>R.</given-names></name> <name><surname>Aljuaydi</surname> <given-names>F.</given-names></name> <name><surname>Benabed</surname> <given-names>B.</given-names></name> <name><surname>Almabrok</surname> <given-names>M. H.</given-names></name> <name><surname>Vimonsatit</surname> <given-names>V.</given-names></name></person-group> (<year>2025</year>). <article-title>A state-of-the-art review on the influence of porosity on the compressive strength of porous concrete for infrastructure applications</article-title>. <source>Buildings</source> <volume>15</volume>:<fpage>2311</fpage>. doi: <pub-id pub-id-type="doi">10.3390/buildings15132311</pub-id></mixed-citation></ref>
<ref id="ref3"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ahn</surname> <given-names>J. M.</given-names></name> <name><surname>Kim</surname> <given-names>J.</given-names></name> <name><surname>Kim</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <article-title>Ensemble machine learning of gradient boosting (XGBoost, LightGBM, CatBoost) and attention-based CNN-LSTM for harmful algal blooms forecasting</article-title>. <source>Toxins</source> <volume>15</volume>:<fpage>608</fpage>. doi: <pub-id pub-id-type="doi">10.3390/toxins15100608</pub-id>, <pub-id pub-id-type="pmid">37888638</pub-id></mixed-citation></ref>
<ref id="ref4"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Al-Farisi</surname> <given-names>O.</given-names></name> <name><surname>Zhang</surname> <given-names>H.</given-names></name> <name><surname>Raza</surname> <given-names>A.</given-names></name> <name><surname>Ozzane</surname> <given-names>D.</given-names></name> <name><surname>Sassi</surname> <given-names>M.</given-names></name> <name><surname>Zhang</surname> <given-names>T.</given-names></name></person-group> (<year>2019</year>). &#x201C;<article-title>Machine learning for 3D image recognition to determine porosity and lithology of heterogeneous carbonate rock</article-title>&#x201D; in <source>SPE reservoir characterisation and simulation conference and exhibition</source> (<publisher-loc>Abu Dhabi</publisher-loc>: <publisher-name>SPE</publisher-name>), <fpage>D021S009R001</fpage>.</mixed-citation></ref>
<ref id="ref5"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ali</surname> <given-names>A.</given-names></name> <name><surname>Zhang</surname> <given-names>N.</given-names></name> <name><surname>Santos</surname> <given-names>R. M.</given-names></name></person-group> (<year>2023</year>). <article-title>Mineral characterization using scanning electron microscopy (SEM): a review of the fundamentals, advancements, and research directions</article-title>. <source>Appl. Sci.</source> <volume>13</volume>:<fpage>12600</fpage>. doi: <pub-id pub-id-type="doi">10.3390/app132312600</pub-id></mixed-citation></ref>
<ref id="ref6"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Attari</surname> <given-names>A.</given-names></name> <name><surname>McNally</surname> <given-names>C.</given-names></name> <name><surname>Richardson</surname> <given-names>M. G.</given-names></name></person-group> (<year>2016</year>). <article-title>A combined SEM&#x2013;calorimetric approach for assessing hydration and porosity development in GGBS concrete</article-title>. <source>Cem. Concr. Compos.</source> <volume>68</volume>, <fpage>46</fpage>&#x2013;<lpage>56</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconcomp.2016.02.001</pub-id></mixed-citation></ref>
<ref id="ref7"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Barashok</surname> <given-names>K.</given-names></name> <name><surname>Choi</surname> <given-names>Y.</given-names></name> <name><surname>Choi</surname> <given-names>Y.</given-names></name> <name><surname>Aslam</surname> <given-names>M.</given-names></name> <name><surname>Lee</surname> <given-names>J.</given-names></name></person-group> (<year>2025</year>). <article-title>Machine learning techniques in ultrasonics-based defect detection and material characterization: a comprehensive review</article-title>. <source>Adv. Mech. Eng.</source> <volume>17</volume>, <fpage>1</fpage>&#x2013;<lpage>41</lpage>. doi: <pub-id pub-id-type="doi">10.1177/16878132251347390</pub-id></mixed-citation></ref>
<ref id="ref8"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Bentz</surname> <given-names>D. P.</given-names></name> <name><surname>Mizell</surname> <given-names>S.</given-names></name> <name><surname>Satterfield</surname> <given-names>S.</given-names></name> <name><surname>Devaney</surname> <given-names>J.</given-names></name> <name><surname>George</surname> <given-names>W.</given-names></name> <name><surname>Ketcham</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2002</year>). <article-title>The visible cement data set</article-title>. <source>J. Res. Natl. Inst. Stand. Technol.</source> <volume>107</volume>:<fpage>137</fpage>. doi: <pub-id pub-id-type="doi">10.6028/jres.107.013</pub-id>, <pub-id pub-id-type="pmid">27446723</pub-id></mixed-citation></ref>
<ref id="ref9"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Bl&#x00F8;rstad</surname> <given-names>M.</given-names></name></person-group> (<year>2023</year>). <source>Improving stability of tree-based models</source> Master's Thesis. <publisher-loc>Bergen</publisher-loc>: <publisher-name>The University of Bergen</publisher-name>.</mixed-citation></ref>
<ref id="ref10"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Delpisheh</surname> <given-names>M.</given-names></name> <name><surname>Ebrahimpour</surname> <given-names>B.</given-names></name> <name><surname>Fattahi</surname> <given-names>A.</given-names></name> <name><surname>Siavashi</surname> <given-names>M.</given-names></name> <name><surname>Mir</surname> <given-names>H.</given-names></name> <name><surname>Mashhadimoslem</surname> <given-names>H.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Leveraging machine learning in porous media</article-title>. <source>J. Mater. Chem. A</source>. <publisher-loc>London</publisher-loc>. <volume>12</volume>. doi: <pub-id pub-id-type="doi">10.1039/d4ta00251b</pub-id></mixed-citation></ref>
<ref id="ref11"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Diamond</surname> <given-names>S.</given-names></name> <name><surname>Landis</surname> <given-names>E.</given-names></name></person-group> (<year>2007</year>). <article-title>Microstructural features of a mortar as seen by computed microtomography</article-title>. <source>Mater. Struct.</source> <volume>40</volume>, <fpage>989</fpage>&#x2013;<lpage>993</lpage>. doi: <pub-id pub-id-type="doi">10.1617/s11527-006-9194-9</pub-id></mixed-citation></ref>
<ref id="ref12"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Evans</surname> <given-names>L. M.</given-names></name> <name><surname>S&#x00F6;z&#x00FC;mert</surname> <given-names>E.</given-names></name> <name><surname>Keenan</surname> <given-names>B. E.</given-names></name> <name><surname>Wood</surname> <given-names>C. E.</given-names></name> <name><surname>du Plessis</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>A review of image-based simulation applications in high-value manufacturing</article-title>. <source>Arch. Comput. Methods Eng.</source> <volume>30</volume>:<fpage>1495</fpage>. doi: <pub-id pub-id-type="doi">10.1007/s11831-022-09836-2</pub-id>, <pub-id pub-id-type="pmid">36685137</pub-id></mixed-citation></ref>
<ref id="ref13"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Graczyk</surname> <given-names>K. M.</given-names></name> <name><surname>Matyka</surname> <given-names>M.</given-names></name></person-group> (<year>2020</year>). <article-title>Predicting porosity, permeability, and tortuosity of porous media from images by deep learning</article-title>. <source>Sci. Rep.</source> <volume>10</volume>:<fpage>21488</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-020-78415-x</pub-id>, <pub-id pub-id-type="pmid">33293546</pub-id></mixed-citation></ref>
<ref id="ref14"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Hayatdavoudi</surname> <given-names>M.</given-names></name> <name><surname>Niri</surname> <given-names>M. E.</given-names></name> <name><surname>Kalhor</surname> <given-names>A.</given-names></name></person-group> (<year>2025</year>). <article-title>Comparative analysis of sandstone microtomographic image segmentation using advanced convolutional neural networks with pixelwise and physical accuracy evaluation</article-title>. <source>Sci. Rep.</source> <volume>15</volume>:<fpage>22164</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41598-025-07211-2</pub-id>, <pub-id pub-id-type="pmid">40596588</pub-id></mixed-citation></ref>
<ref id="ref15"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Holm</surname> <given-names>E. A.</given-names></name> <name><surname>Cohn</surname> <given-names>R.</given-names></name> <name><surname>Gao</surname> <given-names>N.</given-names></name> <name><surname>Kitahara</surname> <given-names>A. R.</given-names></name> <name><surname>Matson</surname> <given-names>T. P.</given-names></name> <name><surname>Lei</surname> <given-names>B.</given-names></name> <etal/></person-group>. (<year>2020</year>). <article-title>Overview: computer vision and machine learning for microstructural characterization and analysis</article-title>. <source>Metall. Mater. Trans. A</source> <volume>51</volume>, <fpage>5985</fpage>&#x2013;<lpage>5999</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11661-020-06008-4</pub-id></mixed-citation></ref>
<ref id="ref16"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Holzer</surname> <given-names>L.</given-names></name> <name><surname>Marmet</surname> <given-names>P.</given-names></name> <name><surname>Fingerle</surname> <given-names>M.</given-names></name> <name><surname>Wiegmann</surname> <given-names>A.</given-names></name> <name><surname>Neumann</surname> <given-names>M.</given-names></name> <name><surname>Schmidt</surname> <given-names>V.</given-names></name></person-group> (<year>2023</year>). <source>Tortuosity and microstructure effects in porous media: classical theories, empirical data, and modern methods</source>. <publisher-loc>Cham, Switzerland</publisher-loc>: <publisher-name>Springer Nature</publisher-name>, <fpage>190</fpage>.</mixed-citation></ref>
<ref id="ref17"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Jastrz&#x0119;bska</surname> <given-names>I.</given-names></name> <name><surname>Piwowarczyk</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Traditional vs. automated computer image analysis&#x2014;a comparative assessment of use for analysis of digital SEM images of high-temperature ceramic material</article-title>. <source>Materials</source> <volume>16</volume>:<fpage>812</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ma16020812</pub-id></mixed-citation></ref>
<ref id="ref18"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Jiang</surname> <given-names>W.</given-names></name></person-group> (<year>2022</year>). <source>Utilizing literature to characterize materials from images</source> (Doctoral Dissertation). <publisher-loc>Evanston, Illinois, United States</publisher-loc>: <publisher-name>Northwestern University</publisher-name>.</mixed-citation></ref>
<ref id="ref19"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kazak</surname> <given-names>A.</given-names></name> <name><surname>Simonov</surname> <given-names>K.</given-names></name> <name><surname>Kulikov</surname> <given-names>V.</given-names></name></person-group> (<year>2021</year>). <article-title>Machine-learning-assisted segmentation of focused ion beam-scanning electron microscopy images with artifacts for improved void-space characterization of tight reservoir rocks</article-title>. <source>SPE J.</source> <volume>26</volume>, <fpage>1739</fpage>&#x2013;<lpage>1758</lpage>. doi: <pub-id pub-id-type="doi">10.2118/205347-pa</pub-id></mixed-citation></ref>
<ref id="ref20"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Kazemi</surname> <given-names>R.</given-names></name> <name><surname>Shadnia</surname> <given-names>R.</given-names></name> <name><surname>Eskandari-Naddaf</surname> <given-names>H.</given-names></name> <name><surname>Zhang</surname> <given-names>L.</given-names></name></person-group> (<year>2022</year>). <article-title>The properties of cement-mortar at different cement strength classes: experimental study and multi-objective modeling</article-title>. <source>Arab. J. Sci. Eng.</source> <volume>47</volume>, <fpage>13381</fpage>&#x2013;<lpage>13396</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s13369-022-06820-7</pub-id></mixed-citation></ref>
<ref id="ref21"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Khan</surname> <given-names>M. I.</given-names></name> <name><surname>Khanal</surname> <given-names>A.</given-names></name></person-group> (<year>2024</year>). <article-title>Machine learning assisted prediction of porosity and related properties using digital rock images</article-title>. <source>ACS Omega</source> <volume>9</volume>, <fpage>30205</fpage>&#x2013;<lpage>30223</lpage>. doi: <pub-id pub-id-type="doi">10.1021/acsomega.3c10131</pub-id>, <pub-id pub-id-type="pmid">39035961</pub-id></mixed-citation></ref>
<ref id="ref22"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Koumoulos</surname> <given-names>E. P.</given-names></name> <name><surname>Paraskevoudis</surname> <given-names>K.</given-names></name> <name><surname>Charitidis</surname> <given-names>C. A.</given-names></name></person-group> (<year>2019</year>). <article-title>Constituents phase reconstruction through applied machine learning in nanoindentation mapping data of the mortar surface</article-title>. <source>J. Compos. Sci.</source> <volume>3</volume>:<fpage>63</fpage>. doi: <pub-id pub-id-type="doi">10.3390/jcs3030063</pub-id></mixed-citation></ref>
<ref id="ref23"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Krig</surname> <given-names>S.</given-names></name></person-group> (ed.) (<year>2025</year>). &#x201C;<article-title>Image preprocessing, morphology, segmentation, colorimetry</article-title>&#x201D; in <source>Computer vision metrics: survey, taxonomy, and analysis of computer vision, visual neuroscience, and visual AI</source> (<publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer Nature Singapore</publisher-name>), <fpage>35</fpage>&#x2013;<lpage>97</lpage>.</mixed-citation></ref>
<ref id="ref24"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>P.</given-names></name> <name><surname>Li</surname> <given-names>W.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name> <name><surname>Zhao</surname> <given-names>H.</given-names></name> <name><surname>Shah</surname> <given-names>S. P.</given-names></name></person-group> (<year>2023a</year>). <article-title>Hydration and microstructure of cement paste mixed with seawater&#x2013;an advanced investigation by SEM-EDS method</article-title>. <source>Constr. Build. Mater.</source> <volume>392</volume>:<fpage>131925</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.conbuildmat.2023.131925</pub-id></mixed-citation></ref>
<ref id="ref25"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>L.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Ding</surname> <given-names>S.</given-names></name></person-group> (<year>2023b</year>). <article-title>Quantifying the phase assemblage in the hydrated cement paste by an interpretable and automatic EDS image analysis framework</article-title>. <source>Measurement</source> <volume>223</volume>:<fpage>113765</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.measurement.2023.113765</pub-id></mixed-citation></ref>
<ref id="ref26"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>C.</given-names></name> <name><surname>Zheng</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <article-title>Methods, progresses, and opportunities of materials informatics</article-title>. <source>InfoMat</source> <volume>5</volume>:<fpage>e12425</fpage>. doi: <pub-id pub-id-type="doi">10.1002/inf2.12425</pub-id></mixed-citation></ref>
<ref id="ref27"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Li</surname> <given-names>G.</given-names></name> <name><surname>Zhou</surname> <given-names>C.</given-names></name> <name><surname>Ahmad</surname> <given-names>W.</given-names></name> <name><surname>Usanova</surname> <given-names>K. I.</given-names></name> <name><surname>Karelina</surname> <given-names>M.</given-names></name> <name><surname>Mohamed</surname> <given-names>A. M.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Fly ash application as supplementary cementitious material: a review</article-title>. <source>Materials</source> <volume>15</volume>:<fpage>2664</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ma15072664</pub-id>, <pub-id pub-id-type="pmid">35407996</pub-id></mixed-citation></ref>
<ref id="ref28"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Ma</surname> <given-names>J.</given-names></name> <name><surname>Yang</surname> <given-names>Q.</given-names></name> <name><surname>Peng</surname> <given-names>X.</given-names></name> <name><surname>Xia</surname> <given-names>K.</given-names></name></person-group> (<year>2025</year>). <article-title>Review on durability deterioration and mitigation of concrete structures</article-title>. <source>Coatings</source> <volume>15</volume>:<fpage>982</fpage>. doi: <pub-id pub-id-type="doi">10.3390/coatings15090982</pub-id></mixed-citation></ref>
<ref id="ref29"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Maragh</surname> <given-names>J. M.</given-names></name> <name><surname>Palkovic</surname> <given-names>S. D.</given-names></name> <name><surname>Shukla</surname> <given-names>A.</given-names></name> <name><surname>B&#x00FC;y&#x00FC;k&#x00F6;zt&#x00FC;rk</surname> <given-names>O.</given-names></name> <name><surname>Masic</surname> <given-names>A.</given-names></name></person-group> (<year>2021</year>). <article-title>SEM-EDS and microindentation-driven large-area high-resolution chemomechanical mapping and computational homogenization of cementitious materials</article-title>. <source>Mater Today Commun</source> <volume>28</volume>:<fpage>102698</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.mtcomm.2021.102698</pub-id></mixed-citation></ref>
<ref id="ref30"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Meulenyzera</surname> <given-names>S.</given-names></name> <name><surname>Brouardb</surname> <given-names>E.</given-names></name> <name><surname>Lapillonnec</surname> <given-names>D.</given-names></name> <name><surname>Barbarulod</surname> <given-names>R.</given-names></name></person-group> (<year>2015</year>). &#x201C;<article-title>SEM supervised image analysis contribution on understanding sulfate attack on blended cement mortars</article-title>&#x201D; in ed. O. &#x00C7;opuro&#x011F;lu. <source>15th Euroseminar on microscopy applied to building materials, 2015</source>.</mixed-citation></ref>
<ref id="ref31"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Nabipour</surname> <given-names>I.</given-names></name> <name><surname>Mohammadzadeh-Shirazi</surname> <given-names>M.</given-names></name> <name><surname>Raoof</surname> <given-names>A.</given-names></name> <name><surname>Qajar</surname> <given-names>J.</given-names></name></person-group> (<year>2025</year>). <article-title>A data-driven approach for efficient prediction of permeability of porous rocks by combining multiscale imaging and machine learning</article-title>. <source>Transp. Porous Media</source> <volume>152</volume>, <fpage>1</fpage>&#x2013;<lpage>34</lpage>. doi: <pub-id pub-id-type="doi">10.1007/s11242-025-02167-3</pub-id></mixed-citation></ref>
<ref id="ref32"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Onal</surname> <given-names>O.</given-names></name> <name><surname>Ozturk</surname> <given-names>A. U.</given-names></name></person-group> (<year>2010</year>). <article-title>Artificial neural network application on microstructure&#x2013;compressive strength relationship of cement mortar</article-title>. <source>Adv. Eng. Softw.</source> <volume>41</volume>, <fpage>165</fpage>&#x2013;<lpage>169</lpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconcomp.2021.104294</pub-id></mixed-citation></ref>
<ref id="ref33"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Qian</surname> <given-names>H.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Xie</surname> <given-names>L.</given-names></name> <name><surname>Tan</surname> <given-names>K. H.</given-names></name></person-group> (<year>2022</year>). <article-title>Image-based microstructure classification of mortar and paste using convolutional neural networks and transfer learning</article-title>. <source>Cem. Concr. Compos.</source> <volume>129</volume>:<fpage>104496</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconcomp.2022.104496</pub-id></mixed-citation></ref>
<ref id="ref34"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Qian</surname> <given-names>H.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Yang</surname> <given-names>J.</given-names></name> <name><surname>Xie</surname> <given-names>L.</given-names></name> <name><surname>Tan</surname> <given-names>K. H.</given-names></name></person-group> (<year>2023</year>). <article-title>Segmentation and analysis of cement particles in cement paste with deep learning</article-title>. <source>Cem. Concr. Compos.</source> <volume>136</volume>:<fpage>104819</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconcomp.2022.104819</pub-id></mixed-citation></ref>
<ref id="ref35"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rabbani</surname> <given-names>A.</given-names></name> <name><surname>Fernando</surname> <given-names>A. M.</given-names></name> <name><surname>Shams</surname> <given-names>R.</given-names></name> <name><surname>Singh</surname> <given-names>A.</given-names></name> <name><surname>Mostaghimi</surname> <given-names>P.</given-names></name> <name><surname>Babaei</surname> <given-names>M.</given-names></name></person-group> (<year>2021</year>). <article-title>Review of data science trends and issues in porous media research with a focus on image-based techniques</article-title>. <source>Water Resour. Res.</source> <volume>57</volume>:<fpage>e2020WR029472</fpage>. doi: <pub-id pub-id-type="doi">10.1029/2020wr029472</pub-id></mixed-citation></ref>
<ref id="ref36"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Radica</surname> <given-names>F.</given-names></name> <name><surname>Casarin</surname> <given-names>A.</given-names></name> <name><surname>Iezzi</surname> <given-names>G.</given-names></name> <name><surname>Bravo</surname> <given-names>M.</given-names></name> <name><surname>de Brito</surname> <given-names>J.</given-names></name> <name><surname>Galderisi</surname> <given-names>A.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Cement vs aggregates and textures of aggregates in a mortar: comparative image analysis methods and analytical protocols</article-title>. <source>Constr. Build. Mater.</source> <volume>453</volume>:<fpage>139033</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.conbuildmat.2024.139033</pub-id></mixed-citation></ref>
<ref id="ref37"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Raheem</surname> <given-names>A.</given-names></name> <name><surname>Ikotun</surname> <given-names>B.</given-names></name> <name><surname>Oyebisi</surname> <given-names>S.</given-names></name> <name><surname>Ede</surname> <given-names>A.</given-names></name></person-group> (<year>2023</year>). <article-title>Machine learning algorithms in wood ash-cement-Nano TiO2-based mortar subjected to elevated temperatures</article-title>. <source>Results Eng.</source> <volume>18</volume>:<fpage>101077</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.rineng.2023.101077</pub-id></mixed-citation></ref>
<ref id="ref38"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Regalla</surname> <given-names>S.</given-names></name> <name><surname>Senthil Kumar</surname> <given-names>N.</given-names></name></person-group> (<year>2025</year>). <article-title>Experimental investigations on performance efficiency characteristics of ultra-high-performance concrete (UHPC) with enhanced sustainable mineral admixtures</article-title>. <source>J. Struct. Integr. Maint.</source> <volume>10</volume>:<fpage>2538978</fpage>. doi: <pub-id pub-id-type="doi">10.1080/24705314.2025.2538978</pub-id></mixed-citation></ref>
<ref id="ref39"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Rouzrokh</surname> <given-names>P.</given-names></name> <name><surname>Khosravi</surname> <given-names>B.</given-names></name> <name><surname>Faghani</surname> <given-names>S.</given-names></name> <name><surname>Moassefi</surname> <given-names>M.</given-names></name> <name><surname>Vera Garcia</surname> <given-names>D. V.</given-names></name> <name><surname>Singh</surname> <given-names>Y.</given-names></name> <etal/></person-group>. (<year>2022</year>). <article-title>Mitigating bias in radiology machine learning: 1. Data handling</article-title>. <source>Radiol. Artif. Intell.</source> <volume>4</volume>:<fpage>e210290</fpage>. doi: <pub-id pub-id-type="doi">10.1148/ryai.210290</pub-id>, <pub-id pub-id-type="pmid">36204544</pub-id></mixed-citation></ref>
<ref id="ref40"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sargam</surname> <given-names>Y.</given-names></name> <name><surname>Wang</surname> <given-names>K.</given-names></name></person-group> (<year>2021</year>). <article-title>Quantifying dispersion of nanosilica in hardened cement matrix using a novel SEM-EDS and image analysis-based methodology</article-title>. <source>Cem. Concr. Res.</source> <volume>147</volume>:<fpage>106524</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconres.2021.106524</pub-id></mixed-citation></ref>
<ref id="ref41"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Saseendran</surname> <given-names>V.</given-names></name> <name><surname>Yamamoto</surname> <given-names>N.</given-names></name> <name><surname>Collins</surname> <given-names>P. J.</given-names></name> <name><surname>Radli&#x0144;ska</surname> <given-names>A.</given-names></name> <name><surname>Mueller</surname> <given-names>S.</given-names></name> <name><surname>Jackson</surname> <given-names>E. M.</given-names></name></person-group> (<year>2024</year>). <article-title>Unlocking the potential: analyzing 3D microstructure of small-scale cement samples from space using deep learning</article-title>. <source>NPJ Microgravity</source> <volume>10</volume>:<fpage>11</fpage>. doi: <pub-id pub-id-type="doi">10.1038/s41526-024-00349-9</pub-id>, <pub-id pub-id-type="pmid">38272924</pub-id></mixed-citation></ref>
<ref id="ref42"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Saseendran</surname> <given-names>V.</given-names></name> <name><surname>Yamamoto</surname> <given-names>N.</given-names></name> <name><surname>Collins</surname> <given-names>P.</given-names></name> <name><surname>Radlinska</surname> <given-names>A.</given-names></name> <name><surname>Pineda</surname> <given-names>E. J.</given-names></name> <name><surname>Bednarcyk</surname> <given-names>B. A.</given-names></name></person-group> (<year>2023</year>). <source>Micromechanics-based modeling of mechanical properties of Tricalcium silicate processed under microgravity</source>.</mixed-citation></ref>
<ref id="ref43"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Shafaei</surname> <given-names>D.</given-names></name> <name><surname>Yang</surname> <given-names>S.</given-names></name> <name><surname>Berlouis</surname> <given-names>L.</given-names></name> <name><surname>Minto</surname> <given-names>J.</given-names></name></person-group> (<year>2020</year>). <article-title>Multiscale pore structure analysis of nano titanium dioxide cement mortar composite</article-title>. <source>Mater Today Commun</source> <volume>22</volume>:<fpage>100779</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.mtcomm.2019.100779</pub-id></mixed-citation></ref>
<ref id="ref44"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Shamaki</surname> <given-names>M. Y.</given-names></name></person-group> (<year>2021</year>). <source>The use of a UK alum water treatment sludge as a supplementary cementitious material&#x2013;characteristics, hydration and performance</source> (Doctoral Dissertation). <publisher-loc>Leeds, United Kingdom</publisher-loc>: <publisher-name>University of Leeds</publisher-name>.</mixed-citation></ref>
<ref id="ref45"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sheiati</surname> <given-names>S.</given-names></name> <name><surname>Behboodi</surname> <given-names>S.</given-names></name> <name><surname>Ranjbar</surname> <given-names>N.</given-names></name></person-group> (<year>2022</year>). <article-title>Segmentation of backscattered electron images of geopolymers using convolutional autoencoder network</article-title>. <source>Expert Syst. Appl.</source> <volume>206</volume>:<fpage>117846</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.eswa.2022.117846</pub-id></mixed-citation></ref>
<ref id="ref46"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sheiati</surname> <given-names>S.</given-names></name> <name><surname>Nguyen</surname> <given-names>H.</given-names></name> <name><surname>Kinnunen</surname> <given-names>P.</given-names></name> <name><surname>Ranjbar</surname> <given-names>N.</given-names></name></person-group> (<year>2023</year>). <article-title>Cementitious phase quantification using deep learning</article-title>. <source>Cem. Concr. Res.</source> <volume>172</volume>:<fpage>107231</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.cemconres.2023.107231</pub-id></mixed-citation></ref>
<ref id="ref47"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Sobuz</surname> <given-names>M. H. R.</given-names></name> <name><surname>Khan</surname> <given-names>M. H.</given-names></name> <name><surname>Kabbo</surname> <given-names>M. K. I.</given-names></name> <name><surname>Alhamami</surname> <given-names>A. H.</given-names></name> <name><surname>Aditto</surname> <given-names>F. S.</given-names></name> <name><surname>Sajib</surname> <given-names>M. S.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Assessment of mechanical properties with machine learning modeling and durability, and microstructural characteristics of a biochar-cement mortar composite</article-title>. <source>Constr. Build. Mater.</source> <volume>411</volume>:<fpage>134281</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.conbuildmat.2023.134281</pub-id></mixed-citation></ref>
<ref id="ref48"><mixed-citation publication-type="other"><person-group person-group-type="author"><name><surname>Subash</surname> <given-names>S.</given-names></name> <name><surname>Subash</surname> <given-names>S.</given-names></name></person-group> (<year>2024</year>). &#x201C;<article-title>Scaling in concrete&#x2013;causes, prevention, and repair</article-title>&#x201D; in eds. Yaagoob, I. Y. and Verma, C. <source>Industrial scale inhibition: principles, design, and applications</source>, <fpage>341</fpage>&#x2013;<lpage>356</lpage>.</mixed-citation></ref>
<ref id="ref49"><mixed-citation publication-type="book"><person-group person-group-type="author"><name><surname>Tang</surname> <given-names>K.</given-names></name></person-group> (<year>2023</year>). <source>Applications of deep convolutional neural networks for energy-based materials characterization on digital images</source> (Doctoral Dissertation). <publisher-loc>Sydney, New South Wales, Australia</publisher-loc>: <publisher-name>University of New South Wales (Australia)</publisher-name>.</mixed-citation></ref>
<ref id="ref50"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Temani</surname> <given-names>A.</given-names></name> <name><surname>Feng</surname> <given-names>Y.</given-names></name></person-group> (<year>2025</year>). <article-title>A segmentation-independent workflow for digital rock property prediction using pore-structure-preserving generative adversarial networks and synthetic data: porosity case study</article-title>. <source>SPE J.</source> <volume>31</volume>, <fpage>57</fpage>&#x2013;<lpage>74</lpage>. doi: <pub-id pub-id-type="doi">10.2118/231405-pa</pub-id></mixed-citation></ref>
<ref id="ref51"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Trigka</surname> <given-names>M.</given-names></name> <name><surname>Dritsas</surname> <given-names>E.</given-names></name></person-group> (<year>2025</year>). <article-title>A comprehensive survey of deep learning approaches in image processing</article-title>. <source>Sensors</source> <volume>25</volume>:<fpage>531</fpage>. doi: <pub-id pub-id-type="doi">10.3390/s25020531</pub-id>, <pub-id pub-id-type="pmid">39860903</pub-id></mixed-citation></ref>
<ref id="ref52"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>X.</given-names></name> <name><surname>Chen</surname> <given-names>Z.</given-names></name> <name><surname>Ren</surname> <given-names>J.</given-names></name> <name><surname>Chen</surname> <given-names>S.</given-names></name> <name><surname>Xing</surname> <given-names>F.</given-names></name></person-group> (<year>2022</year>). <article-title>Object status identification of X-ray CT images of microcapsule-based self-healing mortar</article-title>. <source>Cem. Concr. Compos.</source> <volume>125</volume>:<fpage>104294</fpage>.</mixed-citation></ref>
<ref id="ref53"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>H.</given-names></name> <name><surname>Lyu</surname> <given-names>H.</given-names></name> <name><surname>Liu</surname> <given-names>T.</given-names></name> <name><surname>Li</surname> <given-names>Y.</given-names></name> <name><surname>Qian</surname> <given-names>H.</given-names></name></person-group> (<year>2023</year>). <article-title>Identification of experienced temperature in mortar and concrete using microstructural image and deep learning</article-title>. <source>Constr. Build. Mater.</source> <volume>409</volume>:<fpage>133966</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.conbuildmat.2023.133966</pub-id></mixed-citation></ref>
<ref id="ref54"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wang</surname> <given-names>Y.</given-names></name> <name><surname>Zhou</surname> <given-names>B.</given-names></name></person-group> (<year>2022</year>). <article-title>Recent progress in single and combined porosity-evaluation techniques for porous materials</article-title>. <source>Materials</source> <volume>15</volume>:<fpage>2981</fpage>. doi: <pub-id pub-id-type="doi">10.3390/ma15092981</pub-id>, <pub-id pub-id-type="pmid">35591316</pub-id></mixed-citation></ref>
<ref id="ref55"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>X.</given-names></name> <name><surname>Li</surname> <given-names>Q.</given-names></name> <name><surname>Wang</surname> <given-names>L.</given-names></name> <name><surname>Liu</surname> <given-names>S.</given-names></name> <name><surname>Chen</surname> <given-names>H.</given-names></name> <name><surname>Hou</surname> <given-names>P.</given-names></name> <etal/></person-group>. (<year>2026</year>). <article-title>Exploring the correlation between microstructure and compressive strength in Portland cement-based materials: a deep learning-based comparison of &#x03BC; CT and SEM-BSE microstructure images</article-title>. <source>J. Mater. Civ. Eng.</source> <volume>38</volume>:<fpage>04025547</fpage>. doi: <pub-id pub-id-type="doi">10.1061/jmcee7.mteng-20664</pub-id></mixed-citation></ref>
<ref id="ref56"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Wu</surname> <given-names>J.</given-names></name> <name><surname>Luo</surname> <given-names>R.</given-names></name> <name><surname>Luo</surname> <given-names>L.</given-names></name> <name><surname>Lei</surname> <given-names>C.</given-names></name> <name><surname>Chen</surname> <given-names>X.</given-names></name></person-group> (<year>2025</year>). <article-title>Advanced machine learning for low-data porosity and permeability prediction in tight sandstones</article-title>. <source>Geophysics</source> <volume>90</volume>, <fpage>M31</fpage>&#x2013;<lpage>M44</lpage>. doi: <pub-id pub-id-type="doi">10.1190/geo2024-0340.1</pub-id></mixed-citation></ref>
<ref id="ref57"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Xia</surname> <given-names>Y.</given-names></name> <name><surname>Liu</surname> <given-names>W.</given-names></name> <name><surname>Meng</surname> <given-names>J.</given-names></name> <name><surname>Hu</surname> <given-names>J.</given-names></name> <name><surname>Liu</surname> <given-names>W.</given-names></name> <name><surname>Kang</surname> <given-names>J.</given-names></name> <etal/></person-group>. (<year>2024</year>). <article-title>Principles, developments, and applications of spatially resolved spectroscopy in agriculture: a review</article-title>. <source>Front. Plant Sci.</source> <volume>14</volume>:<fpage>1324881</fpage>. doi: <pub-id pub-id-type="doi">10.3389/fpls.2023.1324881</pub-id>, <pub-id pub-id-type="pmid">38269139</pub-id></mixed-citation></ref>
<ref id="ref58"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Yu</surname> <given-names>Y.</given-names></name> <name><surname>Wei</surname> <given-names>W.</given-names></name> <name><surname>Cui</surname> <given-names>W.</given-names></name> <name><surname>Cheng</surname> <given-names>W.</given-names></name> <name><surname>Zang</surname> <given-names>J.</given-names></name> <name><surname>Fang</surname> <given-names>L.</given-names></name> <etal/></person-group>. (<year>2025</year>). <article-title>Multi-phase segmentation methods for micro-tomographic images based on deep learning</article-title>. <source>Geoenergy Sci. Eng.</source> <volume>252</volume>:<fpage>213962</fpage>. doi: <pub-id pub-id-type="doi">10.1016/j.geoen.2025.213962</pub-id></mixed-citation></ref>
<ref id="ref59"><mixed-citation publication-type="journal"><person-group person-group-type="author"><name><surname>Zhang</surname> <given-names>P.</given-names></name> <name><surname>Ding</surname> <given-names>J.</given-names></name> <name><surname>Guo</surname> <given-names>J.</given-names></name> <name><surname>Wang</surname> <given-names>F.</given-names></name></person-group> (<year>2024</year>). <article-title>Fractal analysis of cement-based composite microstructure and its application in evaluation of macroscopic performance of cement-based composites: a review</article-title>. <source>Fractal Fract.</source> <volume>8</volume>:<fpage>304</fpage>. doi: <pub-id pub-id-type="doi">10.3390/fractalfract8060304</pub-id></mixed-citation></ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by" id="fn0001">
<p>Edited by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3162865/overview">Ashhad Imam</ext-link>, Sam Higginbottom University of Agriculture, Technology and Sciences, India</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by" id="fn0002">
<p>Reviewed by: <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/2300226/overview">Vishnu Saseendran</ext-link>, The Pennsylvania State University (PSU), United States</p>
<p><ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3332696/overview">Shohreh Sheiati</ext-link>, Technical University of Denmark, Denmark</p>
</fn>
</fn-group>
</back>
</article>