<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3-mathml3.dtd">
<article article-type="research-article" dtd-version="1.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:ali="http://www.niso.org/schemas/ali/1.0/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Phys.</journal-id>
<journal-title-group>
<journal-title>Frontiers in Physics</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Phys.</abbrev-journal-title>
</journal-title-group>
<issn pub-type="epub">2296-424X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">1742403</article-id>
<article-id pub-id-type="doi">10.3389/fphy.2025.1742403</article-id>
<article-version article-version-type="Version of Record" vocab="NISO-RP-8-2008"/>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Original Research</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Machine learning and digital images of porous materials: from rock to the human brain</article-title>
<alt-title alt-title-type="left-running-head">Sahimi and Sahimi</alt-title>
<alt-title alt-title-type="right-running-head">
<ext-link ext-link-type="uri" xlink:href="https://doi.org/10.3389/fphy.2025.1742403">10.3389/fphy.2025.1742403</ext-link>
</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Sahimi</surname>
<given-names>Ali</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing - original draft</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Methodology" vocab-term-identifier="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Sahimi</surname>
<given-names>Muhammad</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/95877"/>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Conceptualization" vocab-term-identifier="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Project administration" vocab-term-identifier="https://credit.niso.org/contributor-roles/Project administration/">Project administration</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Data curation" vocab-term-identifier="https://credit.niso.org/contributor-roles/data-curation/">Data curation</role>
<role vocab="credit" vocab-identifier="https://credit.niso.org/" vocab-term="Writing &#x2013; original draft" vocab-term-identifier="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing - original draft</role>
</contrib>
</contrib-group>
<aff id="aff1">
<label>1</label>
<institution>Grossman School of Medicine, New York University</institution>, <city>New York</city>, <state>NY</state>, <country country="US">United States</country>
</aff>
<aff id="aff2">
<label>2</label>
<institution>Mork Family Department of Chemical Engineering and Materials Science, University of Southern California</institution>, <city>Los Angeles</city>, <state>CA</state>, <country country="US">United States</country>
</aff>
<author-notes>
<corresp id="c001">
<label>&#x2a;</label>Correspondence: Muhammad Sahimi, <email xlink:href="mailto:moe@usc.edu">moe@usc.edu</email>
</corresp>
</author-notes>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-17">
<day>17</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date publication-format="electronic" date-type="collection">
<year>2025</year>
</pub-date>
<volume>13</volume>
<elocation-id>1742403</elocation-id>
<history>
<date date-type="received">
<day>23</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="rev-recd">
<day>18</day>
<month>12</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>26</day>
<month>12</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2026 Sahimi and Sahimi.</copyright-statement>
<copyright-year>2026</copyright-year>
<copyright-holder>Sahimi and Sahimi</copyright-holder>
<license>
<ali:license_ref start_date="2026-02-17">https://creativecommons.org/licenses/by/4.0/</ali:license_ref>
<license-p>This is an open-access article distributed under the terms of the <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">Creative Commons Attribution License (CC BY)</ext-link>. The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</license-p>
</license>
</permissions>
<abstract>
<p>Porous media and materials are ubiquitous and found everywhere. Some of them are referred to as rock-like porous media (RLPM), which include soil, concrete, asphalt, and oil and gas reservoirs. A second group consists of biological porous materials (BPMs), ranging from skin to organs such as the brain and lungs. The use of digital images of BPMs for the diagnosis and treatment of illnesses has a relatively long history, whereas their utilization in modeling various phenomena in RLPM is relatively recent. Due to the complexity of such images, along with the need to extract as much information from them as possible, the use of machine-learning (ML) approaches&#x2014;in particular, neural networks (NNs)&#x2014;has been increasing at a rapid pace. We describe and discuss recent progress in the applications of ML algorithms, particularly NNs, for the characterization of such images for the two classes of porous media and materials and show that, while they may seem vastly different, they actually have many similarities, and similar issues must be addressed when using and analyzing the images. As a result, the application of ML algorithms to both types of porous materials is largely similar, even though the goals may be very different.</p>
</abstract>
<kwd-group>
<kwd>biological porous media</kwd>
<kwd>digital image</kwd>
<kwd>machine learning</kwd>
<kwd>neural network</kwd>
<kwd>rock-like porous material</kwd>
</kwd-group>
<funding-group>
<funding-statement>The author(s) declared that financial support was received for this work and/or its publication. While the work of M.S. for the problems described in this study did not receive any specific funding, his current work on the applications of ML algorithms and NNs to problems in porous media and materials is supported by three National Science Foundation grants. A.S. is supported by a scholarship from New York University Grossman School of Medicine. The grant numbers are CBET 2230593, Hydrol. 2333378.</funding-statement>
</funding-group>
<counts>
<fig-count count="11"/>
<table-count count="0"/>
<equation-count count="15"/>
<ref-count count="130"/>
<page-count count="00"/>
</counts>
<custom-meta-group>
<custom-meta>
<meta-name>section-at-acceptance</meta-name>
<meta-value>Interdisciplinary Physics</meta-value>
</custom-meta>
</custom-meta-group>
</article-meta>
</front>
<body>
<sec sec-type="intro" id="s1">
<label>1</label>
<title>Introduction</title>
<p>Porous media and materials are ubiquitous and found everywhere, from the nanoscale at the molecular level to the largest, gigascopic scales encountered in landscapes, mountains, and rivers [<xref ref-type="bibr" rid="B1">1</xref>]. Almost all natural porous media and materials, and many synthetic ones, are also <italic>disordered</italic> or <italic>heterogeneous</italic> over some length scales. The type of heterogeneity depends on the length scale of observation and measurement. At small or laboratory scales, heterogeneity manifests itself through the morphology of the pore space, which consists of its geometry and topology. The geometry is represented by spatial variations in pore size, the porosity of the pore space (the pores&#x2019; volume fraction), and the roughness of the internal pore surfaces. The topology describes the interconnectivity of the pores and how pore connectivity at small scales affects macroscopic properties of the pore space at large scales, along with the tortuosity <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the pore space [<xref ref-type="bibr" rid="B2">2</xref>], which is a measure of the actual length of the paths for fluid flow, transport, and reaction in the pore space. Due to such complexities, the characterization of porous media and materials has always been an active area of research.</p>
<p>Two important classes of porous materials are referred to as rock-like porous media (RLPM) and biological porous materials (BPMs). The former includes soil, sandstone, concrete, and several other types of porous media. In the latter group are skin, lung, brain, liver, bone, and many other types of biological tissues and organs. The properties of several types of RLPM have been studied, characterized, and known for decades, while there is still significant ongoing research on other types. Recognizing biological materials as porous media is a much more recent development; therefore, the characterization of their morphology is a subject of ongoing research. It is, therefore, useful to discuss some of the important properties of BPMs.</p>
<p>The outer layer of the epidermis of human skin is referred to as the stratum corneum, which protects the underlying tissues from infection, dehydration, chemicals, and mechanical stress. It is made up of up to twenty layers of flattened cells whose properties include mechanical shear and impact resistance, water flux and hydration regulation, and selective permeability that excludes toxins and other harmful materials. The pore structure of stratum corneum has been characterized [<xref ref-type="bibr" rid="B3">3</xref>, <xref ref-type="bibr" rid="B4">4</xref>] by Raman scattering, as well as optical, chemical, and electron microscopy. <xref ref-type="fig" rid="F1">Figure 1</xref> shows a high-resolution image of the pore structure of human skin.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>High-resolution image of pore structures in skin tissue [<xref ref-type="bibr" rid="B5">5</xref>]. Skin fibers are shown in pink, the white space between the fibers represents the pores, and the purple structures that are boxed and zoomed in on are the sweat glands. The porosity is approximately 0.19, while the mean pore diameter is approximately <inline-formula id="inf2">
<mml:math id="m2">
<mml:mrow>
<mml:mn>220</mml:mn>
<mml:mo>&#xb1;</mml:mo>
<mml:mn>34</mml:mn>
<mml:mspace width="2.77695pt" class="tmspace"/>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> m. A fully automatic microscope was used to scan the skin section.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g001.tif">
<alt-text content-type="machine-generated">Microscopic image of skin tissue showing layers of cells in shades of red and purple. A smaller inset highlights a detailed section with a close-up of cellular structures. Scale bars indicate 500 micrometers and 100 micrometers.</alt-text>
</graphic>
</fig>
<p>Lungs have long been viewed as porous media [<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>] and have been modeled as such to study the various phenomena that occur within them. In particular, their pore-size distribution [<xref ref-type="bibr" rid="B8">8</xref>], their tortuosity, and the fractal dimension <inline-formula id="inf3">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of the three-dimensional (3D) pore space [<xref ref-type="bibr" rid="B9">9</xref>] have been measured and characterized. Measurements indicated [<xref ref-type="bibr" rid="B8">8</xref>] that the pore sizes vary between 5 and 100 nm, while tortuosity was <inline-formula id="inf4">
<mml:math id="m4">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
<mml:mo>&#x2248;</mml:mo>
<mml:mn>1.22</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and the fractal dimension turned out to be <inline-formula id="inf5">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2248;</mml:mo>
<mml:mn>2.34</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> [<xref ref-type="bibr" rid="B9">9</xref>]. <xref ref-type="fig" rid="F2">Figure 2</xref> presents an image of the lung along with an explanation of the various parts of the organ.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Top: Scanning electron micrograph of the lung parenchyma. Bottom: Details of the pore structure (see <ext-link ext-link-type="uri" xlink:href="https://quizlet.com/548658685/microscopic-anatomy-lung-tissue-diagram/">https://quizlet.com/548658685/microscopic-anatomy-lung-tissue-diagram/</ext-link>).</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g002.tif">
<alt-text content-type="machine-generated">Microscopic image of lung tissue showing a porous structure with variously sized dark spots. Below, an anatomical diagram illustrates lung sections with labeled parts: respiratory bronchiole, alveolar duct, alveoli, alveolar sac, and alveolar pores.</alt-text>
</graphic>
</fig>
<p>Digitized images indicate that the microstructure of the interstitial space (IS) of the brain&#x2019;s extracellular space resembles an unconsolidated (granular) porous medium. It is a porous medium because it is a fluid-saturated, two-phase material. One phase is the solid matrix of the cells, while the second phase is a network of pores, or void space, which contains the fluid that makes up roughly 80 percent of the brain, with the remainder being the solid matrix, thereby endowing it with characteristic poroelastic properties. The porosity, or what neuroscientists call the <italic>volume fraction</italic>, has important physical functions without which the brain could not operate. These functions include (a) allowing the flow of interstitial and cerebrospinal fluids that deliver nutrients, maintaining a stable ionic environment for neuronal signals, and disposing of metabolic &#x201c;waste&#x201d; products, such as the protein amyloid<inline-formula id="inf6">
<mml:math id="m6">
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3b2;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>; (b) allowing diffusion of neurotransmitters and ions in the pore network of extracellular space, which plays a fundamental role for cell-to-cell communication; and (c) providing mechanical protection through the fluid that absorbs physical shocks and regulates intracranial pressure. The interstitial space is, of course, deformable, and therefore, it should be viewed as a &#x201c;soft&#x201d; porous medium. Being a vital organ, even small deformations can result in large changes in the properties of the pore space of the IS, which can have important implications for its biological and neurological activities. Although the brain has been modeled [<xref ref-type="bibr" rid="B10">10</xref>] as an unconsolidated porous medium and its various microstructural properties, ranging from porosity to tortuosity and pore-size distribution, have been determined and compared with data [<xref ref-type="bibr" rid="B10">10</xref>, <xref ref-type="bibr" rid="B11">11</xref>], such models did not take into account the effect of deformation, whereas studies of permeability and ionic conductivity of deformable porous media in other contexts have indicated that deformation strongly influences the flow and electrical conductivity [<xref ref-type="bibr" rid="B12">12</xref>, <xref ref-type="bibr" rid="B13">13</xref>].</p>
<p>Of particular interest are the types of transport processes that occur in the pore space of the IS of the brain. Naturally, diffusion is one mechanism of transport. The measurement of diffusion of small molecules in the IS pore space indicated [<xref ref-type="bibr" rid="B14">14</xref>, <xref ref-type="bibr" rid="B15">15</xref>] that the effective diffusivity <inline-formula id="inf7">
<mml:math id="m7">
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is smaller than its value <inline-formula id="inf8">
<mml:math id="m8">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> in the bulk (outside the pore space), which provided strong evidence for the IS of brain&#x2019;s extracellular space being a porous medium since <inline-formula id="inf9">
<mml:math id="m9">
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math id="m10">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are related through the relation <inline-formula id="inf11">
<mml:math id="m11">
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>&#x3d5;</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>/</mml:mo>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf12">
<mml:math id="m12">
<mml:mrow>
<mml:mi>&#x3d5;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the porosity and <inline-formula id="inf13">
<mml:math id="m13">
<mml:mrow>
<mml:mi>&#x3c4;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the tortuosity. Moreover, since the pores of the IS are very small, anomalous diffusive behavior may be expected. For example, it is known that when the ratio of the size of the diffusing molecules and that of the pores is nonzero, diffusion is hindered [<xref ref-type="bibr" rid="B16">16</xref>, <xref ref-type="bibr" rid="B17">17</xref>], i.e., slowed down. In addition, experience with very small nanotubes, of the size that can be found in the range of pore sizes of the IS, indicated that [<xref ref-type="bibr" rid="B18">18</xref>] the Stokes&#x2013;Einstein relation between the viscosity and diffusivity breaks down, which could also be the case in the pore space of the IS. This was speculated on [<xref ref-type="bibr" rid="B11">11</xref>] a long time ago. Most natural porous media also contain dead-end pores where diffusing molecules may spend a long time, giving rise to non-Gaussian diffusion, and the pore space of the IS is no exception [<xref ref-type="bibr" rid="B19">19</xref>]. The second mode of transport, convection, and its role in the transport of molecules within the IS pore space have been a controversial issue. <xref ref-type="fig" rid="F3">Figure 3</xref> presents [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B20">20</xref>] a realistic model of brain extracellular space and compares it with the two-dimensional section of neuropil ultrastructure observed in electron microscopy.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>Brain extracellular space as an unconsolidated porous medium that mimics a two-dimensional section of neuropil structure shown at the bottom. The porosity is <bold>(A)</bold> 0.13; <bold>(B)</bold> 0.16, and <bold>(C)</bold> 0.18.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g003.tif">
<alt-text content-type="machine-generated">The diagrams labeled A, B, and C show clustered, irregular cells, representing models of an electron micrograph that depicts cellular structures with a similar pattern, highlighted by grey textures and blue outlines.</alt-text>
</graphic>
</fig>
<p>With advances in the development of efficient computational methods, techniques, and instrumentation for obtaining high-resolution digitized images of porous media, attention has increasingly focused on using such images not only to characterize their morphology but also to carry out direct numerical simulations of various phenomena in their pore space. Digital rock physics combines microtomographic imaging with efficient computer simulations in order to compute the effective properties of porous media and is intended to complement laboratory studies and measurements in order to gain a deeper understanding of physical processes related to various fluid flow, diffusion, reaction, and deformation processes in such media. The approach has been used extensively for RLPM. At the same time, imaging technologies play a key role in the diagnosis of abnormalities and medical therapy. Several such technologies exist, ranging from X-ray radiography and computed tomography (CT) to magnetic resonance imaging (MRI) and spectroscopy, medical optical imaging, and ultrasonic and electrical impedance tomography. Such techniques have many applications in the diagnosis of myocardial diseases, cancer of various tissues, neurological disorders, congenital heart disease, complex bone fractures, and other serious medical conditions [<xref ref-type="bibr" rid="B21">21</xref>].</p>
<p>Medical images (MIs), as well as those for RLPM, are complex systems that contain a significant amount of information whose extraction is not straightforward. A key factor in their utility is their resolution, while a second key factor is the ability of the user to extract the maximum amount of information from them. The role of such factors has motivated the application of machine-learning (ML) algorithms, particularly neural networks (NNs), to digitized images as these algorithms and tools have distinguished themselves by their ability to learn complex patterns and extract information from raw data.</p>
<p>The aim of this study is threefold. One is to describe some of the recent advances in this area. We focus on RLPM, on one hand, and on the brain as a prototypical example of a biological porous medium, on the other hand. We argue that ML algorithms, which have been developed to enhance the resolution of images of RLPM and other complex materials, can also be used to enhance the resolution of MIs, thereby enabling more accurate diagnoses. Thus, both communities can benefit from the achievements of each other in using ML approaches to analyze digitized images of their respective porous media. The second aim is to highlight that the same techniques developed for modeling of RLPM can be applied to MIs, particularly those of the brain. Finally, in some respects, the field of RLPM modeling is more advanced than the medical field dealing with biological porous materials, in that more rigorous analyses and highly realistic models have been developed, whereas modeling of biological systems is often characterized by empiricism and <italic>ad hoc</italic> regression. For example, pore-network modeling of various flow phenomena in RLPM is highly advanced, and many rigorous methods and models have been developed. Thus, it can be highly beneficial for the medical community working on biological porous media, as described here, to take advantage of this progress in their efforts to gain a better understanding of the systems of interest.</p>
<p>The connection between imaging of geomaterials and CT scanners used in the medical field is not entirely new and dates back nearly 40 years, when researchers at Shell Oil Company used medical CT scanners to characterize rock [<xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B23">23</xref>]. What we propose in this study is, however, more comprehensive, in that we believe the medical community can take advantage of advances in the modeling and characterization of RLPM, while researchers in the area of RLPM can benefit from advances in instrumentation for medical imaging and from the way such images are interpreted and understood.</p>
</sec>
<sec id="s2">
<label>2</label>
<title>Classification of images</title>
<p>Classification of digitized images&#x2014;in particular MIs and those for RLPM&#x2014;is an important problem that, despite significant progress, is still being studied. The approach in both cases is essentially the same, although the goals may be different. In what follows, we describe some of the recent progress that has been made. It should be noted that the accuracy of all the models described below was tested not only by comparing their predictions with the existing data but also through various statistical measures, such as the coefficient of determination <inline-formula id="inf14">
<mml:math id="m14">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<sec id="s2-1">
<label>2.1</label>
<title>Rock-like porous media</title>
<p>Classification of images of RLPM is multifaceted. In one problem, one would like to distinguish pores of different sizes and classify them into micro-, meso-, and macropores. Note, however, that there are multiple definitions of the three length scales, or respective ranges of pore sizes, so that those working on imaging of biological systems and geologists may not be entirely consistent in their interpretation of what the pore sizes represent. Many porous media contain microfractures, and therefore, another problem is <italic>detection</italic>, i.e., identifying microfractures and distinguishing them from large pores. One issue in both cases is a lack of extensive data for training the ML algorithm. Addressing this problem is relatively straightforward and is usually accomplished through <italic>transfer learning</italic> (TL), i.e., using NNs that have already been trained on other types of data related to porous materials that are relatively abundant and then fine-tuning them for the properties of interest. For example, [<xref ref-type="bibr" rid="B24">24</xref>] used CT images together with five well-known pre-trained ML models, namely, VGG-16, ResNet-50, InceptionV3 [<xref ref-type="bibr" rid="B25">25</xref>] (introduced by Google), DenseNet121 [<xref ref-type="bibr" rid="B26">26</xref>], and MobileNet [<xref ref-type="bibr" rid="B27">27</xref>], all of which represent deep convolutional NNs (CNNs), to generate accurate synthetic data for pore sizes and then used the NNs to classify images of soils based on their pore-size distribution, classifying the pores into various types mentioned above.</p>
<sec id="s2-1-1">
<label>2.1.1</label>
<title>Detection</title>
<p>Several studies address the detection problem. One notable example is the work of [<xref ref-type="bibr" rid="B28">28</xref>], who developed an ML-based algorithm using eighteen images for thin sections of carbonate rock that were scanned at a resolution of 6.35 microns/pixel. To correctly classify the structure, only open-mode microfractures and pores were considered, and healed microfractures, usually referred to as microveins, were ignored. The edges of all images were cropped to remove the blank slide edges, after which they were pre-processed by denoising and sharpening. Various methods can be used for denoising, such as those based on wavelet and curvelet transformations [<xref ref-type="bibr" rid="B29">29</xref>, <xref ref-type="bibr" rid="B30">30</xref>] and other methods [<xref ref-type="bibr" rid="B31">31</xref>] that preserve the edges of the images. If, after denoising, parts of the images were not sharp, an unsharp mask filter was used to restore the sharpness. Thresholding was used to segment the pores that were filled with the blue epoxy. A complication was that microfractures that seemed continuous were actually fragmented into smaller segments after thresholding. To address the issue and obtain the correct microfracture connectivity, a device-independent 3D color space was used that accurately maps all perceivable colors, the result of which was then combined with the original segmented image after post-processing both images. Microporous matrix zones and microporous grains were segmented as macropores, and due to the sheer number of microporous zones, they were considered pores. Many smaller pores were poorly resolved, with their true shapes lost. To address the issue, [<xref ref-type="bibr" rid="B28">28</xref>] visually estimated the smallest pore size that could adequately be resolved, which in their images was 30 pixels in area. Thus, all objects below the threshold were removed from the images, while care was taken to ensure that microfractures were not removed since their number throughout the dataset was limited.</p>
<p>The binary systems were processed in Python for labeling and feature extraction. Feature selection was supervised and based on <italic>a priori</italic> knowledge of the features and their correlations. The features included area, perimeter, filled area (defined as the number of pixels in the object with holes filled), major and minor axis lengths (defined as the lengths of the two axes of the best-fitting ellipses), and a few others. To improve the accuracy of the ML approach, a manual approach was used to remove the outliers. The idea was to examine the data points, and if they were more than 10 standard deviations from the mean of both size and shape features, they were visually corroborated with their corresponding images before being considered outliers. The labeled dataset consisted of 400 pores and 400 microfractures, with the pores selected randomly, while the microfractures were selected manually. Secondary labels were added to each sampled object that pertained to the type of pore or microfracture. Four types of microfracture were identified, namely, straight, curvilinear, curved, and branching. Seventy percent of the labeled dataset was randomly selected for training, with the remaining thirty percent used for testing, while maintaining the same proportions of pores and microfractures in both sets.</p>
<p>Several supervised ML models were then tested. They included random forests (which combine several decision-trees with randomly distributed features to generate a majority vote for classification, which means that they mix several weaker learners into a single robust learner), variants of support vector machines (which work based on identifying the most effective boundary to separate different classes of data points), and the <inline-formula id="inf15">
<mml:math id="m15">
<mml:mrow>
<mml:mi>K</mml:mi>
<mml:mo>&#x2212;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>nearest neighbor algorithm. The hyperparameters of the various algorithms were then optimized. As the next step, learning curves, which graphically represent the degree to which the ML model learns the classification task on incrementally larger portions of a training dataset, were generated for the ML models in order to evaluate their stability and detect any overfitting. The training and resampling increments were set at ten percent of the training dataset. The accuracy of the ML algorithms was assessed using the standard method, i.e., by identifying the true positive (TP), false positive (FP), true negative (TN), and false negative (FN) and defining accuracy <inline-formula id="inf16">
<mml:math id="m16">
<mml:mrow>
<mml:mi>A</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> as follows:<disp-formula id="equ1">
<mml:math id="m17">
<mml:mrow>
<mml:mi>A</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>In addition, sensitivity <inline-formula id="inf17">
<mml:math id="m18">
<mml:mrow>
<mml:mi>S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> of the algorithm is defined as follows:<disp-formula id="equ2">
<mml:math id="m19">
<mml:mrow>
<mml:mi>S</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>while specificity <inline-formula id="inf18">
<mml:math id="m20">
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> was taken to be,<disp-formula id="equ3">
<mml:math id="m21">
<mml:mrow>
<mml:mi mathvariant="script">S</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi mathvariant="normal">F</mml:mi>
<mml:mi mathvariant="normal">P</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi mathvariant="normal">T</mml:mi>
<mml:mi mathvariant="normal">N</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>
</p>
<p>All the supervised ML models produced highly accurate predictions, within a narrow range of 93.64&#x2013;94.63, implying that ML algorithms can accurately classify images of RLPM. <xref ref-type="fig" rid="F4">Figure 4</xref> presents an example of carbonate rock. [<xref ref-type="bibr" rid="B28">28</xref>] used two binary masks [a mask is a specific piece of data used for bitwise operations (such as AND and OR) on the bits, i.e., 1 and 0], namely, hue&#x2013;saturation&#x2013;brightness (HSB) masks and LAB masks. The HSB color space describes colors using three properties, namely, hue (representing a pure color), saturation (that varies from gray to vivid), and brightness (from black to bright color), and is a cylindrical model used in digital art. LAB (or CIELAB) color space is a model that describes colors with three values, namely, L for lightness, A for green-to-red, and B for blue-to-yellow.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Composites of the HSB and LAB binary masks, segmenting the microfractures. Red represents the HSB binary mask; green is the LAB binary mask, while yellow indicates the union of the two. Overall, the HSB mask produces a stronger segmentation, but the LAB mask provides a notable boost in connectivity (based on [<xref ref-type="bibr" rid="B28">28</xref>]).</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g004.tif">
<alt-text content-type="machine-generated">Three images show colored lines, representing microfractures, on a black background that is carbonate rock. The top and middle images feature wavy, multicolored microfractures in green, yellow, and red, running diagonally. The bottom image displays a vertical thick yellow line, a fracture, with smaller red lines scattered nearby, representing microfractures.</alt-text>
</graphic>
</fig>
<p>Other variations of this basic model are described extensively by [<xref ref-type="bibr" rid="B32">32</xref>].</p>
</sec>
<sec id="s2-1-2">
<label>2.1.2</label>
<title>Segmentation</title>
<p>Another problem in the classification of images of porous media is <italic>segmentation</italic>, i.e., dividing the pores and minerals into separate phases, an important problem that also arises in multiphase fluid flow when one must not only distinguish the pores and minerals but also various fluid phases, along with the wettability of the pore space. Segmentation models represent classification models in which the task is executed for each pixel in an image. Segmentation of, for example, micro-CT images by ML algorithm began with very basic methods, such as random forest classification, which proved more successful than most deterministic segmentation approaches (see, for example, [<xref ref-type="bibr" rid="B33">33</xref>]). Random forest classification has been successfully implemented in several open-source and commercial software programs, such as Fiji (trainable WEKA segmentation [<xref ref-type="bibr" rid="B34">34</xref>]), Ilastik, Reactiv&#x2019;IP SDK, and Dragonfly. Random forest classification is also the underlying method in many other software packages, where, for example, a higher-resolution SEM image is used to up-resolve the EDX-map image in electron microscopy [<xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B36">36</xref>].</p>
<p>Deep NNs have also been used for classification, particularly segmentation, of RLPM. [<xref ref-type="bibr" rid="B37">37</xref>] developed an ML-based algorithm for the segmentation of images of RLPM. They used SegNet [<xref ref-type="bibr" rid="B38">38</xref>], a fully convolutional encoder&#x2013;decoder network in which the encoder uses a CNN, but its fully-connected layers were removed in order to produce a low-resolution feature map of the input image. The decoder utilizes a similar architecture to generate a high-resolution feature map, which is then fed to a multiphase softmax layer of the CNN to classify it into a pixel-wise multiphase segmented output. Two SegNet networks were used. One was the standard network [<xref ref-type="bibr" rid="B38">38</xref>] with four encoders and four decoders, with each encoder/decoder having a <inline-formula id="inf19">
<mml:math id="m22">
<mml:mrow>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> convolutional/deconvolutional layer and a <inline-formula id="inf20">
<mml:math id="m23">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> max-pooling/un-pooling window and stride 2. The rectified linear unit (ReLU) activation function was used only in the decoder part. The second SegNet was similar to the standard version but featured a larger architecture with five encoder/decoder pairs, two convolutional/deconvolutional layers, and three convolutional/deconvolutional layers in the six central parts. Since, as mentioned above, massive amounts of data are needed for training the NNs, [<xref ref-type="bibr" rid="B37">37</xref>] used the cross-correlation simulation (CCSIM) method of [<xref ref-type="bibr" rid="B39">39</xref>&#x2013;<xref ref-type="bibr" rid="B42">42</xref>] to generate a large number of synthetic images based on the original ones. The results were then used to train the autoencoder CNN and to carry out the segmentation of images of several Berea sandstone samples.</p>
<p>Segmentation was carried out by thresholding the values of the pixels, resulting in automatic segmentation. An example is shown in <xref ref-type="fig" rid="F5">Figure 5b</xref>. It turned out, however, that two critical issues must be addressed for correct segmentation. As <xref ref-type="fig" rid="F5">Figure 5a</xref> indicates, grain boundaries are brighter than grain surfaces, and thus, they can be misclassified. In addition, various minerals can be similarly classified, due to their close color values or intensities. To address these issues, [<xref ref-type="bibr" rid="B37">37</xref>] manually labeled each misclassified pixel as an &#x201c;expert supervisor,&#x201d; thereby resulting in a semi-automatic segmentation algorithm. <xref ref-type="fig" rid="F5">Figure 5c</xref> shows the result of the semi-manual segmentation. Five phases were identified: pore space, quartz, K-feldspar, zircon, and other minerals that were mainly clays.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>
<bold>(a)</bold> Original CT image of a sandstone; <bold>(b)</bold> automatic segmentation and <bold>(c)</bold> semi-manual segmentation of the image in <bold>(a)</bold> [<xref ref-type="bibr" rid="B37">37</xref>]. In the right image, gray, light gray, white, semi-gray, and dark gray represent, respectively, quartz, ankerite, zircon, K-spar, and clay.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g005.tif">
<alt-text content-type="machine-generated">Three panels, labeled (a), (b), and (c), display grayscale images of rock textures. Panel (a) shows the initial rock texture. Panel (b) depicts an image with enhanced contrast and distinct boundaries. Panel (c) presents a further processed image, highlighting finer details and textures.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="s2-2">
<label>2.2</label>
<title>Biological porous media</title>
<p>Precisely, the same principles used for classifying images of RLPM can also be applied to MIs, for which two types of classification are of interest, as follows.</p>
<sec id="s2-2-1">
<label>2.2.1</label>
<title>Exam and lesion classifications</title>
<p>In <italic>exam classification</italic>, the input to the ML algorithm is one or several images (exams), but there is a single output, such as &#x201c;is the disease present or not?&#x201d; Clearly, dataset is small, and therefore, a TL method must be used to meet the large-data requirement of deep NN training. Two TL strategies have been used. One, as discussed above, is based on using a pre-trained NN as a feature extractor, which has the advantage of not requiring the user to train a deep NN, thereby allowing the extracted features to be easily inserted into the existing image analyzers. The second approach fine-tunes the pre-trained NN on medical data. Fine-tuning trains the parameters of an already pre-trained NN on new data, which can be carried out on the entire NN or on only a subset of its layers. In CNNs, it is common to keep the earlier layers frozen as they are closest to the input layer and capture lower-level features, while subsequent layers often discern high-level features that are more directly related to the task on which the model is trained. In many studies, however, the TL strategy has been arbitrarily configured. [<xref ref-type="bibr" rid="B43">43</xref>] provided a comprehensive review of efforts to offer guidance for selecting a model and TL approaches for the classification of MIs. [<xref ref-type="bibr" rid="B44">44</xref>] and [<xref ref-type="bibr" rid="B45">45</xref>] fine-tuned a pre-trained InceptionV3 architecture on medical data and reported near-human expert performance. [<xref ref-type="bibr" rid="B46">46</xref>] reviewed properties of TL for medical imaging. On the other hand, [<xref ref-type="bibr" rid="B47">47</xref>] carried out computations to compare training from scratch with fine-tuning of pre-trained NNs. Their work showed that fine-tuning can be more accurate if a small dataset of approximately <inline-formula id="inf21">
<mml:math id="m24">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> images of skin lesions is available, although the computer experiments were conducted on too small a scale to allow any general conclusions to be drawn from the results.</p>
<p>The second type of classification of MIs is <italic>lesion classification</italic>, the medical version of the problem of object classification, in which one focuses on classifying a small part of an MI into two or more classes, such as nodule classification in chest CT. Accurate classification requires, however, both local information on lesion appearance and global contextual information on lesion location, which is typically not possible in generic deep learning (DL) architectures. To address this problem, multi-stream architectures, such as multiple CNNs, are used in a multiscale fashion, allowing all contextual information to be processed, regardless of the image size. Accurate lesion classification often requires using 3D information. Note, however, that images in computer vision are usually in 2D, and therefore, NNs pre-trained on such images cannot be used directly when 3D information is available. There are various approaches to addressing this issue. For example, [<xref ref-type="bibr" rid="B48">48</xref>] utilized a multi-stream CNN to classify points of interest in chest CT as a nodule or non-nodule, with up to nine patches with various orientations extracted from the candidate image used in separate streams and merged in the fully connected layers to obtain the final classification output. Note that the typical loss function used in classification is based on cross-entropy loss, which is given by<disp-formula id="e1">
<mml:math id="m25">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>ln</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf22">
<mml:math id="m26">
<mml:mrow>
<mml:mi>C</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the number of classes classified and <inline-formula id="inf23">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the label, which is 1 when and only when <inline-formula id="inf24">
<mml:math id="m28">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>; otherwise, <inline-formula id="inf25">
<mml:math id="m29">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. <inline-formula id="inf26">
<mml:math id="m30">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is the predicted probability that the model assigns the current image to class <inline-formula id="inf27">
<mml:math id="m31">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. As for the location loss, the mean-squared error in position coordinates is typically used.</p>
</sec>
<sec id="s2-2-2">
<label>2.2.2</label>
<title>Detection</title>
<p>Similar to RLPM, the problem of detection is also important for MIs, where detection represents a more refined approach to image classification. This is because every pixel in an image is classified, and thus, the class balance can be skewed strongly toward the non-object class in a training dataset. While the majority of non-object samples are typically straightforward to discriminate, this prevents the ML method from focusing on the difficult datasets. As described above, the same problem arises in RLPM when, for example, one wishes to differentiate between microfractures and macropores. To address this issue, [<xref ref-type="bibr" rid="B49">49</xref>] proposed a method for selective sampling of data, in which datasets that are incorrectly classified are fed back to the NN more frequently to focus on difficult areas in the images, which, in their case, were retinal images. In addition, because classifying each pixel in a sliding window leads to orders of magnitude of redundant or useless calculation, CNN [<xref ref-type="bibr" rid="B50">50</xref>] can be an important aspect of an object detection approach. In the standard approach of using CNNs to classify each pixel in an image individually, patches of extracted data (the sliding window) around the particular pixel are fed into the network. The problem is that input patches from neighboring pixels have very significant overlap, and therefore, the same convolutions are computed too many times&#x2014;a waste of computation time. The same problem arises in the detection of images of RLPM. The convolution and inner (dot) product are both linear operators; this implies that one can be expressed in terms of the other. Thus, if fully connected layers are rewritten as convolutions, then a CNN can take input images larger than those on which it was trained and produce a <italic>likelihood map</italic>, rather than an output for a single pixel. The result is a fully CNN.</p>
<p>Detecting patterns or landmarks in the human body often requires a spatial 3D dataset, which is not straightforward to use with NNs, representing a complex process. One way of addressing this issue is to represent the 3D image as a set of 2D orthogonal slices. For example, to identify landmarks on the distal femur surface, [<xref ref-type="bibr" rid="B51">51</xref>] used three independent sets of 2D MRI slices with CNNs and defined the spatial position of the landmark as the intersection of the three 2D slices with the highest classification output. A more advanced technique was developed by [<xref ref-type="bibr" rid="B52">52</xref>], who detected regions of interest around anatomical regions, such as the heart and aortic arch, by identifying a rectangular 3D bounding box after 2D analyzing of the 3D CT volume. Note that, unlike many problems for which ML algorithms have been used, when it comes to the MIs that are about a specific problem, one may not have access to a considerable volume of data, and therefore, one often uses pre-trained CNNs to learn better feature representations [<xref ref-type="bibr" rid="B53">53</xref>]. Each landmark can also be represented, for example, by a Gaussian distribution, and a map of such landmarks [<xref ref-type="bibr" rid="B54">54</xref>] can be used as input to the CNN for training and subsequent prediction of the map. Another approach [<xref ref-type="bibr" rid="B55">55</xref>] is based on applying reinforcement learning to detect landmarks. To solve the difficult problem of direct detection and localization of landmarks and regions in 3D images, [<xref ref-type="bibr" rid="B56">56</xref>] partitioned 3D convolution into three 1D ones in order to detect carotid artery bifurcation in CT images.</p>
<p>Temporal data contained in, for example, a video, have also been analyzed for detecting landmarks. For example, [<xref ref-type="bibr" rid="B57">57</xref>] trained CNNs on the data contained in a video in order to detect up to twelve standardized scan planes in mid-pregnancy fetal ultrasounds. To locate the brain and spine in the scan plan, [<xref ref-type="bibr" rid="B58">58</xref>] utilized saliency maps. [<xref ref-type="bibr" rid="B59">59</xref>] developed a 4D (three spatial dimensions plus time) reconstruction method based on a deep CNN, dubbed 4D CINENet, for cases in which 3D Cartesian CINE imaging is undersampled, i.e., when insufficient data are available. Their NN was based on <inline-formula id="inf28">
<mml:math id="m32">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>D complex-valued spatio-temporal convolutions and multi-coil data processing. They trained and evaluated the proposed CINENet on 3D CINE data from 20 healthy subjects and 15 patients with suspected cardiovascular disease and showed that CINENet performs much more accurately than iterative reconstructions in terms of visual image quality and contrast. In addition to CNNs, recurrent NNs (RNNs) and long short-term memory (LSTM)-RNNs have also been used with information contained in medical videos. An example is the work of [<xref ref-type="bibr" rid="B59">59</xref>], who utilized LSTM models to incorporate dynamic information contained in consecutive sequences in ultrasound videos to detect fetal standard planes, while [<xref ref-type="bibr" rid="B60">60</xref>] combined an LSTM-RNN with a CNN to detect the end-diastole and end-systole frames in cardiovascular MRI, also called cine-MRI, or the &#x201c;gold-standard&#x201d; for the assessment of cardiac function of the heart.</p>
<p>In general, detection models in MIs, including those described above, are typically one of the two types, namely, single-stage or two-stage models. A single-stage detection (SSD) model must operate in a timely manner, with a low runtime, in order to be efficient. The first version of the algorithm, dubbed [<xref ref-type="bibr" rid="B61">61</xref>] YOLO (&#x201c;You only Look Once&#x201d;), formulated object detection as a regression problem. To do so, it discretized the image into a grid or mesh, instead of initializing the detection frame, thereby generating spatially separated grid blocks or bounding boxes and the associated class probabilities. A single NN was then used to predict bounding boxes and class probabilities directly from the full images in a single evaluation. Since the entire detection algorithm was based on a single NN, it could be optimized end-to-end directly on detection performance. However, if multiple objects fall on the grid at the same real time, YOLO version 1 becomes ineffective. The second version of YOLO, YOLO9000, developed by [<xref ref-type="bibr" rid="B62">62</xref>], addressed this shortcoming by adding to the grid the number of blocks to be selected and fine-grained features through the passthrough layer. Using a convolution operation, instead of a pooling operation, YOLOv3, also developed by [<xref ref-type="bibr" rid="B63">63</xref>], made an &#x201c;incremental improvement&#x201d; (in the authors&#x2019; words) by incorporating a residual structure and a feature fusion strategy, which not only improved its generalization performance but also reduced the number of parameters. YOLOv4, developed by [<xref ref-type="bibr" rid="B64">64</xref>], utilized data processing, a backbone network, an activation function, a loss function, and other strategies in order to speed up the required computation, turning it into an efficient detection model. Another detection approach, the single-shot multibox detector (SSD), developed by [<xref ref-type="bibr" rid="B65">65</xref>], also discretized the output space of bounding boxes into a set of default blocks with various aspect ratios and scales per feature map location. To make a prediction, the deep CNN generates scores for the presence of each object category in each default box and produces adjustments to the box to better match the object shape. Moreover, the proposed approach combines predictions from multiple feature maps with varying resolutions in order to handle objects of various sizes effectively.</p>
<p>A prototype of a two- or multi-stage detection algorithm is that of [<xref ref-type="bibr" rid="B66">66</xref>]. It consisted of two key ideas: (a) applying high-capacity CNNs to bottom&#x2013;up region proposal networks (RPNs) in order to localize and segment objects and (b) carrying out supervised pre-training for an auxiliary task when labeled training data are scarce, followed by domain-specific (or region-specific) fine-tuning (see above), which improves the performance significantly. An RPN [<xref ref-type="bibr" rid="B67">67</xref>] takes an image of any size as input and produces a set of rectangular object proposals, i.e., where an object could be, each with an objectness score, which is a confidence measure that is used in object detection models. It indicates the probability that a proposed region of interest contains an object, irrespective of its specific type. Because [<xref ref-type="bibr" rid="B66">66</xref>] combined RPNs with CNNs, they dubbed the method R-CNN, in which regions are represented using CNN features. An improved and computationally faster version of R-CNN, dubbed Fast R-CNN, was proposed by [<xref ref-type="bibr" rid="B68">68</xref>]. Since many detection algorithms that use deep CNNs require a fixed-size input image, [<xref ref-type="bibr" rid="B69">69</xref>] developed a method that no longer needed the constraint. In their model, dubbed Special Pyramid Pooling-Net (SSP-net), the last pooling layer (after the last convolutional layer) was replaced with a spatial pyramid pooling layer with a large number of bins. In each spatial bin, the approach pools (max pooling was used) the responses of each filter. The outputs of SPP are <inline-formula id="inf29">
<mml:math id="m33">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>-dimensional vectors, with <inline-formula id="inf30">
<mml:math id="m34">
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf31">
<mml:math id="m35">
<mml:mrow>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, being respectively, the number of bins and the number of filters in the last convolutional layer. The vectors have fixed dimensions and are the input to the fully connected layer of the CNN. Thus, the input image can be of any size, which not only allows arbitrary aspect ratios but also arbitrary scales. Note that the typical loss function in both methods includes terms corresponding to classification and location loss functions. The former is a multivariate cross-entropy loss, given by <xref ref-type="disp-formula" rid="e1">Equation 1</xref>. As for the location loss, the mean-squared error in position coordinates is usually used.</p>
</sec>
<sec id="s2-2-3">
<label>2.2.3</label>
<title>Segmentation</title>
<p>Similar to RLPM, segmentation of MIs is also an important problem. In the context of MIs, segmentation of the images is defined as the identification of a set of voxels (pixels) that constitute either the contour or the interior of the organs of interest in the images. Deep learning-based models of segmentation typically use the architecture of the encoder&#x2013;decoder of SegNet (see above). As discussed above, features at a high level of abstraction are obtained in the encoder by convolution and downsampling, while the feature map is recovered to the original size of the image by convolution and upsampling. The output is the segmentation of the image. The U-Net architecture, which is widely used for medical images, combines an equal number of upsampling and downsampling layers, with so-called skip connections between corresponding convolution and deconvolution layers. Skip connection, also referred to as residual connections, bypasses one or more layers by adding the input of a block directly to its output, thereby forming a shortcut, or skipping path, that helps mitigate the vanishing gradient problem in minimizing the loss function. Skip connection enables training of much deeper NNs since it makes it possible for gradients to flow more easily through an NN and preserve important features from earlier layers. This concatenates features from the contracting and expanding paths so that the entire image is processed in a single forward pass, directly generating a segmentation map directly while taking into account the complete content of the image. It was shown by [<xref ref-type="bibr" rid="B70">70</xref>] that a complete 3D segmentation can be achieved if the U-Net is provided with a few 2D annotated slices from the same 3D image. [<xref ref-type="bibr" rid="B71">71</xref>] improved this basic scheme. U-Net and its variants are now the standard NNs for segmentation of lesions. Let us describe a particular variation of U-Net that has attracted considerable attention.</p>
<p>[<xref ref-type="bibr" rid="B72">72</xref>] developed a deep learning-based segmentation method for MIs, dubbed nnU-Net, which configures itself automatically and carries out multiple tasks, including preprocessing, network architecture, training, and post-processing for any new task in the biomedical domain. The configuration of nnU-Nets consists of two blocks per resolution step in both the encoder and decoder, with each block consisting of a convolution, followed by instance normalization and a leaky ReLU, and is shown in <xref ref-type="fig" rid="F6">Figure 6</xref>. The architecture of nnU-Net is the same as that of the standard U-Net, but with minor adjustments. In order to enable large patch sizes, nnU-Net&#x2019;s batch size is as small as two. Batch normalization is typically used in CNNs for speeding up or stabilizing the NN training. It does not, however, perform well when the batch sizes are small. Therefore, [<xref ref-type="bibr" rid="B72">72</xref>] utilized instance normalization [<xref ref-type="bibr" rid="B73">73</xref>]. In a CNN, batch normalization normalizes all images across the batch and spatial locations. In instance normalization, on the other hand, each element of the batch is normalized independently. [<xref ref-type="bibr" rid="B72">72</xref>] used the leaky ReLU (with a negative slope of 0.01), instead of ReLU. The training was performed with deep supervision, and additional auxiliary losses were added to the decoder at all but the two lowest resolutions, allowing gradients to propagate deeper into the network and facilitating the training of all the layers. [<xref ref-type="bibr" rid="B72">72</xref>] performed downsampling using strided convolution, while upsampling was carried out using transposed convolution. To balance the requirements for performance and memory, the initial number of feature maps was set to 32, and doubled (halved) with each downsampling (upsampling), while to limit the size of the final model, the number of feature maps was limited to 320 and 512 for 3D and 2D U-Nets, respectively.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Configuration of architecture for the segmentation of biomedical images based on nnU-Net. For a given segmentation task, dataset properties are extracted as &#x201c;dataset fingerprint&#x201d; (pink). A set of heuristic rules models parameter interdependencies (thin arrows) and operates on the fingerprint to deduce the data-dependent &#x201c;rule-based parameters&#x201d; (green) of the pipeline (that contains all the relevant information), which are complemented by &#x201c;fixed parameters&#x201d; (blue) that are predefined. Up to three configurations are trained in a five-fold cross-validation. Then, nnU-Net automatically selects empirically the optimal ensemble of the models and determines whether post-processing is required (&#x201c;empirical parameters,&#x201d; shown in yellow) (based on [<xref ref-type="bibr" rid="B72">72</xref>]).</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g006.tif">
<alt-text content-type="machine-generated">Flowchart of nnU-net framework showing connections between components. It includes data fingerprinting, rule-based parameters, and empirical parameters. Key elements involve distribution of spacings, median shape, image modality, intensity distribution, patch size, batch size, network topology, and ensemble selection. Arrows illustrate process flow from train data to prediction, detailing configuration of post-processing and network training in 2D, 3D, and 3DC formats.</alt-text>
</graphic>
</fig>
<p>[<xref ref-type="bibr" rid="B72">72</xref>] trained all networks for <inline-formula id="inf32">
<mml:math id="m36">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> epochs, with one epoch defined as an iteration over 250 mini-batches. Stochastic gradient descent with an initial learning rate of <inline-formula id="inf33">
<mml:math id="m37">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> was used for learning the weights of the NNs. The loss function was the sum of cross-entropy (see above) and Dice loss. Also known as the Dice-S&#xf8;rensen coefficient, the Dice loss is a statistical measure used to quantify the similarity between two samples. For each deep supervision output, a corresponding downsampled ground-truth segmentation mask was used in order to compute the loss. The objective function of the training was the sum of the losses at all resolutions. Samples for the mini-batches were selected from random training cases. Class imbalance occurs when some types of data are much more common than others that are exceedingly rare. For example, most voxels in an MI belong to the non-diseased class. As a result of class imbalance, NNs and other ML algorithms are often trained in class-balanced settings. To address the issue, [<xref ref-type="bibr" rid="B72">72</xref>] implemented oversampling. Two-thirds of the samples were taken from random locations within the selected training case, while the remaining one-third of the patches were ensured to contain one of the foreground classes present in the selected training sample.</p>
<p>Accurate segmentation also requires global and local context, which is why multi-stream networks with distinct scales or non-uniformly sampled patches are used. For example, [<xref ref-type="bibr" rid="B74">74</xref>] developed a deep 3D CNN with eleven layers for the segmentation of brain lesions. Since intensive computations are needed for processing 3D medical scans, [<xref ref-type="bibr" rid="B74">74</xref>] developed an efficient training scheme that joined the processing of adjacent image patches into one pass through the NN, while also adapting to the inherent class imbalance present in the data, i.e., most voxels in an MI belonging to the non-diseased class. The class imbalance problem can be addressed by defining a loss function as a weighted combination of sensitivity and specificity (see above), whereby a larger weight is attributed to specificity to make it less sensitive to the data imbalance. To incorporate both local and global information, a dual pathway architecture was used by [<xref ref-type="bibr" rid="B74">74</xref>], which simultaneously processed the input images at multiple scales while using a 3D fully connected conditional random field to effectively remove false positives and to carry out post-processing of the network&#x2019;s soft segmentation. [<xref ref-type="bibr" rid="B75">75</xref>] trained CNNs using non-uniformly sampled patches to achieve a wider coverage of the sampled regions, which captured more contextual information, proving very useful for the analysis of medical images.</p>
</sec>
</sec>
</sec>
<sec id="s3">
<label>3</label>
<title>Enhancement of the images of porous materials</title>
<p>Enhancement techniques are used to refine images of porous materials so that their important features become easier to understand and can be detected by automated image analysis systems. At a fundamental level, image enhancement represents the mapping of one image onto another, which is not necessarily one-to-one, meaning that two seemingly different images can be transformed into the same or similar output images after enhancement. The process is not without problems because the enhancement of some features in the images may produce undesirable effects. In addition, important information may be lost.</p>
<sec id="s3-1">
<label>3.1</label>
<title>Rock-like porous media</title>
<p>The use of ML algorithms in problems involving porous media can be limited by the lack of large datasets required for training the algorithms. One way to address this problem is through the enhancement of images of porous media, which allows for the expansion of the dataset and more accurate estimation of the physical properties of the media. However, the application of ML algorithms to problems involving porous materials and media, particularly their images, initially encountered challenges. For example, pixel values in an image must first be transformed into a feature vector that can be detected by the ML method. After deep-learning methods were developed in early 2006, trained using multiple levels of representation, progress began to accelerate. Raw data can then be used as input to accomplish important tasks such as the detection or classification of certain features in datasets and image enhancement.</p>
<sec id="s3-1-1">
<label>3.1.1</label>
<title>Super-resolution</title>
<p>Super-resolution (SR) [<xref ref-type="bibr" rid="B76">76</xref>, <xref ref-type="bibr" rid="B77">77</xref>] is a classical ill-posed problem studied by computer scientists, which aims to reconstruct high-resolution (HR) images from low-resolution (LR) images and directly enhance image resolution beyond hardware limitations. The problem is ill-posed because the solution that the method produces is not unique. Thus, coarse images taken across large fields of view (FOVs) are artificially enhanced to the resolution needed for carrying out accurate modeling, circumventing traditional trade-offs in analyzing multiscale porous media. In the classical reconstruction problem, one uses a given dataset for a porous medium&#x2014;images of the pore space in this case&#x2014;in order to develop a model of the pore space that not only honors the data but also, when used in numerical simulation of various phenomena in the pore space, provides accurate estimates for the properties of the porous medium [<xref ref-type="bibr" rid="B78">78</xref>]. Super-resolution generates an image from a single LR image using a feedforward CNN, referred to as the <italic>generator</italic> <inline-formula id="inf34">
<mml:math id="m38">
<mml:mrow>
<mml:mi mathvariant="script">G</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that consists of multiple layers with the usual weights <inline-formula id="inf35">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and biases <inline-formula id="inf36">
<mml:math id="m40">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, which, as usual, are determined by optimizing a loss function. If the LR image is 2D, it can, as the input, be described as a tensor of shapes <inline-formula id="inf37">
<mml:math id="m41">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf38">
<mml:math id="m42">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf39">
<mml:math id="m43">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are, respectively, the dimensions of the image and the number of color channels present. In 3D, a third dimension is added so that an image of size <inline-formula id="inf40">
<mml:math id="m44">
<mml:mrow>
<mml:mfenced open="" close=")">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> is super-resolved. A scale factor <inline-formula id="inf41">
<mml:math id="m45">
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is typically used so that one obtains, for example, an SR image of size <inline-formula id="inf42">
<mml:math id="m46">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>c</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. Generative adversarial networks (GANs) and CNNs have been utilized with SR for image enhancement.</p>
<p>[<xref ref-type="bibr" rid="B79">79</xref>] proposed an image SR using a deeply-recursive CNN with a very deep recursive layer with up to 16 recursions. Increasing recursion depth improves the performance of the algorithm without introducing new parameters for additional convolutions. However, the difficulty with their method is that training the NN using the standard gradient descent was very complex, due to exploding or vanishing gradients during the optimization of hyperparameters. [<xref ref-type="bibr" rid="B80">80</xref>] proposed a very deep fully convolutional encoding&#x2013;decoding architecture for restoring images, such as denoising and SR. Their NN consisted of multiple layers of convolution and deconvolution operators that learned end-to-end mappings from corrupted images to the original images. The convolutional layers acted as the feature extractor, capturing the abstraction of image contents while eliminating noise and corruption. Deconvolutional layers were used to recover the details of the image. Symmetrically linked convolutional and deconvolutional layers with skip connections were used. The reason for using skip connection was that, as discussed above, the performance of deep NNS deteriorates with increasing the depth of its architecture, which is referred to as <italic>the degradation problem.</italic> Skip connections were introduced to address this problem by, for example, gates that controlled and learned the flow of information to deeper layers. Using the skip-layer connection, the training converged much faster because it allowed the data to be back-propagated to the bottom layers directly and, thus, addressed the problem of gradient vanishing, which facilitated training deep NNs. In addition, the skip connections transferred image details from the convolutional to the deconvolutional layers, which is useful for recovering the original image.</p>
<p>[<xref ref-type="bibr" rid="B81">81</xref>, <xref ref-type="bibr" rid="B82">82</xref>] used CNNs to super-resolve an image with an NN that had three to five activated convolutional layers. They applied the algorithm to a bicubically upscaled LR image to recover the HR details by a direct mapping. The original architecture, which was dubbed super-resolution CNN (SRCNN), was improved iteratively with deeper, more complex NNs that improved the mapping from LR to HR images. Integrated LR-to-HR SRCNNs do not require bicubic upscaling as a processing step because the computational cost for a fully sized input image increases. [<xref ref-type="bibr" rid="B82">82</xref>] removed the deconvolutional layers at the end of the NN and replaced them with subpixel convolution in order to reduce checkerboard artifacting [<xref ref-type="bibr" rid="B83">83</xref>, <xref ref-type="bibr" rid="B84">84</xref>]. The problem of checkerboard artifacts arises when NNs use a deconvolution operation that allows them to use every point in the small image to &#x201c;paint&#x201d; a square in the larger one. However, deconvolution can easily have &#x201c;uneven overlap&#x201d; that puts more of the &#x201c;paint&#x201d; in some places more than others, particularly when the kernel size&#x2014;the output window size&#x2014;is not divisible by the stride, the spacing between points. When that happens in 2D, the uneven overlaps on the two axes multiply together and generate a characteristic checkerboard-like pattern of varying magnitudes. Deeper networks produce improved results using the skip connection that adds outputs from shallow layers to deeper layers, which preserves important shallow feature sets and improves gradient scaling.</p>
<p>[<xref ref-type="bibr" rid="B85">85</xref>] used SRCNNs for image enhancement, using DeepRock-SR [<xref ref-type="bibr" rid="B86">86</xref>], a diverse compilation of raw and processed micro-CT images in 2D and 3D. The architecture that they used was based on the enhanced deep SR (EDSR) developed by [<xref ref-type="bibr" rid="B87">87</xref>] and the SR-Residual Net [<xref ref-type="bibr" rid="B88">88</xref>] NNs. The EDSR encompasses a sequence of convolutional layers, residual blocks, and an upsampling module, which also alleviates the problem of gradient vanishing. [<xref ref-type="bibr" rid="B85">85</xref>] retained the overall EDSR architecture but replaced the ReLU activation layers with parametric ReLU (PReLU) layers from SR-ResNet in order to avoid using batch normalization layers in SRCNNs, which have been shown to be detrimental to training, while the PReLU layers provide some improvement in the performance for a slightly increased computation [<xref ref-type="bibr" rid="B89">89</xref>]. The EDSR network that [<xref ref-type="bibr" rid="B85">85</xref>] used had convolutional layers of kernel size 3 with 64 filters over 16 residual layers and skip connections throughout the layers prior to upscaling. The network was set up in both 2D and 3D forms, with the only difference being the extra deep dimension that the convolutional kernels had for the 3D case; otherwise, the architectures of the 2D and 3D NNs were structurally identical. The loss function was defined as the pixel-wise mean of the sum of the differences between the generated SR and the original &#x201c;ground-truth&#x201d; HR image:<disp-formula id="e2">
<mml:math id="m47">
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>which presumably leads to more accurate results in a shorter time. The modified EDSR was trained for 100 epochs at 1,000 iterations per epoch with mini-batches of 16 cropped <inline-formula id="inf43">
<mml:math id="m48">
<mml:mrow>
<mml:mn>192</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>192</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images in 2D and batches of 4 cropped of size <inline-formula id="inf44">
<mml:math id="m49">
<mml:mrow>
<mml:mn>96</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>96</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>96</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> in 3D using the loss function defined in <xref ref-type="disp-formula" rid="e2">Equation 2</xref> and a learning rate, <inline-formula id="inf45">
<mml:math id="m50">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>4</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> using the Adam optimizer. The total number of trainable parameters for the 3D case was approximately 198 million, as opposed to approximately 102 million for the 2D case. At the end of each epoch during the training, which consisted of 1,000 mini-batch iterations, 1,200 full-size <inline-formula id="inf46">
<mml:math id="m51">
<mml:mrow>
<mml:mn>500</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>500</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images from the 2D dataset and 300 cubes of <inline-formula id="inf47">
<mml:math id="m52">
<mml:mrow>
<mml:mn>100</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>100</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>100</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> images from the 3D dataset were used to validate the model.</p>
<p>To further test the method, particularly validating it with data that are for quantities other than those related to the quality of the SR images, such as datasets for fluid flow and transport in the pore spaces represented by the images, [<xref ref-type="bibr" rid="B90">90</xref>] used SRCNNs for enhancing images of porous samples using experimental data, which were for two distinct Bentheimer sandstone cores of diameter 1.25 cm and length 6&#x2013;7 cm, with varying centimeter-scale stratification, and consisted of <inline-formula id="inf48">
<mml:math id="m53">
<mml:mrow>
<mml:mn>6</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>m-resolution images of the entire cores in dry, brine-saturated, and partially saturated states during steady-state two-phase flow experiments. The dataset was enhanced by carrying out X-ray imaging of the cores in their dry state at varying resolution, using optical magnification, which generated paired data in four subvolumes at <inline-formula id="inf49">
<mml:math id="m54">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x2212;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf50">
<mml:math id="m55">
<mml:mrow>
<mml:mn>6</mml:mn>
<mml:mo>&#x2212;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf51">
<mml:math id="m56">
<mml:mrow>
<mml:mn>18</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> m resolution. The paired data from core 1, subvolume 1, were used for training, while the other three subvolumes were utilized for validation. Image patches were extracted from the LR and HR data with sizes <inline-formula id="inf52">
<mml:math id="m57">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf53">
<mml:math id="m58">
<mml:mrow>
<mml:mn>9</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> voxels, respectively, generating 8,000 patches, from which 6,400 patches were used for training and 1,600 patches for validation. The patch size was selected to cover the largest typical pores present in the Bentheimer samples. A 3D EDSR model was again used. Moreover, the number of residual blocks and filters (12 residual blocks and 24 filters in each convolutional layer) was reduced to improve computational efficiency and mitigate overfitting to artifacts apparent in the HR micro-CT images. The Adam optimization algorithm was used for training with a batch size of 32 and a total of 50 epochs, selected after some preliminary computations. The learning rate was initially <inline-formula id="inf54">
<mml:math id="m59">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and was decreased tenfold every 10 epochs to consolidate the training process. The loss function utilized was defined as follows:<disp-formula id="e3">
<mml:math id="m60">
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">C</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">C</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo stretchy="false">&#x7c;</mml:mo>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>
<xref ref-type="disp-formula" rid="e3">Equation 3</xref> is similar to <xref ref-type="disp-formula" rid="e1">Equation 1</xref>. Here, <inline-formula id="inf55">
<mml:math id="m61">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">C</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">H</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf56">
<mml:math id="m62">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="normal">C</mml:mi>
<mml:mi mathvariant="normal">T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="normal">S</mml:mi>
<mml:mi mathvariant="normal">R</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the pixel grayscale CT values for the HR and generated SR images, respectively. Convergence was achieved after approximately 10 epochs. After training, the EDSR network was utilized to generate SR images with a threefold increase in resolution from any input LR images. The entire algorithm is shown in <xref ref-type="fig" rid="F7">Figure 7</xref>. The EDSR was then used to generate the corresponding SR images for the unseen LR images from those samples that were not used in the training.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Deep-learning workflow and testing and training of the algorithm. <bold>(a)</bold> EDSR network architecture. <bold>(b)</bold> Example testing images, where the LR image is at <inline-formula id="inf57">
<mml:math id="m63">
<mml:mrow>
<mml:mn>6</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>m resolution, while the HR image is at <inline-formula id="inf58">
<mml:math id="m64">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>&#x3bc;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>m resolution. <bold>(c)</bold> Loss function for the EDSR network training and testing over 50 epochs. <inline-formula id="inf59">
<mml:math id="m65">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>L</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> refers to a loss function similar to that shown in <xref ref-type="disp-formula" rid="e2">Equation 2</xref> (after [<xref ref-type="bibr" rid="B90">90</xref>]; courtesy of Professor Peyman Mostaghimi).</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g007.tif">
<alt-text content-type="machine-generated">The deep-learning workflow, and testing and training of the algorithm for super-resolution. Panel (a) shows the enhanced deep super-resolution network architecture; (b) shows two testing images, one with low resolution (LR) and another one with high resolution (HR), while (c) presents the loss function L1.</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="s3-2">
<label>3.2</label>
<title>Biological porous media</title>
<p>Enhancement of images to higher resolutions, given images with lower resolution, in order to obtain more data with good accuracy, is an important problem in medical physics that has been studied for a long time. The ML-based algorithms that were described above for enhancing images of RLPM can also be used in the context of MIs. [<xref ref-type="bibr" rid="B91">91</xref>] utilized multi-stream networks to reconstruct high-resolution cardiac MRI from one or more low-resolution 3D MRI images, precisely the same problem that was studied by [<xref ref-type="bibr" rid="B92">92</xref>] in the context of RLPM, which is described in the next section. Enhancement of medical images also has other applications. For example, one can not only recover missing spatial information but also utilize the enhanced image to infer advanced MRI diffusion parameters from limited data, as demonstrated by [<xref ref-type="bibr" rid="B93">93</xref>]. Comprehensive recent reviews are provided by [<xref ref-type="bibr" rid="B94">94</xref>] and [<xref ref-type="bibr" rid="B95">95</xref>].</p>
</sec>
</sec>
<sec id="s4">
<label>4</label>
<title>Data augmentation and reconstruction of porous materials</title>
<p>As already mentioned, the use of ML algorithms for analyzing images and extracting information from them may be impeded by the fact that the number of available images may be severely limited, while training any ML algorithm essentially requires extensive data. In such cases, data augmentation is used, which, in the present context, refers to expanding the database by generating synthetic data based on the available images. One way to do this is through the classical problem of reconstruction, which is defined formally as, <italic>given a limited amount of data for a porous medium, how can one reconstruct its realizations that not only honor the data but also provide accurate predictions for its properties?</italic> It should be clear that, unless one copies a 3D image of a porous medium, one can never reconstruct its exact replica. In other words, any reconstruction method can only generate plausible realizations of the original image whose accuracy depends on the reconstruction procedure and the amount of available data. This also implies that no reconstruction method can provide a <italic>unique</italic> solution to the problem. Several approaches to the reconstruction problem have been comprehensively reviewed and discussed by [<xref ref-type="bibr" rid="B78">78</xref>]. In this section, we describe two ML-based algorithms to address the problem of reconstruction of images of porous materials, both of RLPM and biological types. Both methods have been proven to be accurate.</p>
<sec id="s4-1">
<label>4.1</label>
<title>Rock-like porous media</title>
<p>One ML-based approach to the reconstruction of porous media uses generative adversarial networks (GANs), developed by [<xref ref-type="bibr" rid="B96">96</xref>, <xref ref-type="bibr" rid="B97">97</xref>]. Assume that a digital image <bold>DI</bold> is a sample of a real probability distribution function (PDF) of the images <inline-formula id="inf60">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, of which we have a number of sub-domains that are extracted from <bold>DI</bold> without any overlap. Every GAN consists of a pair of NNs, a <italic>generator</italic> <inline-formula id="inf61">
<mml:math id="m67">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, defined by its parameter <inline-formula id="inf62">
<mml:math id="m68">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> that generates realizations of the unknown PDF <inline-formula id="inf63">
<mml:math id="m69">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, and a <italic>discriminator</italic> <inline-formula id="inf64">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, with <inline-formula id="inf65">
<mml:math id="m71">
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> being a parameter that tries to distinguish between the realizations of the training set and the synthetic ones that are generated by <inline-formula id="inf66">
<mml:math id="m72">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The latter learns to classify inputs as either real or false, whereas the former generates sufficiently realistic fake inputs to the discriminator so that they are classified as real. For example, one feeds random noise into a generator that is trained to transform the noise into data that closely mimic the real data from the training set. The discriminator is trained to distinguish the actual training dataset from the generated data. As the training of the two NNs progresses, the discriminator becomes more accurate at identifying the fake data, whereas the generator improves its ability to generate fake data for deceiving the discriminator. Thus, GANs are capable of generating realizations from an arbitrary PDF without any restriction on its form, which is the feature that makes them a powerful method for reconstruction. This is particularly true when the input for reconstruction is an image. The generator performs a mapping from a random prior <bold>Z</bold> onto the domain of the image, defined as follows.<disp-formula id="e4">
<mml:math id="m73">
<mml:mrow>
<mml:mi>Z</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:mrow>
<mml:mi mathvariant="script">N</mml:mi>
</mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>0,1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>
<disp-formula id="e5">
<mml:math id="m74">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mo>&#x2009;&#x2009;</mml:mo>
<mml:mi mathvariant="bold">Z</mml:mi>
<mml:mo>&#x2192;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msup>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>In <xref ref-type="disp-formula" rid="e5">Equation 5</xref> N represents the normal distribution, while d in <xref ref-type="disp-formula" rid="e6">Equation 6</xref> is the dimension of <bold>Z</bold>, with <inline-formula id="inf68">
<mml:math id="m76">
<mml:mrow>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> being the linear size of a subdomain extracted from the original image. <inline-formula id="inf70">
<mml:math id="m78">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> assigns a probability to an image <bold>DI</bold> from a sample drawn from <inline-formula id="inf71">
<mml:math id="m79">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the distribution of the true data: <inline-formula id="inf72">
<mml:math id="m80">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>:</mml:mo>
<mml:mspace width="2.77695pt" class="tmspace"/>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="double-struck">R</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>n</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. Values close to 1 represent a high probability of being a sample of <inline-formula id="inf73">
<mml:math id="m81">
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>d</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Both the generator and discriminator must be optimized, which can be carried out by a variety of methods. One can, for example, represent both <inline-formula id="inf74">
<mml:math id="m82">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">Z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf75">
<mml:math id="m83">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> using differentiable NNs with parameters <inline-formula id="inf76">
<mml:math id="m84">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>and <inline-formula id="inf77">
<mml:math id="m85">
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. The optimization process is carried out in two steps:<list list-type="roman-lower">
<list-item>
<p>First, the discriminator is trained in order to maximize its ability to distinguish between real and fake samples. Training is carried out by a supervised deep-learning algorithm using the known real samples (label 1) and the realizations generated by <inline-formula id="inf78">
<mml:math id="m86">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (label <inline-formula id="inf79">
<mml:math id="m87">
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>). The loss (cost) function to compute the misclassification error is given by:</p>
</list-item>
</list>
<disp-formula id="e6">
<mml:math id="m88">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munder>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:munder>
</mml:mstyle>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2061;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mfenced>
<mml:mi>log</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(6)</label>
</disp-formula>
<xref ref-type="disp-formula" rid="e6">Equation 6</xref> defines the loss function as a cross-entropy in which y&#x2019; contains the output probability assigned by the discriminator to each element of a given mini-batch of samples. Thus, <inline-formula id="inf80">
<mml:math id="m89">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is optimized for each mini-batch of the real images, whereas <inline-formula id="inf81">
<mml:math id="m90">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is optimized for all the fake samples. The error is back-propagated while keeping the parameters of the generator constant.<list list-type="simple">
<list-item>
<label>ii.</label>
<p>Next, the generator is trained in order to maximize its ability to &#x201c;deceive&#x201d; the discriminator into misclassifying the images produced by the generator as real images. Training is carried out by computing the binary cross-entropy of the discriminator&#x2019;s output on a mini-batch sampled from the generator <inline-formula id="inf82">
<mml:math id="m91">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>G</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b8;</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold">Z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, requiring that the created labels be close to 1, thereby computing <inline-formula id="inf83">
<mml:math id="m92">
<mml:mrow>
<mml:mi>H</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>,</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mi mathvariant="bold">y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2032;</mml:mo>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, which is optimized by varying the parameters of the generator while keeping the parameters of <inline-formula id="inf84">
<mml:math id="m93">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3c9;</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> constant. The computations required for the optimization are intensive and cannot be performed by a single processor for large samples. For 3D images, the memory requirements are also substantial. [<xref ref-type="bibr" rid="B98">98</xref>, <xref ref-type="bibr" rid="B99">99</xref>] used GANs to reconstruct a variety of RLPM.</p>
</list-item>
</list>
</p>
<p>By combining an SRCNN, representing the generator, with an image classification network as the discriminator, an SRGAN is formed. Although SRGANs produce features that appear realistic, due to pixel-wise mismatch, the accuracy of the generated SR images is lower than that of SRCNN. Since micro-CT images usually contain significant amounts of image noise and texture that manifest themselves as high-frequency features, they are also inadvertently recovered. According to [<xref ref-type="bibr" rid="B100">100</xref>], SRCNNs usually suffice because they contain some form of intrinsic noise suppression while maximizing edge recovery. [<xref ref-type="bibr" rid="B85">85</xref>] used the same dataset, DeepRock-SR, with SRGAN and compared its performance with that of SRCNN. <xref ref-type="fig" rid="F7">Figure 7</xref> shows the comparison of the results with those produced by SRCNN and BC-generated images.</p>
<p>The second approach is the CCSIM method [<xref ref-type="bibr" rid="B39">39</xref>&#x2013;<xref ref-type="bibr" rid="B42">42</xref>], an ML approach without using any NN that represents the digital image <bold>DI</bold> using a computational grid <bold>G</bold>, partitioned into overlapping blocks of sizes <inline-formula id="inf85">
<mml:math id="m94">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, where <bold>G</bold> and <bold>DI</bold> have the same sizes. The neighboring blocks share overlapping regions <bold>OL</bold> with sizes <inline-formula id="inf86">
<mml:math id="m95">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (see <xref ref-type="fig" rid="F8">Figure 8</xref>). One then begins from a corner block of <bold>G</bold> (or any other block) and visits each grid block along a 1D raster path. For each grid block, a pattern of heterogeneity from <bold>DI</bold> is selected at random and inserted in the visiting block. The inserted pattern is referred to as the <italic>data event</italic> <inline-formula id="inf87">
<mml:math id="m96">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, with the word &#x201c;event&#x201d; implying that the inserted pattern of heterogeneity in the block may change again. Then, the next pattern is selected based on the similarity between the neighborhood points and <bold>DI</bold>, meaning that, instead of considering all the previously constructed blocks (by filling them with patterns of heterogeneity from <bold>DI</bold>), only those in the neighborhood of the current blocks are used for the calculations. Next, the similarity between, or closeness to, the neighboring blocks and <bold>DI</bold> is quantified based on the a cross-correlation function that represents a convolution between <bold>DI</bold> and <inline-formula id="inf88">
<mml:math id="m97">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>:<disp-formula id="e7">
<mml:math id="m98">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>j</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(7)</label>
</disp-formula>which represents a convolution of <inline-formula id="inf89">
<mml:math id="m99">
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> with <inline-formula id="inf90">
<mml:math id="m100">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf91">
<mml:math id="m101">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="" close=")">
<mml:mrow>
<mml:mfenced open="[" close="">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf92">
<mml:math id="m102">
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mfenced open="" close=")">
<mml:mrow>
<mml:mfenced open="[" close="">
<mml:mrow>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:math>
</inline-formula>. Thus, an overlap region of size <inline-formula id="inf93">
<mml:math id="m103">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#xd7;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> between two neighboring blocks and a data event <inline-formula id="inf94">
<mml:math id="m104">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are used to match the patterns in the <bold>DI</bold>. The overlap region contains a set of pixels or voxels selected from the previously constructed blocks and uses them in <xref ref-type="disp-formula" rid="e7">Equation 7</xref> to identify the next pattern of heterogeneity. For Euclidean distance (difference) between the constructed block and the data to be minimum, <inline-formula id="inf95">
<mml:math id="m105">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> must be maximum or, in practice, exceed a preset threshold. After calculating <inline-formula id="inf96">
<mml:math id="m106">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> and selecting those patterns for which <inline-formula id="inf97">
<mml:math id="m107">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> exceeds the preset threshold, one of the acceptable ones is selected at random and inserted in the block currently being visited in <bold>G</bold>. The process is repeated until all the blocks of the grid <bold>G</bold> have been reconstructed. As a rule of thumb, the neighboring regions might have an overlap of size of approximately 1/5&#x2013;1/6 of the size of the blocks. Large grid blocks increase the computations as computing <inline-formula id="inf98">
<mml:math id="m108">
<mml:mrow>
<mml:mi>&#x3c8;</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>;</mml:mo>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> requires considering many points, whereas small regions may cause discontinuity in the patterns. The method has been shown to successfully reconstruct models of various complex 2D and 3D porous media based on their images.</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Schematic illustration of the CCSIM algorithm. Top left: arrows indicate the direction of the raster path. One randomly selected patch of the training image is selected at random as the data event <inline-formula id="inf99">
<mml:math id="m109">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>D</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>T</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> (see the next) to fill one black of the computational grid G. Top right: One realization of the image on the top left, generated by the CCSIM method. At the bottom, the panel in <bold>(a)</bold> represents the input image, while the three panels in <bold>(b)</bold> show three realizations generated by the CCSIM.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g008.tif">
<alt-text content-type="machine-generated">Schematic illustration of the CCSIM algorithm showing a training image on the left with arrows and patterns indicating analysis steps, and a similar-style output image titled &#x22;One realization&#x22; on the right. Below are three smaller images labeled (a), (b), and (c), each displaying varying patterns and textures.</alt-text>
</graphic>
</fig>
<p>[<xref ref-type="bibr" rid="B92">92</xref>] proposed a method for enhancing images of porous materials and computing its effective properties (such as their permeability) when only a limited number of HR images&#x2014;as few as one&#x2014;are available. The method is based on data augmentation and generating a large and diverse dataset for the training of a CNN. [<xref ref-type="bibr" rid="B92">92</xref>] used the CCSIM algorithm to generate a large and diverse dataset to be used for training the CNN, and thus, they dubbed their approach a hybrid stochastic-CNN (HS-CNN) method. They developed their approach for shales, one of the most complex porous media, but the method is applicable to any type of porous material. They used a limited number of images&#x2014;30 2D images with a size of <inline-formula id="inf100">
<mml:math id="m110">
<mml:mrow>
<mml:mn>500</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>500</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>&#x2014;with the CCSIM algorithm to generate 2,000 plausible realizations of the same, which were then used as the training images with the CNN. The size of 500 of the output realizations was <inline-formula id="inf101">
<mml:math id="m111">
<mml:mrow>
<mml:mn>1000</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1000</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, with the remaining having the same size as the original images. The success of any learning algorithm depends heavily on the number and diversity of the images that are used for its training. Thus, to diversify, the 2,000 training images were transformed using various filters, such as the Gaussian noise, scale, crop, rotation, and flip to increase the diversity between the training datasets. The CNN uses the high-quality training data and extracts their differences with the upscaled images of the same set, using a residual learning strategy. The training images all had high resolution, whereas the image at hand that required enhancement was a LR one, i.e., one with a smaller size. Therefore, the target image was enlarged, i.e., upscaled, to the size of the training images by using an interpolation method with bicubic functions. Then, the CNN learned iteratively how to estimate the residuals. After the training was complete, i.e., after the CNN learned how to estimate the residuals, the high-quality image was reconstructed by adding the original enlarged image to the estimated residuals. Therefore, the upscaled images were used as the input, whereas the estimated residuals represented the output. The CNN that [<xref ref-type="bibr" rid="B92">92</xref>] used had 22 individual layers that learned mapping of HR training images onto the LR input image. The utilized images were similar in their content but differed in their details. To reduce the computational time, 64 random patches of size <inline-formula id="inf102">
<mml:math id="m112">
<mml:mrow>
<mml:mn>41</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>41</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> from the training data were selected from each training image and used in the NN, rather than utilizing the full-size images. While such patches can jeopardize the capture of large-scale structures in the images, the use of the aforementioned filters&#x2014;particularly, the scaling operators, ensured that the large-scale structures were still accounted for. The constructed small patches were then fed to the CNN over several iterations.</p>
<p>The LR image represented the first layer. The next layer was the convolution layer that contained 64 filters of size <inline-formula id="inf103">
<mml:math id="m113">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, thereby assigning one filter to each patch. The performance of the CNNs improves with increasing the number of filters, but the training time also increases. Choosing a smaller filter size is usually preferred in the sense that it reduces the computational time, but it may also result in missing the large-scale structures of the image. On the other hand, larger filters complicate the training process. Thus, the optimal filter size should be decided <italic>a priori</italic> or selected by trial and error. Zero padding was also used in order to generate feature maps with the same input layer. All the convolution layers, except the last one, contained the ReLU activation function. The initial weights were assigned randomly and optimized during the learning. The first layer applied the Softplus, or SmoothReLU, function as the rectifier to the input values without changing their depth information, given by <xref ref-type="disp-formula" rid="e8">Equation 8</xref>:<disp-formula id="e8">
<mml:math id="m114">
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>log</mml:mi>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>exp</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(8)</label>
</disp-formula>
</p>
<p>The last layer of the CNN had a single filter of size <inline-formula id="inf104">
<mml:math id="m115">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>50</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, followed by a regression layer that calculated the loss function defined be <xref ref-type="disp-formula" rid="e9">Equation 9</xref>, given by:<disp-formula id="e9">
<mml:math id="m116">
<mml:mrow>
<mml:mi mathvariant="script">L</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msup>
<mml:mrow>
<mml:mfenced open="[" close="]">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>Resid</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>HS</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>CNN</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mi>y</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
</mml:mrow>
</mml:mfrac>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(9)</label>
</disp-formula>where <inline-formula id="inf105">
<mml:math id="m117">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mtext>Resid</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> represents the value of each pixel (voxel) at <inline-formula id="inf106">
<mml:math id="m118">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for the residual image <inline-formula id="inf107">
<mml:math id="m119">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>Resid</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and similarly for image predicted by the HS-CNN algorithm, <inline-formula id="inf108">
<mml:math id="m120">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="bold">D</mml:mi>
<mml:mi mathvariant="bold">I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mtext>HS</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>CNN</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>The CNN was trained with stochastic gradient descent momentum optimization. Since one deals with complex images and very large datasets, the gradient descent method does not by itself solve the problem of accurately estimating the optimal values of the weights <inline-formula id="inf109">
<mml:math id="m121">
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. Hence, the algorithm was used with momentum at each iteration, with the momentum defined as the moving average of the gradients; this approach retains the update <inline-formula id="inf110">
<mml:math id="m122">
<mml:mrow>
<mml:mi mathvariant="normal">&#x394;</mml:mi>
<mml:mi>w</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, the difference between the weights in two consecutive iterations, which is then used in the subsequent update. The momentum was used to prevent the optimization computation from converging to a local minimum or a saddle point. A high momentum parameter accelerates the speed of convergence to the true minimum of the loss function, but setting the momentum parameter at too high a value may give rise to overshooting the true global minimum, thereby making computations unstable. On the other hand, if the momentum coefficient is too low, it cannot reliably help the minimization in order to avoid local minima, thereby slowing down the training. A moving average was also used because the data were just images.</p>
<p>
<xref ref-type="fig" rid="F9">Figure 9</xref> presents a visual comparison of the results, indicating excellent accuracy of the HS-CNN algorithm. The initial LR image represents a very smooth and opaque view of the pores in a shale sample, whereas the enhanced image generated by the HS-CNN reveals features as observed in the reference image. Similarly, the image generated by the BC interpolation method is a smooth reconstructed image. As expected, the CNN alone, without the large dataset produced by CCSIM algorithms, did not perform well. Note, however, that in the modeling of porous materials, one usually has only a few HR images, which are often insufficient. [<xref ref-type="bibr" rid="B92">92</xref>] also made a quantitative comparison using various statistical measures. One measure, for example, was peak signal-to-noise ratio (PSNR) <inline-formula id="inf111">
<mml:math id="m123">
<mml:mrow>
<mml:mi mathvariant="script">R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>, which was used with the images generated by both the bicubic interpolation and the HS-CNN algorithm and compared with that of the reference high-resolution image. <inline-formula id="inf112">
<mml:math id="m124">
<mml:mrow>
<mml:mi mathvariant="script">R</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is defined as follows:<disp-formula id="e10">
<mml:math id="m125">
<mml:mrow>
<mml:mi mathvariant="script">R</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>10</mml:mn>
<mml:mo>&#x2061;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>log</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:msubsup>
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>I</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(10)</label>
</disp-formula>where <inline-formula id="inf113">
<mml:math id="m126">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>M</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>I</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> denotes the maximum possible pixel value in the image <bold>DI</bold> and <inline-formula id="inf114">
<mml:math id="m127">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> is the mean-squared error. The computed statistical measures, such that given by <xref ref-type="disp-formula" rid="e10">Equation 10</xref>, indicated the high accuracy of the enhanced images.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Comparison of <bold>(a)</bold> the reference CT image and <bold>(b)</bold> the low-resolution CT input, with the image obtained by <bold>(c)</bold> regular CNN; <bold>(d)</bold> the bicubic interpolation; and <bold>(e)</bold> the HS-CNN algorithm. For better comparison, a zoom-in portion is also shown. The zoomed-in portions in each image, corresponding to the top five images in the same order, are also shown for comparison. <bold>(f&#x2013;i)</bold> show the zoomed-in part with various resolutions, and compare the accuracy of the methods, with <bold>(j)</bold> obtained by the HS-CNN method described.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g009.tif">
<alt-text content-type="machine-generated">Series of ten microscopic images labeled (a) to (j). Images (a) to (e) depict similar patterns with red boxes highlighting specific sections. Images (f) to (j) focus on magnified details of the highlighted sections, showing varying levels of contrast and detail within the same pattern. All images include a textured grey background with dark geometric and irregular shapes.</alt-text>
</graphic>
</fig>
</sec>
<sec id="s4-2">
<label>4.2</label>
<title>Biological porous media</title>
<p>Every reconstruction method developed for RLPM can also be applied to biological materials, with the results used for training ML algorithms. As an example, <xref ref-type="fig" rid="F10">Figure 10</xref> presents a 2D MRI image of a baby&#x2019;s brain and three realizations reconstructed from it using the CCSIM algorithm. [<xref ref-type="bibr" rid="B101">101</xref>] used a GAN to generate synthetic chest X-rays for infections caused by COVID-19 to train the model. The training data consisted of actual and synthetic chest X-ray images, which were then fed into a customized CNN for deep learning-based chest radiograph classification in order to distinguish the COVID-19 cases from patients with regular pneumonia, along with normal cases free of the disease. The accuracy of the model for detecting COVID-19 cases were 93.94 and 54.55 percent, respectively, with and without data augmentation, i.e., without generating synthetic data by GAN reconstruction, thereby indicating the accuracy of the GAN.</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>MRI image of a baby&#x2019;s brain and three realizations of the images reconstructed by the CCSIM method.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g010.tif">
<alt-text content-type="machine-generated">Two-dimensional MRI image of a baby&#x2019;s brain (top) and three realizations reconstructed from it using the CCSIM algorithm (bottom).</alt-text>
</graphic>
</fig>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Extracting features from an image</title>
<p>To use images of porous media to compute their effective properties, one must first extract the most important features of the images that are at the level of the pixels. Feature extraction is one of the most important and also one of the most challenging problems in image analysis. In addition to extracting the features, one must also interpret them meaningfully. This is, of course, true of any digitized image, but it is true for images of porous materials. Thus, we describe this aspect of the problem.</p>
<sec id="s5-1">
<label>5.1</label>
<title>Rock-like porous media</title>
<p>We describe the problem using a concrete example and then briefly discuss related works. [<xref ref-type="bibr" rid="B102">102</xref>] used images of sandstone to link the effective permeability of the porous media to their morphology. Their proposed network was composed of two ML structures, namely, a CNN for feature extraction and a regular fully connected NN for estimating the permeability. We describe in detail the CNN that extracted the features. The filters in the convolution layers slide along the input data, with each producing a specific feature map. The input data, 3D images labeled with their permeabilities, were then convolved using 3D filters, each of size <inline-formula id="inf115">
<mml:math id="m128">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>k</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>. For simplicity of describing the procedure, the following equation was used, which represents a linear mapping<disp-formula id="e11">
<mml:math id="m129">
<mml:mrow>
<mml:mi>g</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>m</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>m</mml:mi>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mi>w</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>b</mml:mi>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
<label>(11)</label>
</disp-formula>where <inline-formula id="inf116">
<mml:math id="m130">
<mml:mrow>
<mml:mi>f</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf117">
<mml:math id="m131">
<mml:mrow>
<mml:mi>g</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> represent, respectively, the input and output; <inline-formula id="inf118">
<mml:math id="m132">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> are the coordinates of the voxels in the input; and <inline-formula id="inf119">
<mml:math id="m133">
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf120">
<mml:math id="m134">
<mml:mrow>
<mml:mi>b</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> are the weights and biases, respectively. Each set of the 3D filters produces a 3D output image, which is the input for the next convolutional layer. Note that in <xref ref-type="disp-formula" rid="e11">Equation 11</xref>, only the bias does not depend on the location; otherwise, every pixel value <inline-formula id="inf121">
<mml:math id="m135">
<mml:mrow>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is multiplied by a weight, and therefore, the resulting feature map depends on the location. If, however, two pixels have the same value in the input data and the weights are also the same, then the output will also be the same. This problem is addressed by the pooling layer that CNNs usually contain. The convolutional layer convolves the input images using the filters in order to extract their key features. To filter part or all of the input data, they are multiplied by the corresponding matrices, with the results summed up and placed in a single output pixel, as shown in <xref ref-type="fig" rid="F11">Figure 11</xref>, where a <inline-formula id="inf122">
<mml:math id="m136">
<mml:mrow>
<mml:mn>2</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> filter was used. The process is repeated by sliding the filters along the input data using the stride operation in order to generate new output features. Since the filter moves one unit along the data, the stride is one. The entire process of multiplying and summing up represents the convolution operation as it is similar to the classical convolution. In principle, one may require several filters because image recognition requires various feature maps, and multiple filters capture their essential properties. [<xref ref-type="bibr" rid="B102">102</xref>] used fifty <inline-formula id="inf123">
<mml:math id="m137">
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> filters. As <xref ref-type="fig" rid="F11">Figure 11</xref> indicates, the pixels at the center have more influence in the convolution than those on the border of the input matrix (the 3D images), which can be problematic. To remedy the issue and take the effect of the marginal pixels into account, the edges were padded with extra zero-value pixels. The activation function <inline-formula id="inf124">
<mml:math id="m138">
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> used was ReLU, which was applied to the extracted features at each pixel <inline-formula id="inf125">
<mml:math id="m139">
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> to produce their maps <inline-formula id="inf126">
<mml:math id="m140">
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Therefore, several maps were produced that hopefully captured all the important features that connected the input data to the output.</p>
<fig id="F11" position="float">
<label>FIGURE 11</label>
<caption>
<p>Feature map is generated by applying convolution and the activation function ReLU to the input data.</p>
</caption>
<graphic xlink:href="fphy-13-1742403-g011.tif">
<alt-text content-type="machine-generated">Diagram illustrating a convolutional neural network process. The input data matrix is convolved with a kernel/filter in the convolution layer, producing an intermediate matrix. This matrix is transformed by ReLU, resulting in a feature map.</alt-text>
</graphic>
</fig>
<p>Despite their effectiveness in extracting various features from the input images, the resulting maps may still be limited in terms of recording features&#x2019; locations in the input data. This means that if the same pattern appears in various locations in the input data due to, for example, cropping, rotation, resolution, and size changes, then its feature map will be different after the convolution layer, thereby giving rise to a major problem with the convolved maps, namely, their sensitivity to the features&#x2019; location in the input. The reason is that the filters in the convolutional layers are applied with very small strides, usually stride 1, and thus, the maps generated by the convolutional layer will have very similar locations for each feature in the input. Hence, when a convolutional filter is applied and slid over the image, the output feature map may still depend on the data in each region. To resolve the problem, one may downsample, or coarsen, the feature maps, making them more robust to changes in the features&#x2019; position using the next layer of the CNN after the activation, the pooling layer, that downsamples the feature maps by summarizing the presence of the features in the map&#x2019;s patches, while preserving the essential features of the data. This not only reduces the size of the data entering the fully connected layer but also reduces the computation. Thus, the pooling layer is very useful when one must preserve a specific feature without considering its exact location in the image because the layer combines all the semantically similar features. Reducing the dimensionality of the feature map also reduces the sensitivity to the location. Note that the pooling operation is specified, rather than learned. The pooling layer acts on each feature map separately in order to generate a new set of the same number of pooled feature maps. The size of the pooling operation or filter is smaller than that of the feature map. Two common functions used in the pooling operation are average pooling, which calculates the average value for each patch on the feature map, and maximum pooling (max-pooling), which computes the maximum value for each patch of the feature map. [<xref ref-type="bibr" rid="B98">98</xref>] used the former, which is defined as follows:<disp-formula id="equ4">
<mml:math id="m141">
<mml:mrow>
<mml:mi>g</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>z</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>&#x3d;</mml:mo>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>j</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mrow>
<mml:mo>&#x2211;</mml:mo>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:munderover>
</mml:mstyle>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mfrac>
<mml:mrow>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:mfenced>
<mml:mi>f</mml:mi>
<mml:mfenced open="(" close=")">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mi>x</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>i</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>y</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>j</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>s</mml:mi>
<mml:mi>z</mml:mi>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>k</mml:mi>
</mml:mrow>
</mml:mfenced>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</disp-formula>where <inline-formula id="inf127">
<mml:math id="m142">
<mml:mrow>
<mml:mi>s</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> is the coarsening length scale.</p>
<p>The output of the pooling layers was then flattened to a 1D vector to produce the input to the fully connected layer, the final layer in the proposed network, before the output layer. The fully connected layer keeps a collection of the most important outputs from all the convolutional layers. Since the purpose of the network was regression, the number of neurons in the output layer depended on the number of targets, i.e., the permeabilities to be predicted. Each neuron in the output layer had a continuous value that represented the prediction for each target. All the feature maps in the last convolutional layer were connected to a unit in a fully connected layer, with the output layer producing the prediction <inline-formula id="inf128">
<mml:math id="m143">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>&#x2014;the permeability for image <inline-formula id="inf129">
<mml:math id="m144">
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula>&#x2014;calculated by a logistic regression layer, <inline-formula id="inf130">
<mml:math id="m145">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>w</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:msub>
<mml:mrow>
<mml:mi>v</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>i</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, where <inline-formula id="inf131">
<mml:math id="m146">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi mathvariant="script">A</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>R</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the activation function of the regression layer, taken to be the logistic function, <inline-formula id="inf132">
<mml:math id="m147">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">[</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>exp</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>x</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>This basic scheme has been used by many to link various properties of RLPM to their morphology, including diffusivity, elastic moduli, and the dispersion coefficient in solute transport through porous media [<xref ref-type="bibr" rid="B103">103</xref>].</p>
</sec>
<sec id="s5-2">
<label>5.2</label>
<title>Biological porous media</title>
<p>As mentioned above, extracting feature representations from images at the pixel level is a difficult problem, particularly for MIs. Since CNNs have the capability to learn meaningful features at multiple levels and classify them, their use for feature extraction from medical images has naturally attracted attention. Extracting features from MIs falls within a general class of techniques called content-based image retrieval (CBIR)&#x2014;also referred to as query by image content (QBIC) and content-based visual information retrieval (CBVIR)&#x2014;which use computer vision techniques to address the problem of image retrieval, or the search for digital images in large databases. &#x201c;Content&#x201d; in this context is a reference to colors, shapes, textures, and similar information that can be extracted from the image, and therefore, &#x201c;content-based&#x201d; implies analyzing the contents of the image, rather than such metadata as descriptions associated with the image. A comprehensive review of the application of ML algorithms to the CBIR problem is provided by [<xref ref-type="bibr" rid="B104">104</xref>].</p>
<p>[<xref ref-type="bibr" rid="B105">105</xref>] combined feature descriptors of CNNs to extract <inline-formula id="inf133">
<mml:math id="m148">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> features for overlapping patches in 3D MRIs for the prostate and to construct a large feature matrix over all the data volumes. Hashing forests are an approach for mapping a query onto a target in a database, enabling similar nearest neighbors to be searched for and retrieved efficiently and accurately. This is typically done by learning a hashing function [<xref ref-type="bibr" rid="B106">106</xref>] that maps the database entries onto compact binary codes in the Hamming space. A hashing forest is a supervised variant of random forests developed for the task of nearest neighbor retrieval. A Hamming space is the set of all <inline-formula id="inf134">
<mml:math id="m149">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>2</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> binary strings or vectors of length <inline-formula id="inf135">
<mml:math id="m150">
<mml:mrow>
<mml:mi>N</mml:mi>
</mml:mrow>
</mml:math>
</inline-formula> in which two binary strings are considered to be adjacent if they differ only in one position. Then, the distance between any two binary strings, called the Hamming distance, is the number of positions at which the corresponding bits are different. [<xref ref-type="bibr" rid="B107">107</xref>] used a CNN with five layers to extract features from the fully connected layers of the CNN. They utilized the last layer and a pre-trained CNN and fed the extracted features to a support vector machine classifier to obtain the distance metric. The conclusion was that incorporating gender information produces better performance than just CNN features.</p>
</sec>
</sec>
<sec id="s6">
<label>6</label>
<title>Applications</title>
<p>The use of MIs in the diagnosis of ill patients has a long history. The use of digital images in analyzing problems with various phenomena in RLPM is much more recent. As computational power increased dramatically over the past three decades and new computational approaches for calculating various properties of RLPM, such as the lattice Boltzmann method [<xref ref-type="bibr" rid="B108">108</xref>], were developed to simulate various phenomena directly within images, the use of such images has increased significantly. Although the application of simple NNs for regression models correlating various properties of porous media dates back at least 30 years, it was the development of CNNs and other types of NNs and ML algorithms that spurred their widespread use in problems associated with RLPM. An extensive review of ML algorithms to RLPM, as well as problems associated with environmental issues, is provided by [<xref ref-type="bibr" rid="B109">109</xref>] and need not be repeated here, while [<xref ref-type="bibr" rid="B110">110</xref>] and [<xref ref-type="bibr" rid="B95">95</xref>] presented reviews of ML algorithms to MIs. Thus, we describe the applications of the methodologies described above to one particular class of biological porous media, namely, the human brain, which is the research area of one of authors (A.S.).</p>
<sec id="s6-1">
<label>6.1</label>
<title>Brain images</title>
<p>The three most common types of brain lesions are gliomas, meningiomas, and pituitary tumors. Thus, we first describe each type of lesion and then discuss the progress that has been made in using ML methods for analyzing and interpreting MIs associated with brain tumors.</p>
<p>According to Mayo Clinic [<xref ref-type="bibr" rid="B111">111</xref>], &#x201c;Glioma is a growth of cells that starts in the brain or spinal cord. The cells in a glioma resemble healthy brain cells called glial cells. Glial cells surround nerve cells and help them function. As a glioma grows, it forms a mass of cells called a tumor. The tumor can grow to press on the brain or spinal cord tissue and cause symptoms.&#x201d; Gliomas are primary tumors that grow within the central nervous system; a subset of them, glioblastoma (GBM), is the most common primary adult brain tumor. Diagnosis of GBM involves a combination of neurological examination, advanced neuroimaging, such as MRI, and histopathological analysis of composite subtypes after surgical resection or biopsy [<xref ref-type="bibr" rid="B112">112</xref>]. The current &#x201c;gold standard&#x201d; of treatment for GBM is radiation therapy (RT) with concurrent temozolomide (TMZ) or equivalent alkylating chemotherapy. The former, when added to surgery, improves survival by approximately 6 months [<xref ref-type="bibr" rid="B113">113</xref>]. More recent studies have demonstrated that the addition of temozolomide, both concurrently with and following radiation, increases survival by an additional two months [<xref ref-type="bibr" rid="B114">114</xref>]. Temozolomide, an FDA-approved chemotherapy drug sold under the brand name Temodar, is used to treat certain types of brain tumors, including GBM and anaplastic astrocytoma. It is an alkylating agent that, by damaging the DNA of cancer cells, slows or stops their growth. Recurrence is, however, inevitable and will almost always occur within the previously radiated area. GBM is highly heterogeneous with multiple cellular subtypes identified within a given tumor on transcriptomic sequencing. Studies have demonstrated that molecular subtypes have varying responsiveness to RT. Moreover, the impact of subtype analysis on radiation delivery is limited as it is usually obtained too late to impact RT planning due to the need to start radiation immediately after surgery. In addition, the high rate of recurrence of GBM suggests that a subset of radio-resistant cells is persistent and that there are likely infiltrative tumor regions not readily apparent on conventional MRI, which, therefore, receive sub-therapeutic doses of radiation.</p>
<p>Given the aforementioned problems&#x2013;the varying radio-sensitivity, the longitudinal tumor dynamics, and difficulties in identifying recurrent cell reservoirs&#x2013;there exists a critical gap in the clinical capabilities for providing adaptable, longitudinal care that adequately addresses each person&#x2019;s unique tumor burden. Current qualitative neuro-imaging approaches provide some distinguishing information that can predict prognostic molecular markers based on such features as texture and density [<xref ref-type="bibr" rid="B115">115</xref>], but conventional systems lack the ability to determine the extent of disease infiltration and the underlying molecular heterogeneities to adequately target them based on each patient&#x2019;s unique tumor burden. Due to such difficulties and the knowledge gap, it was hypothesized that one can bridge the gap by leveraging new MRI technologies to extract relevant imaging features, in addition to patient outcome data, and construct a predictive NN based on the data. Addressing these limitations, investigators have recently reported on leveraging conventional MRI-guided GBM tissue biopsies to contrast-enhancing and non-enhancing MRI regions in illustrating the potential for ML-based development of predictive radiogenomic modeling of underlying intratumor molecular heterogeneity [<xref ref-type="bibr" rid="B116">116</xref>].</p>
<p>A meningioma, according to Mayo Clinic [<xref ref-type="bibr" rid="B117">117</xref>], is &#x201c;a tumor that grows from the membranes that surround the brain and spinal cord, called the meninges. A meningioma is not a brain tumor, but it may press on the nearby brain, nerves, and vessels. Meningioma is the most common type of tumor that forms in the head. Meningiomas occur more often in women. Most meningiomas grow very slowly over many years. They can grow without causing symptoms. However, sometimes, their effects on nearby brain tissue, nerves, or vessels may cause serious disability.&#x201d; Most meningiomas are not cancerous, but some can be cancerous. More than 39,000 Americans are diagnosed with meningioma every year.</p>
<p>In humans, pituitary tumors, also called pituitary adenomas, are unusual growths that develop in the pituitary gland, or hypophysis, which is an endocrine gland (a network of glands and organs throughout the body) in vertebrates. It is behind the nose at the base of the brain. Some of these tumors cause the pituitary gland to produce excessive amounts of certain hormones that regulate important body functions, while others may cause the gland to produce insufficient amounts of the same hormones. Most pituitary tumors are benign, not cancerous, and typically do not spread to other parts of the body.</p>
<p>Deep NNs have been used extensively for analyzing brain images with multiple applications, including classification of disorders (such as schizophrenia and Alzheimer&#x2019;s disease); segmentation of tissues, lesions, and tumors; detection and classification of tumors and lesions; predicting survival rate against certain diseases; and image reconstruction. Most ML-based methods learn mappings from local patches to representations to labels. However, if the local patches do not contain the contextual information required for anatomical tasks, then a more refined method may be needed. For example, as mentioned above, [<xref ref-type="bibr" rid="B75">75</xref>] used patches obtained through non-uniformly sampling, which was carried out by gradually lowering the sampling rate in patch sides in order to span a larger context. Convolutional NNs have been used extensively in many analyses of brain images, typically obtained by brain MRI, for which there are many public MRI datasets on brain tumor classification, such as Kaggle, OpenNEURO, and fastMRI.</p>
<p>[<xref ref-type="bibr" rid="B118">118</xref>] used a three-layer CNN to classify the three types of brain tumors described above. The ReLU activation function was used in the first layer, followed by a structure in which the activation function could be adjusted to normalize the input layer. Dropout operations (i.e., randomly setting the outputs of hidden neurons to zero, usually carried out during the training of an NN) were used in both the second and third layers to mitigate the risk of overfitting. One CNN was used to distinguish between meningioma, glioma, and pituitary tumors with 3,064 enhanced images for 233 patients. The World Health Organization (WHO) classifies the risk levels associated with gliomas into four grades, with grade I being the least dangerous and grade IV being the most dangerous. [<xref ref-type="bibr" rid="B118">118</xref>] used a second CNN that differentiated between the three glioma grades of higher risk&#x2014;grades II, III, and IV&#x2014;using 516 images for 73 patients. The accuracies of the two distinct cases were, respectively, 96.13 and 98.7 percent. [<xref ref-type="bibr" rid="B119">119</xref>] undertook a similar study to classify glioma and meningioma tumors of the brain. They used a CNN for the classification problem and a second one, faster region-CNN (faster R-CNN), for segmentation. Faster R-CNN, proposed by [<xref ref-type="bibr" rid="B120">120</xref>], shares full-image convolutional features with the detection NN, which makes it possible to obtain region proposals by essentially cost-free computations. It consists of two modules, with one being a deep CNN that proposes regions and the second one being the faster R-CNN detector [<xref ref-type="bibr" rid="B53">53</xref>, <xref ref-type="bibr" rid="B55">55</xref>], which uses the proposed regions. The entire system is a single, unified NN for object detection. [<xref ref-type="bibr" rid="B120">120</xref>] used the faster R-CNN with 218 images as the training set, employing a CNN with two convolutional layers. Their model achieved an accuracy of 100 percent for meningioma classification and 87.5 percent for glioma classification, with an average confidence level of 94.6 percent in the segmentation of meningioma tumors.</p>
<p>CNNs were used by [<xref ref-type="bibr" rid="B121">121</xref>] (see also [<xref ref-type="bibr" rid="B122">122</xref>]) to classify tumor grades (see above), which were segmented using a pre-trained CascadeCNN, after which the deep features were extracted and the tumors were classified. CascadedCNNs were introduced by [<xref ref-type="bibr" rid="B123">123</xref>] for the segmentation of brain tumors. Their idea was that since the predictions should, in principle, be influenced by the model&#x2019;s beliefs about the value of nearby labels, one should feed the output probabilities of a first CNN as additional inputs to the layers of a second CNN by concatenating the convolutional layers. Thus, [<xref ref-type="bibr" rid="B123">123</xref>] concatenated the output layer of the first CNN with <italic>any</italic> of the layers in the second CNN, which, effectively, represented a cascade of two CNNs with a two-way processing of an image. [<xref ref-type="bibr" rid="B121">121</xref>] carried out the classification using a fine-tuned VGG-19 network (see above), a CNN with 16 convolutional layers and 3 fully connected layers. Two datasets, Radiopaedia and brain tumor [<xref ref-type="bibr" rid="B124">124</xref>], were employed, which were extended using eight different augmentation techniques. The datasets were divided into 50, 25, and 25 percent sets for, respectively, training, cross-validation, and testing. The accuracy for predicting the Radiopaedia dataset for grade I was, respectively, 90.03 and 95.5 percent without and with data augmentation, with similar accuracies for grades II, III, and IV. The corresponding numbers for the brain tumor dataset were 88.41 and 96.12 percent.</p>
<p>There have also been a few published studies for the specific case of glioblastoma. [<xref ref-type="bibr" rid="B125">125</xref>], for example, studied the problem of resection after surgery, considered to be one of the main prognostic factors for patients who have been diagnosed with glioblastoma. To achieve this, one must carry out segmentation and classification of residual tumors from postoperative MR images, and [<xref ref-type="bibr" rid="B125">125</xref>] used DCNNs to study the problem. The dataset used consisted of pre- and early post-operative MRI scans from 956 patients (in 12 hospitals in the United States and Europe) who had undergone surgical resection of glioblastoma. With very few exceptions, all the EPMR MRIs (early postoperative magnetic MRIs) were obtained within 72 h after surgery. To have anatomical consistency among the various input sequences, an initial image-to-image registration procedure was implemented. The brain masks were automatically generated using a pre-trained slab-wise AGU-Net model with input of <inline-formula id="inf136">
<mml:math id="m151">
<mml:mrow>
<mml:mn>256</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>192</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> voxels. AGU-Net (annotation-guided U-net) was introduced by [<xref ref-type="bibr" rid="B126">126</xref>] for rapid one-shot video object segmentation. It consists of two parts, a fully convolutional Siamese NN and a U-Net. In the first part, Siamese NNs are used to label the foreground and background in a new frame that corresponds to a reference object image, while the second part fuses the annotation information obtained by Siamese networks into the skip connection of a U-net to form an annotation-guided U-net for segmenting a specific target. A Siamese NN uses the same weights while working concurrently on two different input vectors to compute comparable output vectors. Since the AGU-Net analysis required a lower resolution, all inputs were resampled to an isotropic spacing and resized to <inline-formula id="inf137">
<mml:math id="m152">
<mml:mrow>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>144</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> voxels. Based on the analysis of the dataset by the nnU-Net framework (see above), the 3D full-resolution U-Net was selected for further analysis, using <inline-formula id="inf138">
<mml:math id="m153">
<mml:mrow>
<mml:mn>192</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>80</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> voxels as the size of the input patch, with the NN using five levels, downsampling using strided convolutional layers (a stride of one was used), and upsampling utilizing transposed convolutional layers. The size of the kernel was <inline-formula id="inf139">
<mml:math id="m154">
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> voxels in the first level and <inline-formula id="inf140">
<mml:math id="m155">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> for the remaining four, while those of the filter were <inline-formula id="inf141">
<mml:math id="m156">
<mml:mrow>
<mml:mo stretchy="false">{</mml:mo>
<mml:mrow>
<mml:mn>32,64,128,256,320</mml:mn>
</mml:mrow>
<mml:mo stretchy="false">}</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for the corresponding levels. The loss function was a combination of the Dice score and cross-entropy (see above). All models were initialized using pre-trained weights from the best pre-operative glioblastoma segmentation model. The input layer was the only layer that was adapted to account for the various input combinations. The Adam optimizer was used with an initial learning rate, <inline-formula id="inf142">
<mml:math id="m157">
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mi>l</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>r</mml:mi>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
<mml:msup>
<mml:mrow>
<mml:mn>0</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>. Residual tumor tissue was predicted during inference by each trained model, generating a probability map of the same resolution as the EPMR CE scan. The accuracy of the trained NNs was evaluated based on their ability to segment the residual tumor and classify patients into those with gross total resection and those with residual tumor. Only two classes were considered for segmentation, whereby a positive voxel exhibited tumor tissue, whereas a negative voxel indicated either background or normal tissue. A rest tumor volume threshold was selected for classification that served as the cut-off value. A Dice score of 61 percent was the best achieved segmentation performance, while the best classification performance reached 80 percent. More importantly, the segmentation performance of the most accurate models was comparable to that of human expert raters.</p>
<p>[<xref ref-type="bibr" rid="B127">127</xref>] developed an enhanced YOLO (so above) architecture for detecting brain tumors, which improved significantly the accuracy of detection using convolutional block attention modules (CBAMs), squeeze-and-excitation blocks (SE), and residual blocks. The CBAM increases the ability to focus on critical spatial and channel-wise features in MIs, while the SE blocks further improve feature representation by emphasizing important channels. RepVGG blocks [<xref ref-type="bibr" rid="B128">128</xref>] can efficiently extract features, while residual blocks help mitigate the vanishing gradient problem. The model was evaluated based on three brain tumor datasets and was demonstrated to provide significant improvements over previous models.</p>
<p>Other medical problems associated with the human brain have also been studied recently using ML approaches. [<xref ref-type="bibr" rid="B129">129</xref>] studied the problem of automatic identification of epilepsy in electroencephalography (EEG) signals. They pre-processed the data using adaptive wavelet denoising models in order to remove noise while preserving the actual information-carrying part of the signal. Then, multiple variable permutation entropy and multiple variable multi-scale fuzzy entropy (mvMFE) were extracted from the data that contained complexity and frequency-specific variations. To enhance the discriminative power of the features, uniform manifold approximation and projection were also utilized for non-linear dimensionality reduction. Next, the ResNet neural network (see above), integrated with bi-directional LSTM, was used to capture temporal dynamics and spatial features. When implemented in Python, the model achieved an accuracy of 94 percent, an F1-score of 96 percent, a recall value of 93 percent, a specificity of 87.70 percent, and a precision of 82.21 percent. [<xref ref-type="bibr" rid="B130">130</xref>] used deep CNNs, together with various transfer learning models (see above), to analyze resting state functional MRI (fMRI) images in order to predict the autism disorder features. Their work indicated that VGG16 NN achieves a classification accuracy of 95.8 percent. Their work supports our contention that ML models can be generalized across various neuroimaging-based diagnoses.</p>
</sec>
</sec>
<sec id="s7">
<label>7</label>
<title>Summary and conclusion</title>
<p>This study described and discussed recent advances in the applications of machine-learning algorithms, particularly neural networks, to addressing important problems in porous media and materials. The main issue described and discussed in this study is that two important classes of such materials that may seemingly be unrelated have many commonalities. One class consists of what are referred to as rock-like porous media (RLPM), such as soil, concrete, asphalt, and oil and gas reservoirs. A second group consists of BPMs and organs, such as skin, the brain, and lungs. Digital images of BPMs have been used for the diagnosis and treatment of illnesses for several decades, while their utilization in the modeling of various phenomena in RLPM is relatively new, giving rise to what is popularly referred to as &#x201c;digital rock.&#x201d; Due to the complexity of such images, along with the need to extract as much information from them as possible, the use of ML approaches, particularly NNs, has been increasing rapidly. We argue that, while the two classes of porous media and materials may appear completely different, they actually have many similarities and that using the images to solve various problems for both classes requires addressing similar issues. As a result, the application of ML algorithms to both types of porous materials is completely similar, even though the goals are seemingly very different. Thus, each scientific community can take advantage of advances in the other to gain a deeper understanding of the problems they study. In particular, we propose that, given that many biological materials and organs are porous materials, the medical community can take advantage of highly advanced approaches for the characterization and modeling of rock-like porous media that have been developed by those who work on such systems as the field of modeling of BPM is replete with <italic>ad hoc</italic> regression models and, in many cases, unrealistic models of BPMs.</p>
</sec>
</body>
<back>
<sec sec-type="data-availability" id="s8">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/supplementary material; further inquiries can be directed to the corresponding author.</p>
</sec>
<sec sec-type="author-contributions" id="s9">
<title>Author contributions</title>
<p>AS: Data curation, Writing &#x2013; original draft, Methodology. MS: Conceptualization, Project administration, Data curation, Writing &#x2013; original draft.</p>
</sec>
<sec sec-type="COI-statement" id="s11">
<title>Conflict of interest</title>
<p>The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="ai-statement" id="s12">
<title>Generative AI statement</title>
<p>The author(s) declared that generative AI was not used in the creation of this manuscript.</p>
<p>Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.</p>
</sec>
<sec sec-type="disclaimer" id="s13">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <source>Flow and transport in porous media and fractured rock</source>. <edition>2nd edn</edition>. <publisher-loc>Weinheim</publisher-loc>: <publisher-name>Wiley-VHC</publisher-name> (<year>2011</year>).</mixed-citation>
</ref>
<ref id="B2">
<label>2.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ghanbarian-Alavijeh</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Hunt</surname>
<given-names>AG</given-names>
</name>
<name>
<surname>Ewing</surname>
<given-names>RE</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Tortuosity in porous media: a critical review</article-title>. <source>Soil Sci Soc Am J</source> (<year>2013</year>) <volume>77</volume>:<fpage>1461</fpage>&#x2013;<lpage>77</lpage>. <pub-id pub-id-type="doi">10.2136/sssaj2012.0435</pub-id>
</mixed-citation>
</ref>
<ref id="B3">
<label>3.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Menon</surname>
<given-names>GK</given-names>
</name>
<name>
<surname>Cleary</surname>
<given-names>GM</given-names>
</name>
<name>
<surname>Lane</surname>
<given-names>KE</given-names>
</name>
</person-group>. <article-title>The structure and function of the stratum corneum</article-title>. <source>Int J Pharm</source> (<year>2012</year>) <volume>435</volume>:<fpage>3</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1016/j.ijpharm.2012.06.005</pub-id>
<pub-id pub-id-type="pmid">22705878</pub-id>
</mixed-citation>
</ref>
<ref id="B4">
<label>4.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Osseiran</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Dela Cruz</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Jeong</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Fthenakis</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Evans</surname>
<given-names>CL</given-names>
</name>
</person-group>. <article-title>Characterizing stratum corneum structure, barrier function, and chemical content of human skin with coherent raman scattering imaging</article-title>. <source>Biomed Opt Express</source> (<year>2018</year>) <volume>26</volume>:<fpage>6425</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1364/BOE.9.006425</pub-id>
<pub-id pub-id-type="pmid">31065440</pub-id>
</mixed-citation>
</ref>
<ref id="B5">
<label>5.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>Y</given-names>
</name>
</person-group>. <article-title>Three-dimensional model of normal human dermal tissue using serial tissue sections</article-title>. <source>Front Bioeng Biotechnol</source> (<year>2024</year>) <volume>12</volume>:<fpage>1347159</fpage>. <pub-id pub-id-type="doi">10.3389/fbioe.2024.1347159</pub-id>
<pub-id pub-id-type="pmid">38511132</pub-id>
</mixed-citation>
</ref>
<ref id="B6">
<label>6.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Miguel</surname>
<given-names>AF</given-names>
</name>
</person-group>. <article-title>Lungs as a natural porous media: architecture, airflow characteristics and transport of suspended particles</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Delgado</surname>
<given-names>J</given-names>
</name>
</person-group>, editor. <source>Heat and mass transforfer in porous media</source>. <publisher-loc>Berlin</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2012</year>). p. <fpage>13</fpage>.</mixed-citation>
</ref>
<ref id="B7">
<label>7.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>DeGroot</surname>
<given-names>CT</given-names>
</name>
<name>
<surname>Straatman</surname>
<given-names>AG</given-names>
</name>
</person-group>. <article-title>Towards a porous media model of the human lung</article-title>. <source>AIP Conf Proc</source> (<year>2012</year>) <volume>1453</volume>:<fpage>69</fpage>&#x2013;<lpage>74</lpage>. <pub-id pub-id-type="doi">10.1063/1.4711155</pub-id>
</mixed-citation>
</ref>
<ref id="B8">
<label>8.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Shan</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Zhan</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Pan</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Dou</surname>
<given-names>Y</given-names>
</name>
<etal/>
</person-group> <article-title>Evaluation of lung function by pore size distribution characteristics of lung tissue based on non-invasive imaging</article-title>. <source>J Rad Res Appl Sci</source> (<year>2023</year>) <volume>16</volume>:<fpage>100666</fpage>. <pub-id pub-id-type="doi">10.1016/j.jrras.2023.100666</pub-id>
</mixed-citation>
</ref>
<ref id="B9">
<label>9.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Helmberger</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Pienn</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Urschler</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Kullnig</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Stollberger</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Kovacs</surname>
<given-names>G</given-names>
</name>
<etal/>
</person-group> <article-title>Quantification of tortuosity and fractal dimension of the lung vessels in pulmonary hypertension patients</article-title>. <source>PLoS One</source> (<year>2014</year>) <volume>9</volume>:<fpage>e87515</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0087515</pub-id>
<pub-id pub-id-type="pmid">24498123</pub-id>
</mixed-citation>
</ref>
<ref id="B10">
<label>10.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nicholson</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Sheet and void porous media models for brain interstitial space</article-title>. <source>J Roy Soc Interf</source> (<year>2023</year>) <volume>20</volume>:<fpage>20230223</fpage>. <pub-id pub-id-type="doi">10.1098/rsif.2023.0223</pub-id>
<pub-id pub-id-type="pmid">37553990</pub-id>
</mixed-citation>
</ref>
<ref id="B11">
<label>11.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rusakov</surname>
<given-names>DA</given-names>
</name>
<name>
<surname>Kullmann</surname>
<given-names>DM</given-names>
</name>
</person-group>. <article-title>Geometric and viscous components of the tortuosity of the extracellular space in the brain</article-title>. <source>Proc Natl Acad Sci USA</source> (<year>1998</year>) <volume>95</volume>:<fpage>8975</fpage>&#x2013;<lpage>80</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.95.15.8975</pub-id>
<pub-id pub-id-type="pmid">9671789</pub-id>
</mixed-citation>
</ref>
<ref id="B12">
<label>12.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Richesson</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Flow and transport properties of deforming porous media. I. Permeability</article-title>. <source>Transp Porous Media</source> (<year>2021</year>) <volume>138</volume>:<fpage>577</fpage>&#x2013;<lpage>609</lpage>. <pub-id pub-id-type="doi">10.1007/s11242-021-01633-y</pub-id>
</mixed-citation>
</ref>
<ref id="B13">
<label>13.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Richesson</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Flow and transport properties of deforming porous media. II</article-title>. <source>Electr Conductivity Transp Porous Media</source> (<year>2021</year>) <volume>138</volume>:<fpage>611</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1007/s11242-021-01634-x</pub-id>
</mixed-citation>
</ref>
<ref id="B14">
<label>14.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nicholson</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Phillips</surname>
<given-names>JM</given-names>
</name>
</person-group>. <article-title>Ion diffusion modified by tortuosity and volume fraction in the extracellular microenvironment of the rat cerebellum</article-title>. <source>J Physiol</source> (<year>1981</year>) <volume>321</volume>:<fpage>225</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1113/jphysiol.1981.sp013981</pub-id>
<pub-id pub-id-type="pmid">7338810</pub-id>
</mixed-citation>
</ref>
<ref id="B15">
<label>15.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sykov&#xe1;</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Nicholson</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Diffusion in brain extracellular space</article-title>. <source>Physiol Rev</source> (<year>2008</year>) <volume>88</volume>:<fpage>1277</fpage>&#x2013;<lpage>340</lpage>. <pub-id pub-id-type="doi">10.1152/physrev.00027.2007</pub-id>
<pub-id pub-id-type="pmid">18923183</pub-id>
</mixed-citation>
</ref>
<ref id="B16">
<label>16.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deen</surname>
<given-names>WM</given-names>
</name>
</person-group>. <article-title>Hindered transport of large molecules in liquid-filled pores</article-title>. <source>Aiche J</source> (<year>1987</year>) <volume>33</volume>:<fpage>1409</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1002/aic.690330902</pub-id>
</mixed-citation>
</ref>
<ref id="B17">
<label>17.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Jue</surname>
<given-names>VL</given-names>
</name>
</person-group>. <article-title>Diffusion of large molecules in porous media</article-title>. <source>Phys Rev Lett</source> (<year>1989</year>) <volume>62</volume>:<fpage>629</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.62.629</pub-id>
<pub-id pub-id-type="pmid">10040288</pub-id>
</mixed-citation>
</ref>
<ref id="B18">
<label>18.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khademi</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Static and dynamic properties of supercooled water in small nanotubes</article-title>. <source>J Chem Phys</source> (<year>2016</year>) <volume>145</volume>:<fpage>024502</fpage>. <pub-id pub-id-type="doi">10.1063/1.4955313</pub-id>
<pub-id pub-id-type="pmid">27421415</pub-id>
</mixed-citation>
</ref>
<ref id="B19">
<label>19.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nicholson</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Hrab&#x115;tov&#xe1;</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Brain extracellular space: the final frontier of neuroscience</article-title>. <source>Biophys J</source> (<year>2017</year>) <volume>113</volume>:<fpage>2133</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1016/j.bpj.2017.06.052</pub-id>
<pub-id pub-id-type="pmid">28755756</pub-id>
</mixed-citation>
</ref>
<ref id="B20">
<label>20.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>KC</given-names>
</name>
<name>
<surname>Nicholson</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Changes in brain cell shape create residualextracellular space volume and explaintortuosity behavior duringosmotic challenge</article-title>. <source>Proc Natl Acad Sci USA</source> (<year>2000</year>) <volume>97</volume>:<fpage>8306</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.150338197</pub-id>
<pub-id pub-id-type="pmid">10890922</pub-id>
</mixed-citation>
</ref>
<ref id="B21">
<label>21.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hussain</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Mubeen</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Ullah</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Shah</surname>
<given-names>SS</given-names>
</name>
<name>
<surname>Abduljalil Khan</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Zahoor</surname>
<given-names>M</given-names>
</name>
<etal/>
</person-group> <article-title>Modern diagnostic imaging technique applications and risk factors in the medical field: a review</article-title>. <source>Biomed Res Int</source> (<year>2022</year>) <volume>6</volume>:<fpage>5164970</fpage>. <pub-id pub-id-type="doi">10.1155/2022/5164970</pub-id>
<pub-id pub-id-type="pmid">35707373</pub-id>
</mixed-citation>
</ref>
<ref id="B22">
<label>22.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Vinegar</surname>
<given-names>HJ</given-names>
</name>
<name>
<surname>Wellington</surname>
<given-names>SL</given-names>
</name>
</person-group>. <article-title>Tomographic imaging of three-phase flow experiments</article-title>. <source>Rev Sci Instr</source> (<year>1987</year>) <volume>58</volume>:<fpage>96</fpage>&#x2013;<lpage>107</lpage>. <pub-id pub-id-type="doi">10.1063/1.1139522</pub-id>
</mixed-citation>
</ref>
<ref id="B23">
<label>23.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wellington</surname>
<given-names>SI</given-names>
</name>
<name>
<surname>Vinegar</surname>
<given-names>HJ</given-names>
</name>
</person-group>. <article-title>X-ray computerized tomography</article-title>. <source>J Pet Technol</source> (<year>1987</year>) <volume>39</volume>:<fpage>885</fpage>&#x2013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.2118/16983-pa</pub-id>
</mixed-citation>
</ref>
<ref id="B24">
<label>24.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Charytanowicz</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Kowalski</surname>
<given-names>PA</given-names>
</name>
<name>
<surname>Lukasik</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Kulczycki</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Czachor</surname>
<given-names>H</given-names>
</name>
</person-group>. <article-title>Deep learning for porous media classification based on micro-CT images</article-title>. In: <source>International joint conference on neural networks (IJCNN)</source>. <publisher-loc>Padua, Italy</publisher-loc> (<year>2022</year>). p. <fpage>1</fpage>.</mixed-citation>
</ref>
<ref id="B25">
<label>25.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Jia</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Sermanet</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Reed</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Anguelov</surname>
<given-names>D</given-names>
</name>
<etal/>
</person-group> <article-title>Going deeper with convolutions</article-title>. <comment>arXiv:1409.4842</comment> (<year>2015</year>):<fpage>1</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/cvpr.2015.7298594</pub-id>
</mixed-citation>
</ref>
<ref id="B26">
<label>26.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>van der Maaten</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Weinberger</surname>
<given-names>KQ</given-names>
</name>
</person-group>. <article-title>Densely connected convolutional networks</article-title>. <comment>arXiv:1608.06993</comment> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B27">
<label>27.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Howard</surname>
<given-names>AG</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Kalenichenko</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Weyand</surname>
<given-names>T</given-names>
</name>
<etal/>
</person-group> <article-title>MobileNets: efficient convolutional neural networks for mobile vision applications</article-title>. <comment>arXiv:1704.04861</comment> (<year>2017</year>).</mixed-citation>
</ref>
<ref id="B28">
<label>28.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jayachandran</surname>
<given-names>ISAJ</given-names>
</name>
<name>
<surname>Gibbs</surname>
<given-names>HC</given-names>
</name>
<name>
<surname>Laya</surname>
<given-names>JC</given-names>
</name>
<name>
<surname>Qaiser</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Khan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Ansari</surname>
<given-names>MIMS</given-names>
</name>
<etal/>
</person-group> <article-title>An object-based approach to differentiate pores and microfractures in petrographic analysis using explainable, supervised machine learning</article-title>. <source>Earth Space Sci</source> (<year>2024</year>) <volume>11</volume>:<fpage>e2023EA003291</fpage>. <pub-id pub-id-type="doi">10.1029/2023ea003291</pub-id>
</mixed-citation>
</ref>
<ref id="B29">
<label>29.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rasaei</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Efficient simulation of water flooding in three-dimensional heterogeneous reservoirs using wavelet transformations: application to the SPE-10 model</article-title>. <source>Transp. Porous Media</source> (<year>2008</year>). <volume>72</volume>:<fpage>311</fpage>&#x2013;<lpage>338</lpage>. <pub-id pub-id-type="doi">10.1007/s11242-007-9152-1</pub-id>
</mixed-citation>
</ref>
<ref id="B30">
<label>30.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Aljasmi</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Speeding-up image-based simulation of two-phase flow in porous media with lattice-Boltzmann method using three-dimensional curvelet transforms</article-title>. <source>Phys Fluids</source> (<year>2021</year>) <volume>33</volume>:<fpage>113313</fpage>. <pub-id pub-id-type="doi">10.1063/5.0065857</pub-id>
</mixed-citation>
</ref>
<ref id="B31">
<label>31.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Buades</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Coll</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Morel</surname>
<given-names>JM</given-names>
</name>
</person-group>. <article-title>Non-local means denoising</article-title>. <source>Image Process On Line</source> (<year>2011</year>) <volume>1</volume>:<fpage>208</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.5201/ipol.2011.bcm_nlm</pub-id>
</mixed-citation>
</ref>
<ref id="B32">
<label>32.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <source>Artificial intelligence in science and engineering</source>. <publisher-loc>Weinheim</publisher-loc>: <publisher-name>Wiley VCH</publisher-name> (<year>2026</year>).</mixed-citation>
</ref>
<ref id="B33">
<label>33.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Berg</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Saxena</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Shaik</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Pradhan</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Generation of ground truth images to validate micro-CT image-processing pipelines</article-title>. <source>The Leading Edge</source> (<year>2018</year>) <volume>37</volume>:<fpage>412</fpage>&#x2013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1190/tle37060412.1</pub-id>
</mixed-citation>
</ref>
<ref id="B34">
<label>34.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arganda-Carreras</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Kaynig</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Rueden</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Eliceiri</surname>
<given-names>KW</given-names>
</name>
<name>
<surname>Schindelin</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Cardona</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Trainable weka segmentation: a machine learning tool for microscopy pixel classification</article-title>. <source>Bioinformatics</source> (<year>2017</year>) <volume>33</volume>:<fpage>2424</fpage>&#x2013;<lpage>6</lpage>. <pub-id pub-id-type="doi">10.1093/bioinformatics/btx180</pub-id>
<pub-id pub-id-type="pmid">28369169</pub-id>
</mixed-citation>
</ref>
<ref id="B35">
<label>35.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Marcelis</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Berg</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Brown</surname>
<given-names>B</given-names>
</name>
</person-group>. In: <person-group person-group-type="editor">
<name>
<surname>M&#xe9;dici</surname>
<given-names>EF</given-names>
</name>
<name>
<surname>Otero</surname>
<given-names>AD</given-names>
</name>
</person-group>, editors. <source>Album of porous media</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2023</year>).</mixed-citation>
</ref>
<ref id="B36">
<label>36.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Marcelis</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Steffen</surname>
<given-names>B</given-names>
</name>
</person-group>. <article-title>North Netherlands gas reservoir</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>M&#xe9;dici</surname>
<given-names>EF</given-names>
</name>
<name>
<surname>Otero</surname>
<given-names>AD</given-names>
</name>
</person-group>, editors. <source>Album of porous media</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2023</year>).</mixed-citation>
</ref>
<ref id="B37">
<label>37.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Karimpouli</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>Segmentation of digital rock images using deep convolutional autoencoder networks</article-title>. <source>Comput Geosci</source> (<year>2019</year>) <volume>126</volume>:<fpage>142</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1016/j.cageo.2019.02.003</pub-id>
</mixed-citation>
</ref>
<ref id="B38">
<label>38.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Badrinarayanan</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Kendall</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Cipolla</surname>
<given-names>R</given-names>
</name>
</person-group>. <article-title>SegNet: a deep convolutional encoderdecoder architecture for image segmentation</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source> (<year>2017</year>) <volume>39</volume>:<fpage>2481</fpage>&#x2013;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2016.2644615</pub-id>
<pub-id pub-id-type="pmid">28060704</pub-id>
</mixed-citation>
</ref>
<ref id="B39">
<label>39.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Reconstruction of three-dimensional porous media using a single thin section</article-title>. <source>Phys Rev E</source> (<year>2012</year>) <volume>85</volume>:<fpage>066709</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevE.85.066709</pub-id>
<pub-id pub-id-type="pmid">23005245</pub-id>
</mixed-citation>
</ref>
<ref id="B40">
<label>40.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Cross-correlation function for accurate reconstruction of heterogeneous media</article-title>. <source>Phys Rev Lett</source> (<year>2013</year>) <volume>110</volume>:<fpage>078002</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevLett.110.078002</pub-id>
<pub-id pub-id-type="pmid">25166410</pub-id>
</mixed-citation>
</ref>
<ref id="B41">
<label>41.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Enhancing multiple-point geostatistical modeling. 1: graph theory and pattern adjustment</article-title>. <source>Water Resour Res</source> (<year>2016</year>) <volume>52</volume>:<fpage>2074</fpage>&#x2013;<lpage>98</lpage>. <pub-id pub-id-type="doi">10.1002/2015wr017806</pub-id>
</mixed-citation>
</ref>
<ref id="B42">
<label>42.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Enhancing multiple-point geostatistical modeling. 2: iterative simulation and multiple distance functions</article-title>. <source>Water Resour Res</source> (<year>2016</year>) <volume>52</volume>:<fpage>2099</fpage>&#x2013;<lpage>122</lpage>. <pub-id pub-id-type="doi">10.1002/2015wr017807</pub-id>
</mixed-citation>
</ref>
<ref id="B43">
<label>43.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kim</surname>
<given-names>HE</given-names>
</name>
<name>
<surname>Cosa-Linan</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Santhanam</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Jannesari</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Maros</surname>
<given-names>ME</given-names>
</name>
<name>
<surname>Ganslandt</surname>
<given-names>T</given-names>
</name>
</person-group>. <article-title>Transfer learning for medical image classification: a literature review</article-title>. <source>BMC Med Imaging</source> (<year>2022</year>) <volume>22</volume>:<fpage>69</fpage>. <pub-id pub-id-type="doi">10.1186/s12880-022-00793-7</pub-id>
<pub-id pub-id-type="pmid">35418051</pub-id>
</mixed-citation>
</ref>
<ref id="B44">
<label>44.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gulshan</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Peng</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Coram</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Stumpe</surname>
<given-names>MC</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Narayanaswamy</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs</article-title>. <source>J Am Med Assoc</source> (<year>2016</year>) <volume>316</volume>:<fpage>2402</fpage>&#x2013;<lpage>10</lpage>. <pub-id pub-id-type="doi">10.1001/jama.2016.17216</pub-id>
<pub-id pub-id-type="pmid">27898976</pub-id>
</mixed-citation>
</ref>
<ref id="B45">
<label>45.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Esteva</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Kuprel</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Novoa</surname>
<given-names>RA</given-names>
</name>
<name>
<surname>Ko</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Swetter</surname>
<given-names>SM</given-names>
</name>
<name>
<surname>Blau</surname>
<given-names>HM</given-names>
</name>
<etal/>
</person-group> <article-title>Dermatologist-level classification of skin cancer with deep neural networks</article-title>. <source>Nature</source> (<year>2017</year>) <volume>542</volume>:<fpage>115</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1038/nature21056</pub-id>
<pub-id pub-id-type="pmid">28117445</pub-id>
</mixed-citation>
</ref>
<ref id="B46">
<label>46.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Raghu</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Kleinberg</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Transfusion: understanding transfer learning for medical imaging</article-title>. <comment>arXiv:1902.07208</comment> (<year>2019</year>).</mixed-citation>
</ref>
<ref id="B47">
<label>47.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Menegola</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Fornaciali</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Pires</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Avila</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Valle</surname>
<given-names>E</given-names>
</name>
</person-group>. <article-title>Towards automated melanoma screening: exploring transfer learning schemes</article-title>. <comment>arxiv:1609.01228</comment> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B48">
<label>48.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Setio</surname>
<given-names>AAA</given-names>
</name>
<name>
<surname>Ciompi</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Litjens</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Gerke</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Jacobs</surname>
<given-names>C</given-names>
</name>
<name>
<surname>van Riel</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Pulmonary nodule detection in CT images: false positive reduction using multi-view convolutional networks</article-title>. <source>IEEE Trans Med Imaging</source> (<year>2016</year>) <volume>35</volume>:<fpage>1160</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2016.2536809</pub-id>
<pub-id pub-id-type="pmid">26955024</pub-id>
</mixed-citation>
</ref>
<ref id="B49">
<label>49.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>van Grinsven</surname>
<given-names>MJJP</given-names>
</name>
<name>
<surname>van Ginneken</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Hoyng</surname>
<given-names>CB</given-names>
</name>
<name>
<surname>Theelen</surname>
<given-names>T</given-names>
</name>
<name>
<surname>S&#x2019;anchez</surname>
<given-names>CI</given-names>
</name>
</person-group>. <article-title>Fast convolutional neural network training using selective data sampling: application to hemorrhage detection in color fundus images</article-title>. <source>IEEE Trans Med Imaging</source> (<year>2016</year>) <volume>35</volume>:<fpage>1273</fpage>&#x2013;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2016.2526689</pub-id>
<pub-id pub-id-type="pmid">26886969</pub-id>
</mixed-citation>
</ref>
<ref id="B50">
<label>50.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wolterink</surname>
<given-names>JM</given-names>
</name>
<name>
<surname>Leiner</surname>
<given-names>T</given-names>
</name>
<name>
<surname>de Vos</surname>
<given-names>BD</given-names>
</name>
<name>
<surname>van Hamersvelt</surname>
<given-names>RW</given-names>
</name>
<name>
<surname>Viergever</surname>
<given-names>MA</given-names>
</name>
<name>
<surname>Isgum</surname>
<given-names>I</given-names>
</name>
</person-group>. <article-title>Automatic coronary artery calcium scoring in cardiac CT angiography using paired convolutional neural networks</article-title>. <source>Med Image Anal</source> (<year>2016</year>) <volume>34</volume>:<fpage>123</fpage>&#x2013;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2016.04.004</pub-id>
<pub-id pub-id-type="pmid">27138584</pub-id>
</mixed-citation>
</ref>
<ref id="B51">
<label>51.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Yang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Metaxas</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Automated anatomical landmark detection on distal femur surface using convolutional neural network</article-title>. In: <source>Proceedings of the IEEE international symposium on biomedical imaging (ISBI)</source>. <publisher-loc>Brooklyn, NY</publisher-loc> (<year>2015</year>). p. <fpage>17</fpage>.</mixed-citation>
</ref>
<ref id="B52">
<label>52.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>de Vos</surname>
<given-names>BD</given-names>
</name>
<name>
<surname>Wolterink</surname>
<given-names>JM</given-names>
</name>
<name>
<surname>de Jong</surname>
<given-names>PA</given-names>
</name>
<name>
<surname>Viergever</surname>
<given-names>MA</given-names>
</name>
<name>
<surname>I&#x161;gum</surname>
<given-names>I</given-names>
</name>
</person-group>. <article-title>2D image classification for 3D anatomy localization: employing deep convolutional neural networks</article-title>. In: <source>Proceedings of the SPIE on medical imaging, 9784: image processing</source> (<year>2016</year>) <fpage>97841Y</fpage>.</mixed-citation>
</ref>
<ref id="B53">
<label>53.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kumar</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Sridar</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Quinton</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>RK</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Nanan</surname>
<given-names>R</given-names>
</name>
<etal/>
</person-group> <article-title>Plane identification in fetal ultrasound images using saliency maps and convolutional neural networks</article-title>. In: <source>2016 IEEE 13th international symposium on biomedical imaging (ISBI)</source>. <publisher-loc>Prague, Czech Republic</publisher-loc> (<year>2016</year>). p. <fpage>791</fpage>.</mixed-citation>
</ref>
<ref id="B54">
<label>54.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Payer</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Stern</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Bischof</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Urschler</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Regressing heatmaps for multiple landmark localization using CNNs</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>9901</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>230</fpage>&#x2013;<lpage>8</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46723-8_27</pub-id>
</mixed-citation>
</ref>
<ref id="B55">
<label>55.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ghesu</surname>
<given-names>FC</given-names>
</name>
<name>
<surname>Georgescu</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Mansi</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Neumann</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Hornegger</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Comaniciu</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>An artificial agent for anatomical landmark detection in medical images</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>9902</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>229</fpage>&#x2013;<lpage>37</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46726-9_27</pub-id>
</mixed-citation>
</ref>
<ref id="B56">
<label>56.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Zheng</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Georgescu</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Nguyen</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Comaniciu</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>3D deep learning for efficient and robust landmark detection in volumetric data</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Navab</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Hornegger</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Frangi</surname>
<given-names>A</given-names>
</name>
</person-group>, editors. <source>Proceedings of the medical image computing and computer-assisted intervention</source>, <volume>9349</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2015</year>). p. <fpage>565</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24553-9_69</pub-id>
</mixed-citation>
</ref>
<ref id="B57">
<label>57.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Baumgartner</surname>
<given-names>CF</given-names>
</name>
<name>
<surname>Kamnitsas</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Matthew</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Smith</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Kainz</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Rueckert</surname>
<given-names>D</given-names>
</name>
</person-group>. <article-title>Real time standard scan plane detection and localisation in fetal ultrasound using fully convolutional neural networks</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>9901</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>203</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46723-8_24</pub-id>
</mixed-citation>
</ref>
<ref id="B58">
<label>58.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>K&#xfc;stner</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Fuin</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Hammernik</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Bustin</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Hajhosseiny</surname>
<given-names>R</given-names>
</name>
<etal/>
</person-group> <article-title>CINENet: deep learning-based 3D cardiac CINE MRI reconstruction with multi-coil complex-valued 4D spatio-temporal convolutions</article-title>. <source>Sci Rep</source> (<year>2020</year>) <volume>10</volume>:<fpage>article 13710</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-020-70551-8</pub-id>
<pub-id pub-id-type="pmid">32792507</pub-id>
</mixed-citation>
</ref>
<ref id="B59">
<label>59.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Dou</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Ni</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Cheng</surname>
<given-names>JZ</given-names>
</name>
<name>
<surname>Qin</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Automatic fetal ultrasound standard plane detection using knowledge transferred recurrent neural networks</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Navab</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Hornegger</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Frangi</surname>
<given-names>A</given-names>
</name>
</person-group>, editors. <source>Proceedings of the medical image computing and computer-assisted intervention</source>, <volume>9349</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2015</year>). p. <fpage>507</fpage>&#x2013;<lpage>14</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-24553-9_62</pub-id>
</mixed-citation>
</ref>
<ref id="B60">
<label>60.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kong</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Zhan</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Shin</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Denny</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S</given-names>
</name>
</person-group>. <article-title>Recognizing end-diastole and end-systole frames <italic>via</italic> deep temporal regression network</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>9901</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>264</fpage>&#x2013;<lpage>72</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46726-9_31</pub-id>
</mixed-citation>
</ref>
<ref id="B61">
<label>61.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Redmon</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Divvala</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Farhadi</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>You only look once: unified, real-time object detection</article-title>. <comment>arXiv:1506.02640</comment> (<year>2015</year>).</mixed-citation>
</ref>
<ref id="B62">
<label>62.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Redmon</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Farhadi</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>YOLO9000: better, faster, stronger</article-title>. <comment>arXiv:1612.08242</comment> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B63">
<label>63.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Redmon</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Farhadi</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>YOLOv3: an incremental improvement</article-title>. <comment>arXiv:1804.02767</comment> (<year>2018</year>).</mixed-citation>
</ref>
<ref id="B64">
<label>64.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bochkovskiy</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>CY</given-names>
</name>
<name>
<surname>Liao</surname>
<given-names>HYM</given-names>
</name>
</person-group>. <article-title>YOLOv4: optimal speed and accuracy of object detection</article-title>. <comment>arXiv:2004.10934</comment> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="B65">
<label>65.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Wei</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Lu</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Turkbey</surname>
<given-names>E</given-names>
</name>
<etal/>
</person-group> <article-title>Colitis detection on computed tomography using regional convolutional neural networks</article-title>. <source>Med Phys</source> (<year>2017</year>) <volume>44</volume>:<fpage>4630</fpage>&#x2013;<lpage>42</lpage>. <pub-id pub-id-type="doi">10.1002/mp.12399</pub-id>
<pub-id pub-id-type="pmid">28594460</pub-id>
</mixed-citation>
</ref>
<ref id="B66">
<label>66.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Girshick</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Donahue</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Darrell</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Malik</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Rich feature hierarchies for accurate object detection and semantic segmentation</article-title>. <comment>arXiv:1311.2524</comment> (<year>2013</year>).</mixed-citation>
</ref>
<ref id="B67">
<label>67.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>Very deep convolutional networks for large-scale image recognition</article-title>. <comment>arXiv:1409.1556</comment> (<year>2015</year>).</mixed-citation>
</ref>
<ref id="B68">
<label>68.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Girshick</surname>
<given-names>R</given-names>
</name>
</person-group> <article-title>Fast R-CNN</article-title>, <comment>arXiv:1504.08083</comment> (<year>2014</year>).</mixed-citation>
</ref>
<ref id="B69">
<label>69.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J</given-names>
</name>
</person-group>. <source>Spatial pyramid pooling in deep convolutional networks for visual recognition</source> (<year>2014</year>). <comment>arXiv:1406.4729</comment>.</mixed-citation>
</ref>
<ref id="B70">
<label>70.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Cicek</surname>
<given-names>&#xd6;</given-names>
</name>
<name>
<surname>Abdulkadir</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Lienkamp</surname>
<given-names>SS</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Ronneberger</surname>
<given-names>O</given-names>
</name>
</person-group>. <article-title>3D U-Net: learning dense volumetric segmentation from sparse annotation</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>9901</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>424</fpage>&#x2013;<lpage>32</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46723-8_49</pub-id>
</mixed-citation>
</ref>
<ref id="B71">
<label>71.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Drozdzal</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Vorontsov</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Chartrand</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Kadoury</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Pal</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>The importance of skip connections in biomedical image segmentation</article-title>. <source>arXiv:1608</source> (<year>2016</year>):<fpage>04117</fpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1608.04117</pub-id>
</mixed-citation>
</ref>
<ref id="B72">
<label>72.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Isensee</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Jaeger</surname>
<given-names>PF</given-names>
</name>
<name>
<surname>Kohl</surname>
<given-names>SAA</given-names>
</name>
<name>
<surname>Petersen</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Maier-Hein</surname>
<given-names>KM</given-names>
</name>
</person-group>. <article-title>nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation</article-title>. <source>Nat Methods</source> (<year>2021</year>) <volume>18</volume>:<fpage>203</fpage>&#x2013;<lpage>11</lpage>. <pub-id pub-id-type="doi">10.1038/s41592-020-01008-z</pub-id>
<pub-id pub-id-type="pmid">33288961</pub-id>
</mixed-citation>
</ref>
<ref id="B73">
<label>73.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ulyanov</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Vedaldi</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Lempitsky</surname>
<given-names>V</given-names>
</name>
</person-group>. <article-title>Instance normalization: the missing ingredient for fast stylization</article-title>. <comment>arXiv:1607.08022</comment> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B74">
<label>74.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamnitsas</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Ledig</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Newcombe</surname>
<given-names>VF</given-names>
</name>
<name>
<surname>Simpson</surname>
<given-names>JP</given-names>
</name>
<name>
<surname>Kane</surname>
<given-names>AD</given-names>
</name>
<name>
<surname>Menon</surname>
<given-names>DK</given-names>
</name>
<etal/>
</person-group> <article-title>Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation</article-title>. <source>Med Image Anal</source> (<year>2017</year>) <volume>36</volume>:<fpage>61</fpage>&#x2013;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2016.10.004</pub-id>
<pub-id pub-id-type="pmid">27865153</pub-id>
</mixed-citation>
</ref>
<ref id="B75">
<label>75.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ghafoorian</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Karssemeijer</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Heskes</surname>
<given-names>T</given-names>
</name>
<name>
<surname>van Uden</surname>
<given-names>IWM</given-names>
</name>
<name>
<surname>de Leeuw</surname>
<given-names>FE</given-names>
</name>
<name>
<surname>Marchiori</surname>
<given-names>E</given-names>
</name>
<etal/>
</person-group> <article-title>Non-uniform patch sampling with deep convolutional neural networks for white matter hyperintensity segmentation</article-title>. In: <source>2016 IEEE 13th international symposium on biomedical imaging (ISBI)</source>. <publisher-loc>Prague, Czech Republic</publisher-loc> (<year>2016</year>). p. <fpage>1414</fpage>.</mixed-citation>
</ref>
<ref id="B76">
<label>76.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Park</surname>
<given-names>SC</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>MK</given-names>
</name>
<name>
<surname>Kang</surname>
<given-names>MG</given-names>
</name>
</person-group>. <article-title>Super-resolution image reconstruction: a technical overview</article-title>. <source>IEEE Signal Process. Mag</source> (<year>2003</year>) <volume>20</volume>:<fpage>21</fpage>. <pub-id pub-id-type="doi">10.1109/MSP.2003.1203207</pub-id>
</mixed-citation>
</ref>
<ref id="B77">
<label>77.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Glasner</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Bagon</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Irani</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Super-resolution from a single image</article-title>. In: <source>Proceedings of IEEE 12th international conference on computer vision</source>. <publisher-loc>Kyoto, Japan</publisher-loc> (<year>2009</year>). p. <fpage>349</fpage>.</mixed-citation>
</ref>
<ref id="B78">
<label>78.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>Reconstruction, optimization, and design of heterogeneous materials and media: basic principles, computational algorithms, and applications</article-title>. <source>Phys Rep</source> (<year>2021</year>) <volume>939</volume>:<fpage>1</fpage>&#x2013;<lpage>82</lpage>. <pub-id pub-id-type="doi">10.1016/j.physrep.2021.09.003</pub-id>
</mixed-citation>
</ref>
<ref id="B79">
<label>79.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kim</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>JK</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>KM</given-names>
</name>
</person-group>. <article-title>Deeply-recursive convolutional network for image superresolution</article-title>. <source>arXiv:1511</source> (<year>2015</year>):<fpage>04491</fpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1511.04491</pub-id>
</mixed-citation>
</ref>
<ref id="B80">
<label>80.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mao</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Shen</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>YB</given-names>
</name>
</person-group>. <article-title>Image restoration using very deep convolutional encoderdecoder networks with symmetric skip connections</article-title>. <source>Adv Neural Infor Process. Syst</source> (<year>2016</year>) <volume>29</volume>:<fpage>2802</fpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1603.09056</pub-id>
</mixed-citation>
</ref>
<ref id="B81">
<label>81.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Loy</surname>
<given-names>CC</given-names>
</name>
<name>
<surname>He</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Image super-resolution using deep convolutional networks</article-title>. <source>IEEE Trans Pattern Anal Mach Intell</source> (<year>2014</year>) <volume>38</volume>:<fpage>295</fpage>&#x2013;<lpage>307</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2015.2439281</pub-id>
<pub-id pub-id-type="pmid">26761735</pub-id>
</mixed-citation>
</ref>
<ref id="B82">
<label>82.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Dong</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Loy</surname>
<given-names>CC</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X</given-names>
</name>
</person-group>. <article-title>Accelerating the super-resolution convolutional neural network</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Leibe</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Matas</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Sebe</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M</given-names>
</name>
</person-group>, editors. <source>Computer vision - ECCV 2016</source>, <volume>9906</volume>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>367</fpage>.</mixed-citation>
</ref>
<ref id="B83">
<label>83.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Odena</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Dumoulin</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Olah</surname>
<given-names>C</given-names>
</name>
</person-group>. <article-title>Deconvolution and checkerboard artifacts</article-title>. <source>Distill</source> (<year>2015</year>) <volume>10</volume>:<fpage>23915</fpage>.</mixed-citation>
</ref>
<ref id="B84">
<label>84.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shi</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Caballero</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Huszar</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Totz</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Aitken</surname>
<given-names>AP</given-names>
</name>
<name>
<surname>Bishop</surname>
<given-names>R</given-names>
</name>
<etal/>
</person-group> <article-title>Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network</article-title>. <source>arXiv:1609</source> (<year>2016</year>):<fpage>05158</fpage>. <pub-id pub-id-type="doi">10.48550/arXiv.1609.05158</pub-id>
</mixed-citation>
</ref>
<ref id="B85">
<label>85.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>YD</given-names>
</name>
<name>
<surname>Armstrong</surname>
<given-names>RT</given-names>
</name>
<name>
<surname>Mostaghimi</surname>
<given-names>P</given-names>
</name>
</person-group>. <article-title>Boosting resolution and recovering texture of 2D and 3D micro-CT images with deep learning</article-title>. <source>Water Resour Res</source> (<year>2020</year>) <volume>56</volume>:<fpage>e2019WR026052</fpage>. <pub-id pub-id-type="doi">10.1029/2019wr026052</pub-id>
</mixed-citation>
</ref>
<ref id="B86">
<label>86.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>YD</given-names>
</name>
<name>
<surname>Mostaghimi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Armstrong</surname>
<given-names>RT</given-names>
</name>
</person-group>. <article-title>A super resolution dataset of digital rocks (DRSRD1): sandstone and carbonate</article-title>. <source>arXiv</source> (<year>2019</year>). <pub-id pub-id-type="doi">10.17612/P7D38H</pub-id>
</mixed-citation>
</ref>
<ref id="B87">
<label>87.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Lim</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Son</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Nah</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>KM</given-names>
</name>
</person-group>. <article-title>Enhanced deep residual networks for single image super-resolution</article-title>. In: <source>Proceedings of 2017 IEEE conference on computer vision and pattern recognition workshops (CVPRW), Honolulu, USA</source> (<year>2017</year>). p. <fpage>1132</fpage>.</mixed-citation>
</ref>
<ref id="B88">
<label>88.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Ledig</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Theis</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Huszar</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Caballero</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Cunningham</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Acosta</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Photo-realistic single image super-resolution using a generative adversarial network</article-title>. In: <source>Proceedings of 2017 IEEE conference on computer vision and pattern recognition (CVPR), Honolulu, HI, USA</source> (<year>2017</year>). p. <fpage>105</fpage>.</mixed-citation>
</ref>
<ref id="B89">
<label>89.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Ren</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Delving deep into rectifiers: surpassing human-level performance on ImageNet classification</article-title>. <comment>arXiv:1502.01852</comment> (<year>2015</year>):<fpage>1026</fpage>&#x2013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2015.123</pub-id>
</mixed-citation>
</ref>
<ref id="B90">
<label>90.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jackson</surname>
<given-names>SJ</given-names>
</name>
<name>
<surname>Niu</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Manoorkar</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Mostaghimi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Armstrong</surname>
<given-names>RT</given-names>
</name>
</person-group>. <article-title>Deep learning of multiresolution X-ray micro-computed-tomography images for multiscale modeling</article-title>. <source>Phys Rev Appl</source> (<year>2022</year>) <volume>17</volume>:<fpage>054046</fpage>. <pub-id pub-id-type="doi">10.1103/physrevapplied.17.054046</pub-id>
</mixed-citation>
</ref>
<ref id="B91">
<label>91.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Oktay</surname>
<given-names>O</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Guerrero</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Kamnitsas</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Caballero</surname>
<given-names>J</given-names>
</name>
<etal/>
</person-group> <article-title>Multi-input cardiac image super-resolution using convolutional neural networks</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ourselin</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Joskowicz</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sabuncu</surname>
<given-names>MR</given-names>
</name>
<name>
<surname>Unal</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer-assisted intervention - MICCAI</source>, <volume>Vol. 9902</volume>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2016</year>). p. <fpage>246</fpage>&#x2013;<lpage>54</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-46726-9_29</pub-id>
</mixed-citation>
</ref>
<ref id="B92">
<label>92.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamrava</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Enhancing images of shale formations by a hybrid stochastic and deep learning algorithm</article-title>. <source>Neural Netw</source> (<year>2019</year>) <volume>118</volume>:<fpage>310</fpage>&#x2013;<lpage>20</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2019.07.009</pub-id>
<pub-id pub-id-type="pmid">31326663</pub-id>
</mixed-citation>
</ref>
<ref id="B93">
<label>93.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Golkov</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Dosovitskiy</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Sperl</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Menzel</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Czisch</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Samann</surname>
<given-names>P</given-names>
</name>
<etal/>
</person-group> <article-title>Q-space deep learning: twelve-fold shorter and model-free diffusion MRI scans</article-title>. <source>IEEE Trans Med Imaging</source> (<year>2016</year>) <volume>35</volume>:<fpage>1344</fpage>&#x2013;<lpage>51</lpage>. <pub-id pub-id-type="doi">10.1109/TMI.2016.2551324</pub-id>
<pub-id pub-id-type="pmid">27071165</pub-id>
</mixed-citation>
</ref>
<ref id="B94">
<label>94.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>H</given-names>
</name>
</person-group>. <article-title>Medical image analysis using deep learning algorithms</article-title>. <source>Front Public Health</source> (<year>2023</year>) <volume>11</volume>:<fpage>1273253</fpage>. <pub-id pub-id-type="doi">10.3389/fpubh.2023.1273253</pub-id>
<pub-id pub-id-type="pmid">38026291</pub-id>
</mixed-citation>
</ref>
<ref id="B95">
<label>95.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mall</surname>
<given-names>PK</given-names>
</name>
<name>
<surname>Singh</surname>
<given-names>PK</given-names>
</name>
<name>
<surname>Srivastav</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Narayan</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Paprzycki</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Jaworska</surname>
<given-names>T</given-names>
</name>
<etal/>
</person-group> <article-title>A comprehensive review of deep neural networks for medical image processing: recent developments and future opportunities</article-title>. <source>Health Analytic</source> (<year>2023</year>) <volume>4</volume>:<fpage>100216</fpage>. <pub-id pub-id-type="doi">10.1016/j.health.2023.100216</pub-id>
</mixed-citation>
</ref>
<ref id="B96">
<label>96.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Courville</surname>
<given-names>A</given-names>
</name>
</person-group>. <source>Deep learning</source>. <publisher-loc>Cambridge</publisher-loc>: <publisher-name>MIT Press</publisher-name> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B97">
<label>97.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Mirza</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ozair</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Generative adversarial networks</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ghahramani</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Welling</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Cortes</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Lawrence</surname>
<given-names>ND</given-names>
</name>
<name>
<surname>Weinberger</surname>
<given-names>KQ</given-names>
</name>
</person-group>, editors. <source>Advances in neural information processing systems</source>, <volume>27</volume>. <publisher-loc>New York</publisher-loc>: <publisher-name>Curran Association</publisher-name> (<year>2014</year>). p. <fpage>2672</fpage>.</mixed-citation>
</ref>
<ref id="B98">
<label>98.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mosser</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Dubrule</surname>
<given-names>O</given-names>
</name>
<name>
<surname>Blunt</surname>
<given-names>MJ</given-names>
</name>
</person-group>. <article-title>Reconstruction of three-dimensional porous media using generative adversarial neural networks</article-title>. <source>Phys Rev E</source> (<year>2017</year>) <volume>96</volume>:<fpage>043309</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevE.96.043309</pub-id>
<pub-id pub-id-type="pmid">29347591</pub-id>
</mixed-citation>
</ref>
<ref id="B99">
<label>99.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mosser</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Dubrule</surname>
<given-names>O</given-names>
</name>
<name>
<surname>Blunt</surname>
<given-names>MJ</given-names>
</name>
</person-group>. <article-title>Stochastic seismic waveform inversion using generative adversarial networks as a geological prior</article-title>. <source>Math Geosci</source> (<year>2020</year>) <volume>52</volume>:<fpage>53</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1007/s11004-019-09832-6</pub-id>
</mixed-citation>
</ref>
<ref id="B100">
<label>100.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Iassonov</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Gebrenegus</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Tuller</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Segmentation of X-ray computed tomography images of porous materials: a crucial step for characterization and quantitative analysis of pore structures</article-title>. <source>Water Resour Res</source> (<year>2009</year>) <volume>45</volume>:<fpage>W09415</fpage>. <pub-id pub-id-type="doi">10.1029/2009wr008087</pub-id>
</mixed-citation>
</ref>
<ref id="B101">
<label>101.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sakib</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tazrin</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Fouda</surname>
<given-names>MM</given-names>
</name>
<name>
<surname>Fadlullah</surname>
<given-names>ZM</given-names>
</name>
<name>
<surname>Guizani</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>DL-CRC: deep learningbased chest radiograph classification for COVID-19 detection: a novel approach</article-title>. <source>IEEE Access</source> (<year>2020</year>) <volume>8</volume>:<fpage>171575</fpage>&#x2013;<lpage>89</lpage>. <pub-id pub-id-type="doi">10.1109/ACCESS.2020.3025010</pub-id>
<pub-id pub-id-type="pmid">34976555</pub-id>
</mixed-citation>
</ref>
<ref id="B102">
<label>102.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamrava</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Linking morphology of porous media to their macroscopic permeability by deep learning</article-title>. <source>Transp Porous Media</source> (<year>2020</year>) <volume>131</volume>:<fpage>427</fpage>&#x2013;<lpage>48</lpage>. <pub-id pub-id-type="doi">10.1007/s11242-019-01352-5</pub-id>
</mixed-citation>
</ref>
<ref id="B103">
<label>103.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kamrava</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Im</surname>
<given-names>J</given-names>
</name>
<name>
<surname>de Barros</surname>
<given-names>FPJ</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Estimating dispersion coefficient in flow through heterogeneous porous media by a deep convolutional neural network</article-title>. <source>Geophys Res Lett</source> (<year>2021</year>) <volume>48</volume>:<fpage>e2021GL094443</fpage>. <pub-id pub-id-type="doi">10.1029/2021gl094443</pub-id>
</mixed-citation>
</ref>
<ref id="B104">
<label>104.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dubey</surname>
<given-names>SR</given-names>
</name>
</person-group>. <article-title>A decade survey of content based image retrieval using deep learning</article-title>. <comment>arXiv:2012.00641v2</comment> (<year>2020</year>).</mixed-citation>
</ref>
<ref id="B105">
<label>105.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Shah</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Conjeti</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Navab</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Katouzian</surname>
<given-names>A</given-names>
</name>
</person-group>. <article-title>Deeply learnt hashing forests for content based image retrieval in prostate MR images</article-title>. In: <source>Proceedings of the SPIE 9784</source> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B106">
<label>106.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Conjeti</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Katouzian</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Kazi</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Mesbah</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Beymer</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Syeda-Mahmood</surname>
<given-names>TF</given-names>
</name>
<etal/>
</person-group> <article-title>Metric hashing forests</article-title>. <source>Med Image Anal</source> (<year>2016</year>) <volume>34</volume>:<fpage>13</fpage>&#x2013;<lpage>29</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2016.05.010</pub-id>
<pub-id pub-id-type="pmid">27338173</pub-id>
</mixed-citation>
</ref>
<ref id="B107">
<label>107.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Anavi</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Kogan</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Gelbart</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Geva</surname>
<given-names>O</given-names>
</name>
<name>
<surname>Greenspan</surname>
<given-names>H</given-names>
</name>
</person-group>. <article-title>A comparative study for chest radiograph image retrieval using binary texture and deep learning classification</article-title>. In: <source>Proceedings of 37th annual international conference of the IEEE engineering in medicine and biology society</source>. <publisher-loc>Milan, Italy</publisher-loc>: <publisher-name>EMBC</publisher-name> (<year>2015</year>). p. <fpage>2940</fpage>.</mixed-citation>
</ref>
<ref id="B108">
<label>108.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Doolen</surname>
<given-names>GD</given-names>
</name>
</person-group>. <article-title>Lattice boltzmann method for fluid flows</article-title>. <source>Annu Rev Fluid Mech</source> (<year>1998</year>) <volume>30</volume>:<fpage>329</fpage>&#x2013;<lpage>64</lpage>. <pub-id pub-id-type="doi">10.1146/annurev.fluid.30.1.329</pub-id>
</mixed-citation>
</ref>
<ref id="B109">
<label>109.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tahmasebi</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Kamrava</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Sahimi</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Machine learning in geo- and environmental sciences: from small to large scale</article-title>. <source>Adv Water Resour</source> (<year>2020</year>) <volume>142</volume>:<fpage>103619</fpage>. <pub-id pub-id-type="doi">10.1016/j.advwatres.2020.103619</pub-id>
</mixed-citation>
</ref>
<ref id="B110">
<label>110.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Litjens</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Kooi</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Ehteshami Bejnordi</surname>
<given-names>B</given-names>
</name>
<name>
<surname>Setio</surname>
<given-names>AAA</given-names>
</name>
<name>
<surname>Ciompi</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Ghafoorian</surname>
<given-names>M</given-names>
</name>
<etal/>
</person-group> <article-title>A survey on deep learning in medical image analysis</article-title>. <source>Med Image Anal</source> (<year>2017</year>) <volume>42</volume>:<fpage>60</fpage>&#x2013;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2017.07.005</pub-id>
<pub-id pub-id-type="pmid">28778026</pub-id>
</mixed-citation>
</ref>
<ref id="B111">
<label>111.</label>
<mixed-citation publication-type="web">
<collab>Mayo Clinic</collab>. <article-title>Glioma</article-title> (<year>2025</year>). <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.mayoclinic.org/diseases-conditions/glioma/symptoms-causes/syc-20350251">https://www.mayoclinic.org/diseases-conditions/glioma/symptoms-causes/syc-20350251</ext-link>
</comment> (<comment>Accessed August 15, 2025</comment>).</mixed-citation>
</ref>
<ref id="B112">
<label>112.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wirsching</surname>
<given-names>HG</given-names>
</name>
<name>
<surname>Galanis</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Weller</surname>
<given-names>M</given-names>
</name>
</person-group>. <article-title>Glioblastoma</article-title>. <source>Handb Clin Neurol</source> (<year>2016</year>) <volume>134</volume>:<fpage>381</fpage>&#x2013;<lpage>97</lpage>. <pub-id pub-id-type="doi">10.1016/B978-0-12-802997-8.00023-2</pub-id>
<pub-id pub-id-type="pmid">26948367</pub-id>
</mixed-citation>
</ref>
<ref id="B113">
<label>113.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Walker</surname>
<given-names>MD</given-names>
</name>
<name>
<surname>Green</surname>
<given-names>SB</given-names>
</name>
<name>
<surname>Byar</surname>
<given-names>DP</given-names>
</name>
<name>
<surname>Alexander</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Batzdorf</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Brooks</surname>
<given-names>WH</given-names>
</name>
<etal/>
</person-group> <article-title>Randomized comparisons of radiotherapy and nitrosoureas for the treatment of malignant glioma after surgery</article-title>. <source>N Engl J Med</source> (<year>1980</year>) <volume>303</volume>:<fpage>1323</fpage>&#x2013;<lpage>9</lpage>. <pub-id pub-id-type="doi">10.1056/NEJM198012043032303</pub-id>
<pub-id pub-id-type="pmid">7001230</pub-id>
</mixed-citation>
</ref>
<ref id="B114">
<label>114.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Stupp</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Taillibert</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Kanner</surname>
<given-names>AA</given-names>
</name>
<name>
<surname>Kesari</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Steinberg</surname>
<given-names>DM</given-names>
</name>
<name>
<surname>Toms</surname>
<given-names>SA</given-names>
</name>
<etal/>
</person-group> <article-title>Maintenance therapy with tumor-treating fields plus temozolomide vs temozolomide alone for glioblastoma: a randomized clinical trial</article-title>. <source>JAMA</source> (<year>2015</year>) <volume>31</volume>:<fpage>2535</fpage>&#x2013;<lpage>43</lpage>. <pub-id pub-id-type="doi">10.1001/jama.2015.16669</pub-id>
<pub-id pub-id-type="pmid">26670971</pub-id>
</mixed-citation>
</ref>
<ref id="B115">
<label>115.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kickingereder</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Neuberger</surname>
<given-names>U</given-names>
</name>
<name>
<surname>Bonekamp</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Piechotta</surname>
<given-names>PL</given-names>
</name>
<name>
<surname>G&#xf6;tz</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Wick</surname>
<given-names>A</given-names>
</name>
<etal/>
</person-group> <article-title>Radiomic subtyping improves disease stratification beyond key molecular, clinical, and standard imaging characteristics in patients with glioblastoma</article-title>. <source>Neuro-Oncol</source> (<year>2018</year>) <volume>18</volume>:<fpage>848</fpage>&#x2013;<lpage>57</lpage>. <pub-id pub-id-type="doi">10.1093/neuonc/nox188</pub-id>
<pub-id pub-id-type="pmid">29036412</pub-id>
</mixed-citation>
</ref>
<ref id="B116">
<label>116.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hu</surname>
<given-names>LS</given-names>
</name>
<name>
<surname>D&#x2019;Angelo</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Weiskittel</surname>
<given-names>TM</given-names>
</name>
<name>
<surname>Caruso</surname>
<given-names>FP</given-names>
</name>
<name>
<surname>Fortin Ensign</surname>
<given-names>SP</given-names>
</name>
<name>
<surname>Blomquist</surname>
<given-names>MR</given-names>
</name>
<etal/>
</person-group> <article-title>Integrated molecular and multiparametric MRI mapping of high-grade glioma identifies regional biologic signatures</article-title>. <source>Nat Commun</source> (<year>2023</year>) <volume>28</volume>:<fpage>6066</fpage>. <pub-id pub-id-type="doi">10.1038/s41467-023-41559-1</pub-id>
</mixed-citation>
</ref>
<ref id="B117">
<label>117.</label>
<mixed-citation publication-type="web">
<collab>Mayo Clinic</collab>. <article-title>Meningioma</article-title> (<year>2025</year>). <comment>Available online at: <ext-link ext-link-type="uri" xlink:href="https://www.mayoclinic.org/diseases-conditions/meningioma/symptoms-causes/syc-20355643">https://www.mayoclinic.org/diseases-conditions/meningioma/symptoms-causes/syc-20355643</ext-link>
</comment> (<comment>Accessed August 15, 2025</comment>).</mixed-citation>
</ref>
<ref id="B118">
<label>118.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sultan</surname>
<given-names>HH</given-names>
</name>
<name>
<surname>Salem</surname>
<given-names>NM</given-names>
</name>
<name>
<surname>Al-Atabany</surname>
<given-names>W</given-names>
</name>
</person-group>. <article-title>Multi-classification of brain tumor images using deep neural network</article-title>. <source>IEEE Access</source> (<year>2019</year>) <volume>7</volume>:<fpage>69215</fpage>&#x2013;<lpage>25</lpage>. <pub-id pub-id-type="doi">10.1109/access.2019.2919122</pub-id>
</mixed-citation>
</ref>
<ref id="B119">
<label>119.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Kaldera</surname>
<given-names>HNTK</given-names>
</name>
<name>
<surname>Gunasekara</surname>
<given-names>SR</given-names>
</name>
<name>
<surname>Dissanayake</surname>
<given-names>MB</given-names>
</name>
</person-group>. <article-title>Brain tumor classification and segmentation using faster R-CNN</article-title>. In: <source>Proceedings of 2019 advances in science and engineering technology international conferences (ASET)</source>. <publisher-loc>Dubai, United Arab Emirates</publisher-loc> (<year>2019</year>). p. <fpage>1</fpage>.</mixed-citation>
</ref>
<ref id="B120">
<label>120.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ren</surname>
<given-names>S</given-names>
</name>
<name>
<surname>He</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Sun</surname>
<given-names>J</given-names>
</name>
</person-group>. <article-title>Faster R-CNN: towards real-time object detection with region proposal networks</article-title>. <comment>arXiv:1506.01497v3</comment> (<year>2016</year>).</mixed-citation>
</ref>
<ref id="B121">
<label>121.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sajjad</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Khan</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Muhammad</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Ullah</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Baik</surname>
<given-names>SW</given-names>
</name>
</person-group>. <article-title>Multi-grade brain tumor classification using deep CNN with extensive data augmentation</article-title>. <source>J Comput Sci</source> (<year>2019</year>) <volume>30</volume>:<fpage>174</fpage>&#x2013;<lpage>82</lpage>. <pub-id pub-id-type="doi">10.1016/j.jocs.2018.12.003</pub-id>
</mixed-citation>
</ref>
<ref id="B122">
<label>122.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Swati</surname>
<given-names>ZNK</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>Q</given-names>
</name>
<name>
<surname>Kabir</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ali</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Ali</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Ahmed</surname>
<given-names>S</given-names>
</name>
<etal/>
</person-group> <article-title>Brain tumor classification for MR images using transfer learning and fine-tuning</article-title>. <source>Comput Med Imaging Graph</source> (<year>2019</year>) <volume>75</volume>:<fpage>34</fpage>&#x2013;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1016/j.compmedimag.2019.05.001</pub-id>
<pub-id pub-id-type="pmid">31150950</pub-id>
</mixed-citation>
</ref>
<ref id="B123">
<label>123.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Havaei</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Davy</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Biard</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Courville</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y</given-names>
</name>
<etal/>
</person-group> <article-title>Brain tumor segmentation with deep neural networks</article-title>. <source>Med Image Anal</source> (<year>2017</year>) <volume>35</volume>:<fpage>18</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2016.05.004</pub-id>
<pub-id pub-id-type="pmid">27310171</pub-id>
</mixed-citation>
</ref>
<ref id="B124">
<label>124.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cheng</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Cao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Yun</surname>
<given-names>Z</given-names>
</name>
<etal/>
</person-group> <article-title>Enhanced performance of brain tumor classification <italic>via</italic> tumor region augmentation and partition</article-title>. <source>PloS One</source> (<year>2015</year>) <volume>10</volume>:<fpage>e0140381</fpage>. <pub-id pub-id-type="doi">10.1371/journal.pone.0140381</pub-id>
<pub-id pub-id-type="pmid">26447861</pub-id>
</mixed-citation>
</ref>
<ref id="B125">
<label>125.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Helland</surname>
<given-names>RH</given-names>
</name>
<name>
<surname>Ferles</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Pedersen</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Kommers</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Ardon</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Barkhof</surname>
<given-names>F</given-names>
</name>
<etal/>
</person-group> <article-title>Segmentation of glioblastomas in early post-operative multi-modal MRI with deep neural networks</article-title>. <source>Sci Rep</source> (<year>2023</year>) <volume>13</volume>:<fpage>18897</fpage>. <pub-id pub-id-type="doi">10.1038/s41598-023-45456-x</pub-id>
<pub-id pub-id-type="pmid">37919325</pub-id>
</mixed-citation>
</ref>
<ref id="B126">
<label>126.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yin</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>L</given-names>
</name>
</person-group>. <article-title>AGUnet: annotation-guided U-net for fast one-shot video object segmentation</article-title>. <source>Pattern Recogni</source> (<year>2021</year>) <volume>110</volume>:<fpage>107580</fpage>. <pub-id pub-id-type="doi">10.1016/j.patcog.2020.107580</pub-id>
</mixed-citation>
</ref>
<ref id="B127">
<label>127.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Khan</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Katsagannis</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Sengar</surname>
<given-names>SS</given-names>
</name>
</person-group>. <article-title>RepSE-CBAMNet: a hybrid attention-enhanced CNN for brain tumor detection</article-title>. <source>Stud Health Technol Inform</source> (<year>2025</year>) <volume>327</volume>:<fpage>567</fpage>&#x2013;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.3233/SHTI250401</pub-id>
<pub-id pub-id-type="pmid">40380511</pub-id>
</mixed-citation>
</ref>
<ref id="B128">
<label>128.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Balakrishnan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Sengar</surname>
<given-names>SS</given-names>
</name>
</person-group>. <article-title>RepVGG-GELAN: enhanced GELAN with VGG-STYLE ConvNets for brain tumour detection</article-title>. <comment>arXiv:2405.03541</comment> (<year>2024</year>).</mixed-citation>
</ref>
<ref id="B129">
<label>129.</label>
<mixed-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sikarwar</surname>
<given-names>SS</given-names>
</name>
<name>
<surname>Rana</surname>
<given-names>AK</given-names>
</name>
<name>
<surname>engar</surname>
<given-names>SS</given-names>
</name>
</person-group>. <article-title>Entropy-driven deep learning framework for epilepsy detection using electro encephalogram signals</article-title>. <source>Neurosci</source> (<year>2025</year>) <volume>577</volume>:<fpage>12</fpage>. <pub-id pub-id-type="doi">10.1016/j.neuroscience.2025.05.003</pub-id>
</mixed-citation>
</ref>
<ref id="B130">
<label>130.</label>
<mixed-citation publication-type="book">
<person-group person-group-type="author">
<name>
<surname>Lakshmi</surname>
<given-names>PB</given-names>
</name>
<name>
<surname>Reddy</surname>
<given-names>VD</given-names>
</name>
<name>
<surname>Ghosh</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Sengar</surname>
<given-names>SS</given-names>
</name>
</person-group>. <article-title>Classification of autism spectrum disorder based on brain image data using deep neural networks</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Bhateja</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>XS</given-names>
</name>
<name>
<surname>Ferreira</surname>
<given-names>MC</given-names>
</name>
<name>
<surname>Sengar</surname>
<given-names>SS</given-names>
</name>
<name>
<surname>Travieso-Gonzalez</surname>
<given-names>CM</given-names>
</name>
</person-group>, editors. <source>Evolution in computational intelligence. FICTA 2023. Smart innovation, systems and technologies</source>. <publisher-loc>Singapore</publisher-loc>: <publisher-name>Springer</publisher-name> (<year>2023</year>).</mixed-citation>
</ref>
</ref-list>
<fn-group>
<fn fn-type="custom" custom-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/73058/overview">Alex Hansen</ext-link>, NTNU, Norway</p>
</fn>
<fn fn-type="custom" custom-type="reviewed-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1376257/overview">Sandeep Singh Sengar</ext-link>, Cardiff Metropolitan University, United Kingdom</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/3301051/overview">Steffen Berg</ext-link>, Shell Global Solutions International B.V., Netherlands</p>
</fn>
</fn-group>
</back>
</article>