<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Astron. Space Sci.</journal-id>
<journal-title>Frontiers in Astronomy and Space Sciences</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Astron. Space Sci.</abbrev-journal-title>
<issn pub-type="epub">2296-987X</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">896632</article-id>
<article-id pub-id-type="doi">10.3389/fspas.2022.896632</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Astronomy and Space Sciences</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Towards the Identification and Classification of Solar Granulation Structures Using Semantic Segmentation</article-title>
<alt-title alt-title-type="left-running-head">D&#xed;az Castillo et al.</alt-title>
<alt-title alt-title-type="right-running-head">Semantic Segmentation of Solar Granulation</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>D&#xed;az Castillo</surname>
<given-names>S. M.</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1498533/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Asensio Ramos</surname>
<given-names>A.</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="aff" rid="aff4">
<sup>4</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/958878/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Fischer</surname>
<given-names>C. E.</given-names>
</name>
<xref ref-type="aff" rid="aff5">
<sup>5</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1843976/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Berdyugina</surname>
<given-names>S. V.</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/307000/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Leibniz-Institut f&#xfc;r Sonnenphysik (KIS)</institution>, <addr-line>Freiburg</addr-line>, <country>Germany</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>Physikalisches Institut</institution>, <institution>Albert-Ludwigs-Universit&#xe4;t Freiburg</institution>, <addr-line>Freiburg</addr-line>, <country>Germany</country>
</aff>
<aff id="aff3">
<sup>3</sup>
<institution>Instituto de Astrof&#xed;sica de Canarias (IAC)</institution>, <addr-line>San Crist&#xf3;bal de La Laguna</addr-line>, <country>Spain</country>
</aff>
<aff id="aff4">
<sup>4</sup>
<institution>Departamento de Astrof&#xed;sica</institution>, <institution>Universidad de La Laguna</institution>, <addr-line>San Crist&#xf3;bal de La Laguna</addr-line>, <country>Spain</country>
</aff>
<aff id="aff5">
<sup>5</sup>
<institution>National Solar Observatory (NSO)</institution>, <addr-line>Boulder</addr-line>, <addr-line>CO</addr-line>, <country>United States</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/719600/overview">Bala Poduval</ext-link>, University of New Hampshire, United States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/462158/overview">Reinaldo Roberto Rosa</ext-link>, National Institute of Space Research (INPE), Brazil</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/1743882/overview">Herbert Muthsam</ext-link>, University of Vienna, Austria</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/718020/overview">Jerome Ballot</ext-link>, UMR5277 Institut de Recherche en Astrophysique et Plan&#xe9;tologie (IRAP), France</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: S. M. D&#xed;az Castillo, <email>smdiazcas@leibniz-kis.de</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Astrostatistics, a section of the journal Frontiers in Astronomy and Space Sciences</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>23</day>
<month>06</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>9</volume>
<elocation-id>896632</elocation-id>
<history>
<date date-type="received">
<day>15</day>
<month>03</month>
<year>2022</year>
</date>
<date date-type="accepted">
<day>23</day>
<month>05</month>
<year>2022</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2022 D&#xed;az Castillo, Asensio Ramos, Fischer and Berdyugina.</copyright-statement>
<copyright-year>2022</copyright-year>
<copyright-holder>D&#xed;az Castillo, Asensio Ramos, Fischer and Berdyugina</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>Solar granulation is the visible signature of convective cells at the solar surface. The granulation cellular pattern observed in the continuum intensity images is characterised by diverse structures e.g., bright individual granules of hot rising gas or dark intergranular lanes. Recently, the access to new instrumentation capabilities has given us the possibility to obtain high-resolution images, which have revealed the overwhelming complexity of granulation (e.g., exploding granules and granular lanes). In that sense, any research focused on understanding solar small-scale phenomena on the solar surface is sustained on the effective identification and localization of the different resolved structures. In this work, we present the initial results of a proposed classification model of solar granulation structures based on neural semantic segmentation. We inspect the ability of the <italic>U-net</italic> architecture, a convolutional neural network initially proposed for biomedical image segmentation, to be applied to the dense segmentation of solar granulation. We use continuum intensity maps of the IMaX instrument onboard the <italic>Sunrise I</italic> balloon-borne solar observatory and their corresponding segmented maps as a training set. The training data have been labeled using the multiple-level technique (MLT) and also by hand. We performed several tests of the performance and precision of this approach in order to evaluate the versatility of the <italic>U-net</italic> architecture. We found an appealing potential of the <italic>U-net</italic> architecture to identify cellular patterns in solar granulation images reaching an average accuracy above 80% in the initial training experiments.</p>
</abstract>
<kwd-group>
<kwd>solar physics</kwd>
<kwd>solar granulation</kwd>
<kwd>photosphere&#x2013;convection</kwd>
<kwd>dense segmentation</kwd>
<kwd>deep learning&#x2013;artificial neural network</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>The solar photosphere is the lowest visible layer of the solar atmosphere, where the solar plasma changes from almost completely opaque to almost completely transparent, forming the so-called solar surface (<xref ref-type="bibr" rid="B65">Stix, 2002</xref>). Continuum intensity images of this layer reveal the existence of the <italic>s</italic>olar granulation. It covers most of the solar surface and is characterized by a recurrent and dynamical cellular pattern. Individual elements are called granules, which are relatively small and bright bubble-like structures with horizontal scales in the order of megameters (10<sup>3</sup>&#xa0;km) evolving on timescales of minutes (<xref ref-type="bibr" rid="B45">Nordlund et al., 2009</xref>). Solar granules are evidence of the overturning convection process occurring at the solar interior, where hot plasma rises at their centre, then cools down and sinks downward at the edges (<xref ref-type="bibr" rid="B65">Stix, 2002</xref>). An <italic>intergranular region</italic> forms when the granule&#x2019;s cool plasma drives down into the solar interior. This relatively darker narrow lane surrounding the granules is another identifiable structure at the solar surface. (<xref ref-type="bibr" rid="B45">Nordlund et al., 2009</xref>).</p>
<p>Detailed studies of small-scale phenomena on the solar surface have shown specific and systematic morphological patterns in the granulation. A type of pattern that has been studied extensively is the so-called Exploding granules. They were first described by <xref ref-type="bibr" rid="B9">Carlier et al. (1968</xref>) as special types of granules with sizes 2&#x2013;3&#xa0;times bigger than regular ones, being a product of their rapid horizontal expansion. Based on their morphology, exploding granules are characterized by a reduction in the continuum intensity in their centre, generating a &#x201c;dark dot&#x201d;, which eventuality evolves by fragmenting (<xref ref-type="bibr" rid="B35">Kitai and Kawaguchi, 1979</xref>; <xref ref-type="bibr" rid="B44">Namba, 1986</xref>; <xref ref-type="bibr" rid="B26">Hirzberger et al., 1999</xref>). Several observational and numerical studies revealed that exploding granules have a close relationship with mesogranular dynamics (<xref ref-type="bibr" rid="B13">Dom&#xed;nguez Cerde&#xf1;a, 2003</xref>; <xref ref-type="bibr" rid="B55">Roudier et al., 2003</xref>; <xref ref-type="bibr" rid="B54">Roudier and Muller, 2004</xref>), small-scale magnetic field diffusion and concentration (<xref ref-type="bibr" rid="B56">Roudier et al., 2016</xref>; <xref ref-type="bibr" rid="B41">Malherbe et al., 2018</xref>), and small-scale magnetic flux emergence (<xref ref-type="bibr" rid="B12">De Pontieu, 2002</xref>; <xref ref-type="bibr" rid="B48">Palacios et al., 2012</xref>; <xref ref-type="bibr" rid="B50">Rempel, 2018</xref>; <xref ref-type="bibr" rid="B21">Guglielmino et al., 2020</xref>). Another extensively studied pattern are Bright points, point-like bright elements localized within intergranular lanes and which can be clearly identified in certain photospheric spectral bands such as the Fraunhofer&#x2019;s G band (<xref ref-type="bibr" rid="B43">Muller and Roudier, 1984</xref>). Those are mostly related with magnetic field elements, being perfect tracers of high magnetic field concentrations in intensity images ((<xref ref-type="bibr" rid="B6">Bellot Rubio and Orozco Su&#xe1;rez, 2019</xref>) and references therein). More recently, Granular lanes have been reported as another subgranular pattern of interest (<xref ref-type="bibr" rid="B64">Steiner et al., 2010</xref>). Those are arch-like signatures moving from the boundary of a granule into the granule itself. In general, they do not completely cover the granules and are associated with a linear polarisation signal, which corresponds to the emergence of horizontal magnetic fields (<xref ref-type="bibr" rid="B18">Fischer et al., 2020</xref>). Granular lanes were described in simulations as signatures of underlying tubes of vortex flow with their axis oriented parallel to the solar surface (<xref ref-type="bibr" rid="B64">Steiner et al., 2010</xref>).</p>
<p>The capabilities of the new and upcoming solar telescopes (Daniel K. Inouye Solar Telescope&#x2013;DKIST (<xref ref-type="bibr" rid="B52">Rimmele et al., 2020</xref>) or Balloon-borne telescope Sunrise III (<xref ref-type="bibr" rid="B63">Solanki et al., 2017</xref>)) will provide us with large amounts of unprecedented high-resolution images, which could reveal the next level of complexity of granulation. The statistical study of photospheric plasma dynamics at this level of resolution will rely on the correct identification, classification and localization of systematic structures. For this specific task, automatic solutions can be implemented, for instance, Machine Learning techniques (ML) have demonstrated promising results in classification tasks on solar images (<xref ref-type="bibr" rid="B3">Armstrong and Fletcher, 2019</xref>; <xref ref-type="bibr" rid="B40">Love et al., 2020</xref>; <xref ref-type="bibr" rid="B5">Baek et al., 2021</xref>; <xref ref-type="bibr" rid="B11">Chola and Benifa, 2022</xref>). The demonstrated effectiveness of those algorithms in pattern identification tasks has motivated us toward the exploration of Deep Learning (DL) in semantic segmentation tasks, i.e., producing automatically labelled maps at the pixel level in order to rapidly distinguish diverse granulation patterns, such as described previously.</p>
<p>Machine Learning techniques have acquired high popularity in resolving diverse problems in daily life during the last decade. For instance, giving computers the ability to learn representations without being directly programmed for a specific task has been extensively leveraged in computer vision (<xref ref-type="bibr" rid="B58">Sebe et al., 2005</xref>). Convolutional Neural Networks (CNNs) were particularly developed for image recognition tasks (<xref ref-type="bibr" rid="B37">Le Cun et al., 1997</xref>; <xref ref-type="bibr" rid="B36">Krizhevsky et al., 2012</xref>). Inspired by biological visual perception, CNNs are trained to react to specific image features, starting from simple forms, as lines or edges, and then detecting more complex and abstract patterns in subsequent layers (<xref ref-type="bibr" rid="B19">Ghosh et al., 2020</xref>). Sequentially combining layers inside the network to progressively extract higher-level features is the main line of the DL success (<xref ref-type="bibr" rid="B2">Aloysius and Geetha, 2017</xref>). Taking advantage of large amounts of data, this approach may achieve unprecedented performance on several perception tasks, e.g., instance classification (<xref ref-type="bibr" rid="B61">Simonyan and Zisserman, 2015</xref>; <xref ref-type="bibr" rid="B28">Huang et al., 2017</xref>), object detection (<xref ref-type="bibr" rid="B20">Girshick, 2015</xref>) or optical flow estimation (<xref ref-type="bibr" rid="B29">Ilg et al., 2017</xref>).</p>
<p>Another task that saw an important push forward with DL was dense prediction, i.e., prediction at a pixel level in images, such as semantic segmentation (<xref ref-type="bibr" rid="B59">Shelhamer et al., 2017</xref>; <xref ref-type="bibr" rid="B10">Chen et al., 2018</xref>), which solves the classification problem working at pixel resolution. More specifically, the aim is to group the pixels of an image into categories, providing precise localization of labeled structures. Additionally, semantic segmentation seeks to partition the image into semantic meaningful parts (<xref ref-type="bibr" rid="B66">Szeliski, 2011</xref>). This paradigm has been successfully addressed using Encoder-Decoder architectures (<xref ref-type="bibr" rid="B4">Badrinarayanan et al., 2017</xref>; <xref ref-type="bibr" rid="B69">Yanli and Pengpeng, 2021</xref>). Leveraging the properties of CNNs, this type of architecture is capable of producing spatially consistent classification maps, thus providing precise localization of objects of interest.</p>
<p>In this work, we propose to train supervisedly and evaluate the performance of a CNN to carry out solar granulation segmentation. To this end, we apply an encoder-decoder architecture called <italic>U-net</italic> (<xref ref-type="bibr" rid="B53">Ronneberger et al., 2015</xref>). This architecture was developed for biomedical image segmentation tasks, and it is especially interesting for our objectives since it has been successfully applied to cellular pattern segmentation. It can work with few training images, and it achieves high levels of accuracy in the localization of specific structures (<xref ref-type="bibr" rid="B53">Ronneberger et al., 2015</xref>).</p>
</sec>
<sec id="s2">
<title>2 Methods</title>
<sec id="s2-1">
<title>2.1 <italic>U-Net</italic> Architecture</title>
<p>A <italic>U-net</italic> is composed of fully CNN layers organized in an encoder-decoder architecture (<xref ref-type="bibr" rid="B53">Ronneberger et al., 2015</xref>)<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref>. The encoder part (left side of <xref ref-type="fig" rid="F1">Figure 1</xref>) is responsible for producing a low dimensionality dense representation of an image. In the initial stage, the contracting network receives an image of a specific size <italic>H</italic> (height) &#xd7; <italic>W</italic> (width), which is downsampled by sequential layers, each composed by the following operations:<list list-type="simple">
<list-item>
<p>1. Two 3 &#xd7; 3 padded convolution operations each followed by a rectified linear unit (entire operation represented by orange arrows in the <xref ref-type="fig" rid="F1">Figure 1</xref>). As the fundamental components of the convolutional neural network, 2D convolution operations transform the image into feature maps using a set of filters or <italic>kernels.</italic> Those resulting feature maps then pass through a non-linear activation function. We use a set of 64 3 &#xd7; 3 kernels generating features maps of 64 (depth) &#xd7; <italic>H</italic> &#xd7; <italic>W</italic> in the initial operation. Then for the subsequent one, each kernel will have a depth dimension, which corresponds to the feature map depth previously generated. We used padded operations in order to not change the size of the input map during the convolution operations (<xref ref-type="bibr" rid="B14">Dumoulin and Visin, 2016</xref>). <italic>U-net</italic> convolutional operations use a rectified linear unit (ReLU) as the default activation function, which gives the non-linear character to the network. This function is characterized by being linear for input positive values and zero for input negative values. It is well-behaved and converges fast when using the stochastic gradient descent algorithm. Consequently, it is commonly used as an activation function in deep neural networks (<xref ref-type="bibr" rid="B57">Schmidhuber, 2014</xref>).</p>
</list-item>
<list-item>
<p>2. A 2 &#xd7; 2 max-pooling operation with stride 2, which reduces the dimensions of the input map by computing the maximum value of each successive 2 &#xd7; 2 pixel set to produce a downsampled map (pink arrows in <xref ref-type="fig" rid="F1">Figure 1</xref>). During this process, the spatial information is reduced by a factor of two, while the feature information is increased by a factor of two.</p>
</list-item>
</list>
</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Schematic sketch of the <italic>U-net</italic> architecture used, going sequentially from left to right. Each blue box corresponds to a feature map. The depth of the map is denoted on top of the box. Gray boxes represent copied feature maps. The arrows denote the different operations. We use 64 kernels (feature information) for the initial kernel set, whose number increases sequentially in the contraction levels and decreases in the expansion levels as seen in the feature maps depth values. Based on the intrinsic properties of a fully convolutional neural network, the height (H) and width (W) of the input images can be arbitrary numbers but must be equal (squared maps). For the training procedure, we use input maps of sizes of 128 &#xd7; 128. Sketch modified from (<xref ref-type="bibr" rid="B53">Ronneberger et al., 2015</xref>).</p>
</caption>
<graphic xlink:href="fspas-09-896632-g001.tif"/>
</fig>
<p>When the lower level is reached, the lower feature map is then expanded by upsampling sequential layers in the decoder part (right side of the <xref ref-type="fig" rid="F1">Figure 1</xref>), which is responsible for recovering the initial spatial dimension. This expanding network consists of upsampled layers, each composed by:<list list-type="simple">
<list-item>
<p>1. Two 3 &#xd7; 3 padded convolution operations, each followed by a rectified linear unit (orange arrows in the <xref ref-type="fig" rid="F1">Figure 1</xref>) equivalent to the operations of the encoder part.</p>
</list-item>
<list-item>
<p>2. A 2 &#xd7; 2 transposed convolution with stride 2 as the upsampling operation (green arrows in <xref ref-type="fig" rid="F1">Figure 1</xref>). Those seek to reverse the encoder downsampling operations, while broadcasting input elements via a set of 2 &#xd7; 2 kernels, thereby producing an output that is larger than the input (<xref ref-type="bibr" rid="B14">Dumoulin and Visin, 2016</xref>). During this process, the spatial information is increased by a factor of two, while the feature information is reduced by a factor of two.</p>
</list-item>
</list>
</p>
<p>As the outstanding component of the <italic>U-net</italic> architecture, each expansion layer is concatenated with high-resolution features from the encoder path at the same level (see grey blocks in <xref ref-type="fig" rid="F1">Figure 1</xref>), giving the network the capacity to localize with precision. We use five levels of contraction and expansion, like in the original <italic>U-net</italic> model, giving it its characteristic symmetrical shape. Finally, at the end of the sequence, a 1 &#xd7; 1 convolution operation produces probability maps per class as output with the same sizes as the original map. Using the configuration shown in <xref ref-type="fig" rid="F1">Figure 1</xref>, our model employs around 31 million trainable parameters or hyperparameters.</p>
</sec>
<sec id="s2-2">
<title>2.2 Ground-Truth Data</title>
<sec id="s2-2-1">
<title>2.2.1 Observational Dataset: IMaX/Sunrise</title>
<p>In order to train the model using supervised training, we should provide it with a suitable set of data for the segmentation process, i.e., a ground-truth. We are interested in classifying specific patterns in observational images of the solar surface seen in the continuum. Based on our requirements of high-spatial-resolution data, we select the products of the Imaging Magnetograph eXperiment (IMaX) (<xref ref-type="bibr" rid="B42">Mart&#xed;nez Pillet et al., 2011</xref>), the filterpolarimeter onboard the Sunrise I balloon-borne telescope during its flight in June 2009 (<xref ref-type="bibr" rid="B72">Barthol et al., 2011</xref>; <xref ref-type="bibr" rid="B71">Solanki et al., 2010</xref>). IMaX was tuned to the Fe<sc>i</sc> at 525.02&#xa0;nm highly Zeeman-sensitive line, and provided measurements of the local continuum intensity near this line. After phase diversity reconstruction, each map reached a spatial resolution of around 0<italic>&#x201d;</italic>.15 &#x2248; 100&#xa0;km over the solar surface (pixel scale 0<italic>&#x201d;</italic>.05) and a field of view (FOV) of 50<italic>&#x201d;</italic> &#xd7; 50<italic>&#x201d;</italic> &#x2248; 35 &#xd7; 35&#xa0;Mm (<xref ref-type="bibr" rid="B42">Mart&#xed;nez Pillet et al., 2011</xref>). Those data products are freely accessible on <ext-link ext-link-type="uri" xlink:href="https://star.mps.mpg.de/sunrise/">https://star.mps.mpg.de/sunrise/</ext-link>. We select a time series of 56&#xa0;min with a cadence of 30&#xa0;s resulting in 113 individual frames. We selected the frames with the highest quality and spread out in time to obtain as much as possible a diverse data set. Due to the apodization needed for image reconstruction, a portion of the edges was lost. Taking all the above in consideration, our ground-truth dataset is composed of eight frames of 768 &#xd7; 768 pixels each, with a FOV of 38<italic>&#x201d;</italic> &#xd7; 38<italic>&#x201d;</italic> (see one frame as example in <xref ref-type="fig" rid="F2">Figure 2</xref> map A).</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>Frame example of IMaX/Sunrise during the labeling procedure. <bold>(A)</bold> Reconstructed continuum intensity map, <bold>(B)</bold> Initial segmentation results using 25 descending detection thresholds, <bold>(C)</bold> Merging and shrinking results, <bold>(D)</bold> Result map differentiating integranular lanes and granules cells as single units, <bold>(E)</bold> Manual selection into defined categories: intergranular lane (dark violet), uniform-shaped granules (pink), granules with a dot (white), granules with a lane (light green) and complex-shaped granules (dark green).</p>
</caption>
<graphic xlink:href="fspas-09-896632-g002.tif"/>
</fig>
</sec>
<sec id="s2-2-2">
<title>2.2.2 Labeling Structures</title>
<p>In a supervised learning approach, we need to provide an initial truth segmentation of our selected dataset in order to properly train the model. In previous studies, the identification and tracking of specific granular structures have been done mostly manually with the help of intensity multi-threshold algorithms (<xref ref-type="bibr" rid="B31">Javaherian et al., 2014</xref>; <xref ref-type="bibr" rid="B15">Ellwarth et al., 2021</xref>). For our experiment, we select a common multi-threshold algorithm MLT4 (<xref ref-type="bibr" rid="B7">Bovelet and Wiehr, 2001</xref>; <xref ref-type="bibr" rid="B8">Bovelet and Wiehr, 2007</xref>) used for segmenting photospheric structures for the initial granular identification (see for instance (<xref ref-type="bibr" rid="B51">Riethm&#xfc;ller et al., 2008</xref>; <xref ref-type="bibr" rid="B17">Fischer et al., 2019</xref>; <xref ref-type="bibr" rid="B32">Kaithakkal and Solanki, 2019</xref>)) which is freely available<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref>. We adopt this approach to assess the extent to which user intervention affects the training process on the network. In particular, for labeling our structures at a pixel level we follow a procedure composed of two phases:<list list-type="simple">
<list-item>
<p>1. Semi-automatic granules identification: Using the Multiple-Level Pattern Recognition algorithm&#x2013;MLT4 (<xref ref-type="bibr" rid="B7">Bovelet and Wiehr, 2001</xref>; <xref ref-type="bibr" rid="B8">Bovelet and Wiehr, 2007</xref>), we segregated the intergranular regions and the granular pattern. This is a top-down segmentation technique of brightness structures based on a sequence of descending detection thresholds (<xref ref-type="bibr" rid="B8">Bovelet and Wiehr, 2007</xref>). The algorithm uses the reconstructed and normalized continuum intensity maps as input (see one frame as an example in <xref ref-type="fig" rid="F2">Figure 2</xref> map A). The procedure starts with an initial segmentation of features at equidistant intensity levels as shown in map B of <xref ref-type="fig" rid="F2">Figure 2</xref>, and then the pixel brightness level is normalized within each cell to its maximum value. Consecutively, a semi-automatic procedure of merging over-segmented cells and (4) shrinking these brightness-normalized cells to features of adequate sizes is performed, resulting in maps such as map C of <xref ref-type="fig" rid="F2">Figure 2</xref>. Regarding the setup parameters used, we selected 25 descending thresholds, 0.47 as a normalized reference for merging and 0.38 as the unitary cut threshold for shrinking. The unitary cut threshold controls the final size of the recognized features, initially derived from a normalized brightness histogram for the full sample of recognized cells, which was then tuned by visual inspection. The rest of input parameters were set to their default value (<xref ref-type="bibr" rid="B7">Bovelet and Wiehr, 2001</xref>). The resulting maps are composed of several hundred individual cells (granules) separated from the intergranular space as shown in map D of the <xref ref-type="fig" rid="F2">Figure 2</xref>.</p>
</list-item>
<list-item>
<p>2. Fully manual granules classification: Based on the basic instantaneous morphological features of the granular phenomena that we seek to classify, we propose an initial set of granule classes characterised by the presence of a central dot signatures or an arch-like lane signatures. For completeness, we include two categories that refer to extreme levels of complexity in granules: 1) morphologies with low complexity, i.e. uniform and clean morphologies with circular or ellipsoid shapes, and 2) morphologies with high complexity, i.e. abnormal granules having combinations of dark spots or lanes. In that sense and using the map products of the previous procedure, we classified manually the set of individual cells into four categories: granules with a dot (cells with dark point-shape features close to the centre of the cell), granules with a lane (cells with a dark arch-like lane following a bright rim mark inside the cell structure), uniform-shaped granules (cells with uniform intensity distribution and with elliptical or circular shapes) and complex-shaped granules (all remaining cells). The map E of the <xref ref-type="fig" rid="F2">Figure 2</xref> shows the results of the manual classification, where each colour corresponds to a specific classification, equivalently to the ground-truth maps in the <xref ref-type="fig" rid="F4">Figure 4A</xref>; <xref ref-type="fig" rid="F5">Figure 5A</xref>. During the selection via visual inspection, we pursue to classify all individual granules per class unequivocally to avoid ambiguity.</p>
</list-item>
</list>
</p>
<p>We perform this two-step procedure for all pixels of the eight selected frames, generating one ground-truth labeled map for each continuum intensity map. We are interested in evaluating independently and unbiased the performance of our model and simultaneously providing it with as many training examples as possible, thus we split our dataset in such a way that seven frames are used for the training set and one is used for the validation/test set. As an example, <xref ref-type="fig" rid="F2">Figure 2</xref> shows the intermediate steps in the complete labeling procedure for the validation/test map.</p>
</sec>
</sec>
<sec id="s2-3">
<title>2.3 Training Strategy</title>
<p>Although the <italic>U-net</italic> architecture has demonstrated a good performance even with a few per-class training examples, it is essential to provide it with a large and diverse set of training data. In that sense, we divided the full FOV of all available maps into several sub-maps of a fixed size. As we are interested in predicting the class of each granule, we select sub-maps of the size 128 &#xd7; 128 pixels <inline-formula id="inf1">
<mml:math id="m1">
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mo>&#x223c;</mml:mo>
<mml:mn>6.5</mml:mn>
<mml:mi>&#x201d;</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>6.5</mml:mn>
<mml:mi>&#x201d;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> as input, with the aim of covering entire granules (see <xref ref-type="fig" rid="F1">Figure 1</xref>). In addition, we applied a process of data augmentation, including random rotations, random perspective transformations and warping.</p>
<p>We identified a severely skewed class distribution in the labeled data, where 85% of the pixels of all available maps are associated to two classes (intergranular lane 40% and complex-shaped granules 45%) and the remaining 15% of the pixels belong to the underrepresented classes (granules with a dot 8%, granules with a lane 3% and uniform-shaped granules 4%). This is a known difficulty that affects all classification machine learning algorithms because the metrics used for training assume an equivalent proportion of examples of each class. This assumption decreases the performance of the model for underrepresented classes (<xref ref-type="bibr" rid="B23">He and Garcia, 2009</xref>; <xref ref-type="bibr" rid="B16">Fern&#xe1;ndez et al., 2018</xref>). Many strategies have been developed to overcome this issue in computer vision paradigms [see, e.g., (<xref ref-type="bibr" rid="B25">He et al., 2008</xref>; <xref ref-type="bibr" rid="B23">He and Garcia, 2009</xref>; <xref ref-type="bibr" rid="B24">He and Ma, 2013</xref>; <xref ref-type="bibr" rid="B27">Huang et al., 2016</xref>; <xref ref-type="bibr" rid="B33">Khan et al., 2017</xref>; <xref ref-type="bibr" rid="B46">Oksuz et al., 2020</xref>)], however, it is still an active topic research in semantic segmentation tasks [see, e.g., (<xref ref-type="bibr" rid="B22">Havaei et al., 2017</xref>; <xref ref-type="bibr" rid="B47">Ol&#xe3; Bressan et al., 2022</xref>; <xref ref-type="bibr" rid="B70">Zou et al., 2021</xref>)].</p>
<p>We addressed the imbalance-class issue in this work by including a stratified random sub-map sampling previous to the augmentation procedure as follows. 1) We defined weighted pixel maps for each full image, in which the greater weights were given to areas where underrepresented classes were localized. 2) We applied a softmax function to compute probability distribution maps. Those probability distributions were included in a weighted random choice function, which returned sub-maps centred on underrepresented classes regions. With this method, we increased the pixel proportion of the underrepresented classes to 22% in our training dataset. We noticed that this proportion has an upper limit due to the size of the sub-maps. The reason is that the surface covered by underrepresented classes is smaller than the size of the sub-maps, which is mostly covered by the background classes (i.e., intergranular regions and several complex-shaped granules).</p>
<p>An additional strategy towards solving class imbalance is an appropriate selection of the loss function. Neural networks applications learn via optimization, which requires a suitable cost/loss functions to calculate the model error. The iterative process of hyperparameter tuning is controlled by the loss function minimization, which, at the end of the training, ideally provides the best model setup for the assigned task. In particular, metrics for semantic segmentation have been historically dominated by global approaches, like the Cross-Entropy loss (<xref ref-type="bibr" rid="B1">Aggarwal, 2018</xref>). Defined as <italic>CE</italic> &#x3d; &#x2212; log(<italic>p</italic>
<sub>
<italic>t</italic>
</sub>), where <italic>p</italic>
<sub>
<italic>t</italic>
</sub> corresponds to the estimated probability for a correct classification for a specific class <italic>t</italic>, the cross-entropy loss evaluates the overall proportion of the correctly classified pixels as the precision measurement. However, these scores are dominated by the background classes in skewed datasets. Typically, the addition of a cost-sensitive weighting factor <italic>&#x3b1;</italic> is used in cross-entropy, known as <italic>&#x3b1;</italic>-balance variant. This seeks to balance the importance of well-classified over the wrong-classified examples in cases of skewed datasets. For several classes, the <italic>&#x3b1;</italic> factor can be considered as a weight vector with values inversely proportional to the frequency of each class (<xref ref-type="bibr" rid="B39">Lin et al., 2017</xref>).</p>
<p>For our experiments, we test the accuracy and effectiveness of two different loss functions during the network training, which are commonly used for imbalanced data problems:<list list-type="simple">
<list-item>
<p>1. The Focal Loss was developed for addressing the unbalance-class problem by adding a modulating factor <inline-formula id="inf2">
<mml:math id="m2">
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mi>p</mml:mi>
</mml:mrow>
<mml:mrow>
<mml:mi>t</mml:mi>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mi>&#x3b3;</mml:mi>
</mml:mrow>
</mml:msup>
</mml:math>
</inline-formula> to the cross-entropy loss. It uses the tunable focusing parameter <italic>&#x3b3;</italic> &#x2265; 0, which adjusts the rate at which background examples are down-weighted. This modification downplays the importance of the background classes, making the training to focus on learning the hard examples, i.e. weakly represented classes (<xref ref-type="bibr" rid="B39">Lin et al., 2017</xref>). The use of <italic>&#x3b1;</italic>-balance variant is also applicable in this case.</p>
</list-item>
<list-item>
<p>2. The Intersection-over-Union (IoU) loss or Jaccard index was extensively used in semantic segmentation tasks. It is focused on determining the similarity between finite sample sets (<xref ref-type="bibr" rid="B30">Jaccard, 1912</xref>). For images, the IoU measures the agreement between any predicted region and its corresponding ground-truth region by measuring the intersection between the prediction and the ground-truth normalized by their union. The IoU loss can take into account the frequency of the classes, and thus it is considered robust to the class imbalance problem (<xref ref-type="bibr" rid="B38">Leivas Oliveira, 2019</xref>). For multi-class classification tasks like the one we pursue here, the mean IoU (mIoU) loss function is often used, which initially computes the Jaccard index for each class and then computes the average over all classes.</p>
</list-item>
</list>
</p>
<p>For comparative purposes, we computed the standard evaluation metrics for semantic segmentation. We computed the overall accuracy, measured as the ratio between the correctly predicted pixels and the total number of pixels, and the mean pixel accuracy per class measured as the average of the correctly predicted pixels per class over the total ground-truth pixels per class. Likewise, we compare the test performance with performance parameters during the execution of the training. In this sense, we monitor the average per epoch of the loss value given its loss function (average loss) and the average overall accuracy per epoch (accuracy).</p>
</sec>
</sec>
<sec id="s3">
<title>3 Results</title>
<p>We implemented the model and the training using the open-source <italic>PyTorch</italic> (<xref ref-type="bibr" rid="B49">Paszke et al., 2019</xref>) framework<xref ref-type="fn" rid="fn3">
<sup>3</sup>
</xref>. All our training cases have been performed with an NVIDIA GeForce 2080 Ti GPU and an NVIDIA Tesla P100 GPU. We generated 27,000 sub-maps in the training dataset and 3,000 sub-maps in the validation dataset from the augmentation procedure previously described. We used the Adam stochastic gradient algorithm (<xref ref-type="bibr" rid="B34">Kingma and Ba, 2014</xref>) for optimization. In this case, the gradient is estimated from subsets of the training dataset or <italic>batches</italic>. We used batches of 32 samples, generating around 840 subsets containing all the training dataset, which are used to update the hyperparameters and to computed the metric in each cycle or <italic>epoch</italic>. We considered 100 or 200 epochs depending on the loss function that we used, thus we executed around 84,000 or 168,000 iterations in total respectively. The learning rate was annealed from the initial value of 0.001 with a dynamic adjustment, lowering the value by a factor 0.9 when the minimization did not decrease in 5 consecutive epochs. We performed several training cases by turning off some of the transformations of the data augmentation process, changing the weight values for the stratified random sampling maps, and changing the loss functions and its parameters. Using this setup, each epoch took <inline-formula id="inf3">
<mml:math id="m3">
<mml:mo>&#x223c;</mml:mo>
<mml:mn>210</mml:mn>
</mml:math>
</inline-formula>&#xa0;s.</p>
<p>We present two training cases that reached the highest accuracies from all testing that we ran. Test 1 uses the <italic>U-net</italic> architecture with a <italic>&#x3b1;</italic>-balanced Focal Loss. For this experiment, we consider a proportion of 1:100 for the background classes with respect to the underrepresented classes to build the weighted pixel maps for the stratified sampling selection. On the other hand, Test 2 uses the <italic>U-net</italic> architecture using the mean IoU as the loss function. Similarly, we consider a proportion of 1:100 for the background classes with respect to the underrepresented classes to build the weighted pixel maps. In both experiments, we only used random rotations and perspective transformations as augmentation. <xref ref-type="fig" rid="F3">Figures 3A,B</xref> present the monitoring plots of these two test cases during the training execution. These graphs show the evolution of the performance parameters of the selected metrics per epoch, i.e., average loss and accuracy. For both cases, the performance parameters behave as expected for the training dataset (see blue curves in <xref ref-type="fig" rid="F3">Figures 3A,B</xref>), however, the parameters associated with the validation dataset show differences between each other.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>
<bold>(A)</bold> Tracking plot for the training for test 1: <italic>U-net</italic> model using <italic>&#x3b1;</italic>-balanced Focal loss with <italic>&#x3b3;</italic> &#x3d; 10 as training loss function. <bold>(B)</bold> Tracking training plot for test 2: <italic>U-net</italic> model using Mean IoU as training loss function.</p>
</caption>
<graphic xlink:href="fspas-09-896632-g003.tif"/>
</fig>
<p>For test 1, the overall pixel accuracy, the accuracy per class and the Jaccard index for the validation dataset increase slowly over every iteration reaching 0.84, 0.60, and 0.47 respectively, while its corresponding loss increases heavily. We interpret that the noise and rising trend in the loss profile are due to the accumulation of misclassified examples, such as pixels at the edge of granules or clusters of pixels that have features of multiple classes, which the model slowly corrects thus improving the accuracy. The model reaches high levels of saturation, with hints a slight overfitting close to the end of the training (see <xref ref-type="fig" rid="F3">Figure 3A</xref>).</p>
<p>This effect can be observed in the full map prediction and in the predicted probability distribution per class shown in <xref ref-type="fig" rid="F4">Figures 4A,B</xref>. In the whole map, the model reaches 0.74 of overall pixel accuracy, a mean accuracy per class of 0.52 and a Jaccard index of 0.40. These values are slightly lower than those achieved during training (evaluated in sub-maps of size 128 &#xd7; 128) but still compatible, since they come from the same map. <xref ref-type="fig" rid="F4">Figure 4A</xref> displays the direct comparison between the ground-truth map with the predicted map. As a first conclusion, we highlight the efficiency of the model to segregate granules and intergranular lanes, which contributes mostly to the overall pixel accuracy.</p>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>
<bold>(A)</bold> Comparative maps using the validation full size map&#x2013;Test 1: <italic>U-net</italic> model using <italic>&#x3b1;</italic>-balanced Focal loss with <italic>&#x3b3;</italic> &#x3d; 10 as training loss function. A zoomed region is highlighted showing in detail the diverse constrains of the segmentation. <bold>(B)</bold> Probability maps per class using the validation full size map&#x2013;Test 1: <italic>U-net</italic> model using <italic>&#x3b1;</italic>-balanced Focal loss with <italic>&#x3b3;</italic> &#x3d; 10 as training loss function.</p>
</caption>
<graphic xlink:href="fspas-09-896632-g004.tif"/>
</fig>
<p>Regarding the correct identification of underrepresented morphologies, the model behaviour is different per class. Based on the probability maps in <xref ref-type="fig" rid="F4">Figure 4B</xref>, we identify that the model develops different reliability levels depending on the class. For clean morphologies such as uniformly shaped granules, the model is slightly more confident as compared with more structured and complex classes. In that sense, granules with multiple and similar features give rise to a prediction with a high degree of uncertainty. This is manifested in classes such as granules with a dot, granules with a lane and complex-shaped granules.</p>
<p>On the other hand, test 2 shows a completely different behaviour. As shown in <xref ref-type="fig" rid="F3">Figure 3B</xref>, the performance parameters related to the validation dataset reaches a threshold at an early stage of training without appreciable changes along the epochs. From the first cycle, a value of 0.87 for overall pixel accuracy and a Jaccard index of 0.52 are achieved. However, in terms of the mean accuracy per class, the average threshold value during the first 20 iterations is 0.64, which then decreases and stabilizes around 0.60 correspondingly. Using this training setup, we suggest that the model is able to learn the basic morphological patterns from few batches, but it is unable to extract more specific information from the full dataset provided during the training. Signatures of over-fitting are also observed, but the invariance of the loss for the training and validation datasets indicates an upper limit in the learning process in the defined training time.</p>
<p>
<xref ref-type="fig" rid="F5">Figure 5A</xref> shows the full map prediction and the predicted probability distribution maps of the model with the lowest loss value obtained during the training (epoch 12). In the whole map, the model reaches 0.71 of overall pixel accuracy, a mean accuracy per class of 0.58 and a Jaccard index of 0.40. Again, these values are slightly lower than those achieved during training (evaluated in sub-maps of size 128 &#xd7; 128) but still compatible, since they come from the same map. A good efficiency to segregate granules and intergranular lanes is also obtained in this test, contributing mostly to the overall pixel accuracy as well (see the <xref ref-type="fig" rid="F5">Figure 5A</xref>).</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>
<bold>(A)</bold> Comparative maps using the validation full size map&#x2013;Test 2: <italic>U-net</italic> model using Mean IoU as training loss function. A zoomed region is highlighted showing in detail the diverse constrains of the segmentation. <bold>(B)</bold> Probability maps per class using the validation full size map&#x2013;Test 2: <italic>U-net</italic> model using Mean IoU as training loss function.</p>
</caption>
<graphic xlink:href="fspas-09-896632-g005.tif"/>
</fig>
<p>Based on the probability maps generated (see <xref ref-type="fig" rid="F5">Figure 5B</xref>), we notice that the model reproduces high levels of confidence in all classes, managing to identify detailed morphological patterns associated with the classes, i.e., dots, lanes or combinations even within individual cells, which promotes the over-labelling of structures, i.e., single granules contain pixels of different classes as shown in the predicted map of <xref ref-type="fig" rid="F5">Figure 5A</xref>.</p>
<p>So far, we have been referring to class-average quantities of the performance parameters, which are biased by the well-know imbalance between granulation structures. In <xref ref-type="table" rid="T1">Table 1</xref>, we present a summary of the overall pixel accuracy (OPA) and the Jaccard index per class. As we mentioned, the identification of the intergranular lane provides the major contribution to the effectiveness of the models, reaching accuracy values around 0.90 in both metrics. On the other hand, the identification of specific morphologies has a significantly lower performance, especially for underrepresented classes. We identify that the contribution of precision and recall on the Jaccard index are unequal, i.e. the models are acceptably sensitive (higher recall) in detecting the simple shapes characteristics of each class, but those are inexact during the ground-truth comparison.</p>
<table-wrap id="T1" position="float">
<label>TABLE 1</label>
<caption>
<p>Summary of the performance parameters per class calculated for the full size verification map prediction.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left"/>
<th colspan="2" align="center">IGL<xref ref-type="table-fn" rid="Tfn1">
<sup>a</sup>
</xref>
</th>
<th colspan="2" align="center">UG<xref ref-type="table-fn" rid="Tfn2">
<sup>b</sup>
</xref>
</th>
<th colspan="2" align="center">DG<xref ref-type="table-fn" rid="Tfn3">
<sup>c</sup>
</xref>
</th>
<th colspan="2" align="center">LG<xref ref-type="table-fn" rid="Tfn4">
<sup>d</sup>
</xref>
</th>
<th colspan="2" align="center">CG<xref ref-type="table-fn" rid="Tfn5">
<sup>e</sup>
</xref>
</th>
</tr>
<tr>
<td align="left"/>
<td align="center">OPA<xref ref-type="table-fn" rid="Tfn6">
<sup>f</sup>
</xref>
</td>
<td align="center">Jaccard</td>
<td align="center">OPA</td>
<td align="center">Jaccard</td>
<td align="center">OPA</td>
<td align="center">Jaccard</td>
<td align="center">OPA</td>
<td align="center">Jaccard</td>
<td align="center">OPA</td>
<td align="center">Jaccard</td>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">Test 1: Focal Loss</td>
<td align="char" char=".">0.90</td>
<td align="char" char=".">0.85</td>
<td align="char" char=".">0.12</td>
<td align="char" char=".">0.10</td>
<td align="char" char=".">0.58</td>
<td align="char" char=".">0.29</td>
<td align="char" char=".">0.31</td>
<td align="char" char=".">0.15</td>
<td align="char" char=".">0.71</td>
<td align="char" char=".">0.58</td>
</tr>
<tr>
<td align="left">Test 2: mIoU</td>
<td align="char" char=".">0.95</td>
<td align="char" char=".">0.90</td>
<td align="char" char=".">0.44</td>
<td align="char" char=".">0.20</td>
<td align="char" char=".">0.67</td>
<td align="char" char=".">0.31</td>
<td align="char" char=".">0.33</td>
<td align="char" char=".">0.11</td>
<td align="char" char=".">0.54</td>
<td align="char" char=".">0.49</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn id="Tfn1">
<label>a</label>
<p>Intergranular lane.</p>
</fn>
<fn id="Tfn2">
<label>b</label>
<p>Uniform-shaped granules.</p>
</fn>
<fn id="Tfn3">
<label>c</label>
<p>Granules with a dot.</p>
</fn>
<fn id="Tfn4">
<label>d</label>
<p>Granules with a lane.</p>
</fn>
<fn id="Tfn5">
<label>e</label>
<p>Complex-shaped granules.</p>
</fn>
<fn id="Tfn6">
<label>f</label>
<p>Overall Pixel Accuracy.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>Based on the multi-class confusion matrix at the pixel level for each model shown in the <xref ref-type="fig" rid="F6">Figures 6A,B</xref>, we notice specific behavior per model. For test 1, the high uncertainty levels generate a strong effect on the maximum probability to match a category given the original category. In this case, granules belonging to under-represented classes have a high tendency to be classified as complex-shape granules due to the homogeneous probabilities between classes in granules where similar morphologies are shared. On the other hand, for test 2, the over-labeling effect and the high levels of reliability, tend to homogenize the maximum probabilities by class, negatively affecting classes such as uniform-shaped granules and granules with a lane.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>
<bold>(A)</bold> Probability distribution at pixel level of the category prediction given its original category (confusion matrix) for test 1: <italic>U-net</italic> model using <italic>&#x3b1;</italic>-balanced Focal loss with <italic>&#x3b3;</italic> &#x3d; 10 as training loss function. <bold>(B)</bold> Probability distribution at pixel level of the category prediction given its original category (confusion matrix) for test 2: <italic>U-net</italic> model using Mean IoU as training loss function.</p>
</caption>
<graphic xlink:href="fspas-09-896632-g006.tif"/>
</fig>
</sec>
<sec id="s4">
<title>4 Discussion</title>
<p>In summary, we present the first attempts to classify and identify structures in the solar granulation based on the semantic segmentation paradigm using a deep learning method. As our main objective, we found an interesting potential of the <italic>U-net</italic> architecture to identify and classify cellular patterns in solar granulation images, but modifications to the current model should be implemented to ensure its optimal performance. With the proposed training procedure, the model achieves high levels of accuracy in the identification of the intergranular network which allows the effective separation of granular morphologies. We have also established that the network architecture is sensitive in identifying characteristic patterns in granules, such as granules with a dot (overall accuracy greater than 0.5 in both tests), but it looses efficacy when it comes to discerning between structures with combined morphologies, i.e., granules with multiple features and complex-shaped structures. This outcome drives high uncertainty levels (test 1) or an over-labeling effect in single granules (test 2).</p>
<p>During our experiments, we have identified recurrent hints of overfitting in all performed tests, meeting the highest accuracy for the tests presented here. We implemented some functional strategies such as the Dropout regularization and hyperparameters scaling but without obtaining any improvement. Going further, we identified that the preparation of the ground-truth dataset played a crucial role in the model generalization ability. The semi-manual and manual labeling process introduced unwanted constraints, e.g., over-merging, poor contours separation and small incorrect areas. Moreover, we noted the difficulty in defining closed classification criteria, which would allow us to represent the samples of each category unambiguously. Labeling structures of this specific phenomenon is a complex task, even for human classifications. The phenomena in the photosphere are so diverse that it is effectively easy to under-classify or over-classify morphologies. Thus, it is fundamental to improve the initial labeling for future supervised testing including the use of ground-truth segmentation methods that involve the least amount of user intervention in order to reduce ambiguity.</p>
<p>Another source of over-fitting may be related to the augmentation process, which is highly affected by the wrong-labeled data. In this case, the geometric transformations applied in the limited available labeled samples, especially for the underrepresented classes, can induce over-fitting (<xref ref-type="bibr" rid="B60">Shorten and Khoshgoftaar, 2019</xref>). Granules, as individual elements, are unique at a very detailed level, i.e., in super-high-resolution images. However, as we have already mentioned, they share similar phenomenologies that makes it possible to classify them into groups with comparable patterns at basic levels of similarity. Therefore, the use of extensive and random geometric transformations can produce non-deterministic effects, negatively affecting the training performance. Other strategies exist in the literature to prevent overfitting in skewed data, i.e. transfer learning (<xref ref-type="bibr" rid="B67">Weiss et al., 2016</xref>), pre-training (<xref ref-type="bibr" rid="B62">Singh Punn and Agarwal, 2021</xref>) or one-shot and zero-shot learning (<xref ref-type="bibr" rid="B68">Xian et al., 2017</xref>), which we plan to study in future works.</p>
<p>We extensively highlight these initial experiments as a starting point for further investigation. As this research is still under development, we seek to improve the levels of sensitivity and precision as much as possible to unequivocally detect the existing phenomenologies in solar granulation. We anticipate that extending our approach to include time-series, i.e., video segmentation, and other physical observables such as polarization and Doppler maps can be fruitful. This additional information would reveal other characteristics associated with the considered phenomena, allowing the definition of precise selection criteria, e.g., granular lane cases have been unambiguously detected based on the host granule evolution. Besides, the exploration of self-supervised or unsupervised methods is in our sights for further studies.</p>
</sec>
<sec id="s5">
<title>5 Resource Identification Initiative</title>
<p>All code of the model was constructed based on Python Programming Language, RRID:SCR_008394 version 3.9.7, and <italic>PyTorch</italic> libraries, RRID: SCR_018536 version 1.10.0.</p>
</sec>
</body>
<back>
<sec id="s6">
<title>Data Availability Statement</title>
<p>Publicly available datasets were analyzed in this study. This data can be found here: <ext-link ext-link-type="uri" xlink:href="https://star.mps.mpg.de/sunrise/">https://star.mps.mpg.de/sunrise/</ext-link> and <ext-link ext-link-type="uri" xlink:href="https://gitlab.leibniz-kis.de/smdiazcas/SegGranules_Unet_model.git">https://gitlab.leibniz-kis.de/smdiazcas/SegGranules_Unet_model.git</ext-link>.</p>
</sec>
<sec id="s7">
<title>Author Contributions</title>
<p>SD, CF, and AA conceived of the presented idea. SD and CF prepared the ground-truth dataset. SD and AA constructed the algorithm and the training strategy. SD executed the network training and analysed the results with help of AA All authors contributed to the final version of the manuscript. SB supervised the project.</p>
</sec>
<sec id="s8">
<title>Funding</title>
<p>SMDC and CEF are funded by the Leibniz Association grant for the SAW-2018-KIS-2-QUEST project. The German contribution to Sunrise is funded by the Bundesministerium f&#x00FC;r Wirtschaft und Technologie through Deutsches Zentrum f&#x00FC;r Luft und Raumfahrt e.V. (DLR), Grant No.50 OU 0401, and by the Innovationsfond of the President of the Max Planck Society (MPG). The Spanish contribution has been funded by the Spanish MICINN under projects ESP2006-13030-C06 and AYA2009-14105-C06 (including European FEDER funds). HAO/NCAR is sponsored by the National Science Foundation, and the HAO Contribution to Sunrise was partly funded through NASA grant number NNX08AH38G. The National Solar Observatory (NSO) is operated by the Association of Universities for Research in Astronomy, Inc. (AURA), under cooperative agreement with the National Science Foundation.</p>
</sec>
<sec sec-type="COI-statement" id="s9">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec sec-type="disclaimer" id="s10">
<title>Publisher&#x2019;s Note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<ack>
<p>We acknowledge the Instituto de Astrof&#xed;sica de Canarias (IAC) and the Leibniz-Institut f&#xfc;r Sonnenphysik (KIS) for providing us with compute time on their GPUs: NVIDIA GeForce 2080 Ti GPU and an NVIDIA Tesla P100 GPU respectively, resources extensively used in this research. This research has made use of NASA&#x2019;s Astrophysics Data System Bibliographic Services. We acknowledge the community effort devoted to the development of the following open-source packages that were used in this work: numpy (<ext-link ext-link-type="uri" xlink:href="http://numpy.org">numpy.org</ext-link>), matplotlib (<ext-link ext-link-type="uri" xlink:href="http://matplotlib.org">matplotlib.org</ext-link>) and Pytorch (<ext-link ext-link-type="uri" xlink:href="http://pytorch.org">pytorch.org</ext-link>).</p>
</ack>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>More information can be found at <ext-link ext-link-type="uri" xlink:href="https://lmb.informatik.uni-freiburg.de/people/ronneber/u-net/">https://lmb.informatik.uni-freiburg.de/people/ronneber/u-net/</ext-link>
</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>All code and documentation can be found at <ext-link ext-link-type="uri" xlink:href="http://wwwuser.gwdg.de/">http://wwwuser.gwdg.de/</ext-link>astronom/</p>
</fn>
<fn id="fn3">
<label>3</label>
<p>All codes are placed in a free access repository (<ext-link ext-link-type="uri" xlink:href="https://gitlab.leibniz-kis.de/smdiazcas/SegGranules_Unet_model.git">https://gitlab.leibniz-kis.de/smdiazcas/SegGranules_Unet_model.git</ext-link>)</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Aggarwal</surname>
<given-names>C. C.</given-names>
</name>
</person-group> (<year>2018</year>). <source>An Introduction to Neural Networks</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name>, <fpage>1</fpage>&#x2013;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1007/978-3-319-94463-0_1</pub-id> </citation>
</ref>
<ref id="B2">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Aloysius</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Geetha</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>A Review on Deep Convolutional Neural Networks</article-title>,&#x201d; in <conf-name>2017 International Conference on Communication and Signal Processing (ICCSP)</conf-name>, <conf-loc>Chennai, India</conf-loc>, <conf-date>April 6&#x2013;8, 2017</conf-date>, <fpage>0588</fpage>&#x2013;<lpage>0592</lpage>. <pub-id pub-id-type="doi">10.1109/ICCSP.2017.8286426</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Armstrong</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Fletcher</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Fast Solar Image Classification Using Deep Learning and its Importance for Automation in Solar Physics</article-title>. <source>Sol. Phys.</source> <volume>294</volume>, <fpage>80</fpage>. <pub-id pub-id-type="doi">10.1007/s11207-019-1473-z</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Badrinarayanan</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Kendall</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Cipolla</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Segnet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation</article-title>. <source>IEEE Trans. Pattern Analysis Mach. Intell.</source> <volume>39</volume>, <fpage>2481</fpage>&#x2013;<lpage>2495</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2016.2644615</pub-id> </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Baek</surname>
<given-names>J. H.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Choi</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kim</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Jo</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2021</year>). <article-title>Solar Event Detection Using Deep-Learning-Based Object Detection Methods</article-title>. <source>Sol. Phys.</source> <volume>296</volume>, <fpage>160</fpage>. <pub-id pub-id-type="doi">10.1007/s11207-021-01902-5</pub-id> </citation>
</ref>
<ref id="B72">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Barthol</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Gandorfer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Schussler</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chares</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Curdt</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>The Sunrise Mission</article-title>. <source>Solar Phys.</source> <volume>268</volume>, <fpage>1</fpage>&#x2013;<lpage>34</lpage>. <pub-id pub-id-type="doi">10.1007/s11207-010-9662-9</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bellot Rubio</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Orozco Su&#xe1;rez</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Quiet Sun Magnetic Fields: an Observational View</article-title>. <source>Living Rev. Sol. Phys.</source> <volume>16</volume>, <fpage>1</fpage>. <pub-id pub-id-type="doi">10.1007/s41116-018-0017-1</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bovelet</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wiehr</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2001</year>). <article-title>A New Algorithm for Pattern Recognition and its Application to Granulation and Limb Faculae</article-title>. <source>Sol. Phys.</source> <volume>201</volume>, <fpage>13</fpage>&#x2013;<lpage>26</lpage>. <pub-id pub-id-type="doi">10.1023/A:1010344827952</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bovelet</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Wiehr</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2007</year>). <article-title>Multiple-Scale Pattern Recognition Applied to Faint Intergranular G-Band Structures</article-title>. <source>Sol. Phys.</source> <volume>243</volume>, <fpage>121</fpage>&#x2013;<lpage>129</lpage>. <pub-id pub-id-type="doi">10.1007/s11207-007-9010-x</pub-id> </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Carlier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Chauveau</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Hugon</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>R&#xf6;sch</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>1968</year>). <article-title>Cin&#xe9;matographie &#xe0; Haute R&#xe9;solution Spatiale de la Granulation Photosph&#xe9;rique</article-title>. <source>Acad. Des. Sci. Paris Comptes Rendus Ser. B Sci. Phys.</source> <volume>266</volume>, <fpage>199</fpage>&#x2013;<lpage>201</lpage>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Papandreou</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Kokkinos</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Murphy</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Yuille</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deeplab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected Crfs</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>40</volume>, <fpage>834</fpage>&#x2013;<lpage>848</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2017.2699184</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chola</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Benifa</surname>
<given-names>J. V. B.</given-names>
</name>
</person-group> (<year>2022</year>). <article-title>Detection and Classification of Sunspots via Deep Convolutional Neural Network</article-title>. <source>Glob. Transitions Proc.</source> <pub-id pub-id-type="doi">10.1016/j.gltp.2022.03.006</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>De Pontieu</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2002</year>). <article-title>High&#x2010;Resolution Observations of Small&#x2010;Scale Emerging Flux in the Photosphere</article-title>. <source>Astrophys. J.</source> <volume>569</volume>, <fpage>474</fpage>&#x2013;<lpage>486</lpage>. <pub-id pub-id-type="doi">10.1086/339231</pub-id> </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dom&#xed;nguez Cerde&#xf1;a</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Evidence of Mesogranulation from Magnetograms of the Sun</article-title>. <source>Astron. Astrophys.</source> <volume>412</volume>, <fpage>L65</fpage>&#x2013;<lpage>L68</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361:20034617</pub-id> </citation>
</ref>
<ref id="B14">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Dumoulin</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Visin</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2016</year>). <source>A Guide to Convolution Arithmetic for Deep Learning</source>. <comment>arXiv:1603.07285 [Online]</comment>. </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ellwarth</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>C. E.</given-names>
</name>
<name>
<surname>Vitas</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Schmiz</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Schmidt</surname>
<given-names>W.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Newly Formed Downflow Lanes in Exploding Granules in the Solar Photosphere</article-title>. <source>Astron. Astrophys.</source> <volume>653</volume>, <fpage>A96</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/202038252</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Fern&#xe1;ndez</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Garc&#xed;a</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Galar</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Prati</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Krawczyk</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Herrera</surname>
<given-names>F.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>Foundations on Imbalanced Classification</article-title>,&#x201d; in <source>Learning from Imbalanced Data Sets</source> (<publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>). <pub-id pub-id-type="doi">10.1007/978-3-319-98074-4_2</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fischer</surname>
<given-names>C. E.</given-names>
</name>
<name>
<surname>Borrero</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Bello Gonz&#xe1;lez</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Kaithakkal</surname>
<given-names>A. J.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Observations of Solar Small-Scale Magnetic Flux-Sheet Emergence</article-title>. <source>Astron. Astrophys.</source> <volume>622</volume>, <fpage>L12</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201834628</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fischer</surname>
<given-names>C. E.</given-names>
</name>
<name>
<surname>Vigeesh</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Lindner</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Borrero</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Calvo</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Steiner</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Interaction of Magnetic Fields with a Vortex Tube at Solar Subgranular Scale</article-title>. <source>Astrophys. J. Lett.</source> <volume>903</volume>, <fpage>L10</fpage>. <pub-id pub-id-type="doi">10.3847/2041-8213/abbada</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ghosh</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sufian</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sultana</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Chakrabarti</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>De</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Fundamental Concepts of Convolutional Neural Network</article-title>,&#x201d; in <source>Recent Trends and Advances in Artificial Intelligence and Internet of Things, Intelligent Systems Reference Library</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Balas</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Kumar</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Srivastava</surname>
<given-names>R.</given-names>
</name>
</person-group> (<publisher-name>Springer</publisher-name>), <volume>Vol. 172</volume>. <pub-id pub-id-type="doi">10.1007/978-3-030-32644-9_36</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Girshick</surname>
<given-names>R. B.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Fast R-Cnn</article-title>,&#x201d; in <conf-name>2015 IEEE International Conference on Computer Vision (ICCV)</conf-name> <conf-loc>Santiago, Chile</conf-loc>, <conf-date>December 7&#x2013;13, 2015</conf-date>, <fpage>1440</fpage>&#x2013;<lpage>1448</lpage>. <pub-id pub-id-type="doi">10.1109/iccv.2015.169</pub-id> </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Guglielmino</surname>
<given-names>S. L.</given-names>
</name>
<name>
<surname>Mart&#xed;nez Pillet</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Ruiz Cobo</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bellot Rubio</surname>
<given-names>L. R.</given-names>
</name>
<name>
<surname>del Toro Iniesta</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>On the Magnetic Nature of an Exploding Granule as Revealed by Sunrise/IMaX</article-title>. <source>Astrophys. J.</source> <volume>896</volume>, <fpage>62</fpage>. <pub-id pub-id-type="doi">10.3847/1538-4357/ab917b</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Havaei</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Davy</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Biard</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Courville</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Brain Tumor Segmentation with Deep Neural Networks</article-title>. <source>Med. Image Anal.</source> <volume>35</volume>, <fpage>18</fpage>&#x2013;<lpage>31</lpage>. <pub-id pub-id-type="doi">10.1016/j.media.2016.05.004</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Garcia</surname>
<given-names>E. A.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Learning from Imbalanced Data</article-title>. <source>IEEE Trans. Knowl. Data Eng.</source> <volume>21</volume>, <fpage>1263</fpage>&#x2013;<lpage>1284</lpage>. <pub-id pub-id-type="doi">10.1109/TKDE.2008.239</pub-id> </citation>
</ref>
<ref id="B24">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Ma</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>2013</year>). <source>Imbalanced Learning: Foundations, Algorithms, and Applications</source>. <edition>1st edn</edition>. <publisher-loc>Hoboken, New Jersey</publisher-loc>: <publisher-name>Wiley-IEEE Press</publisher-name>. </citation>
</ref>
<ref id="B25">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Bai</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Garcia</surname>
<given-names>E. A.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2008</year>). &#x201c;<article-title>Adasyn: Adaptive Synthetic Sampling Approach for Imbalanced Learning</article-title>,&#x201d; in <conf-name>2008 IEEE International Joint Conference on Neural Networks</conf-name> <conf-loc>Hong Kong, China</conf-loc>, <conf-date>June 1&#x2013;8, 2008</conf-date> (<publisher-name>IEEE World Congress on Computational Intelligence</publisher-name>), <fpage>1322</fpage>&#x2013;<lpage>1328</lpage>. <pub-id pub-id-type="doi">10.1109/IJCNN.2008.4633969</pub-id> </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hirzberger</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Bonet</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>V&#xe1;zquez</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hanslmeier</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>1999</year>). <article-title>Time Series of Solar Granulation Images. III. Dynamics of Exploding Granules and Related Phenomena</article-title>. <source>Astrophys. J.</source> <volume>527</volume>, <fpage>405</fpage>&#x2013;<lpage>414</lpage>. <pub-id pub-id-type="doi">10.1086/308065</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Loy</surname>
<given-names>C. C.</given-names>
</name>
<name>
<surname>Tang</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Learning Deep Representation for Imbalanced Classification</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Las Vegas, NV</conf-loc>, <conf-date>June 27&#x2013;30, 2016</conf-date>. <pub-id pub-id-type="doi">10.1109/cvpr.2016.580</pub-id> </citation>
</ref>
<ref id="B28">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Huang</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Van Der Maaten</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Weinberger</surname>
<given-names>K. Q.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Densely Connected Convolutional Networks</article-title>,&#x201d; in <conf-name>2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Honolulu, HI</conf-loc>, <conf-date>July 21&#x2013;26, 2017</conf-date>, <fpage>2261</fpage>&#x2013;<lpage>2269</lpage>. <pub-id pub-id-type="doi">10.1109/CVPR.2017.243</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ilg</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Mayer</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Saikia</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Keuper</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Dosovitskiy</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Flownet 2.0: Evolution of Optical Flow Estimation with Deep Networks</article-title>,&#x201d; in <conf-name>IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, <conf-loc>Honolulu, HI</conf-loc>, <conf-date>July 21&#x2013;26, 2017</conf-date>. <pub-id pub-id-type="doi">10.1109/cvpr.2017.179</pub-id> </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Jaccard</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>1912</year>). <article-title>The Distribution of the Flora in the Alpine zone.1</article-title>. <source>New Phytol.</source> <volume>11</volume>, <fpage>37</fpage>&#x2013;<lpage>50</lpage>. <pub-id pub-id-type="doi">10.1111/j.1469-8137.1912.tb05611.x</pub-id> </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Javaherian</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Safari</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Amiri</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Ziaei</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Automatic Method for Identifying Photospheric Bright Points and Granules Observed by Sunrise</article-title>. <source>Sol. Phys.</source> <volume>289</volume>, <fpage>3969</fpage>&#x2013;<lpage>3983</lpage>. <pub-id pub-id-type="doi">10.1007/s11207-014-0555-1</pub-id> </citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kaithakkal</surname>
<given-names>A. J.</given-names>
</name>
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Cancelation of Small-Scale Magnetic Features</article-title>. <source>Astron. Astrophys.</source> <volume>622</volume>, <fpage>A200</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201833770</pub-id> </citation>
</ref>
<ref id="B33">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Khan</surname>
<given-names>S. H.</given-names>
</name>
<name>
<surname>Hayat</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bennamoun</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Sohel</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Togneri</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Cost Sensitive Learning of Deep Feature Representations from Imbalanced Data</article-title>,&#x201d; in <source>IEEE Transactions on Neural Networks and Learning Systems</source> <volume>vol. 29</volume>, <fpage>3573</fpage>&#x2013;<lpage>3587</lpage>. <pub-id pub-id-type="doi">10.1109/TNNLS.2017.2732482</pub-id> </citation>
</ref>
<ref id="B34">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Adam: A Method for Stochastic Optimization</article-title>,&#x201d; in <source>3rd International Conference on Learning Representations, ICLR 2015</source> <conf-loc>San Diego, CA</conf-loc>, <conf-date>May 7&#x2013;9, 2015</conf-date>. </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kitai</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Kawaguchi</surname>
<given-names>I.</given-names>
</name>
</person-group> (<year>1979</year>). <article-title>Morphological Study of the Solar Granulation</article-title>. <source>Sol. Phys.</source> <volume>64</volume>, <fpage>3</fpage>&#x2013;<lpage>12</lpage>. <pub-id pub-id-type="doi">10.1007/BF00151111</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Krizhevsky</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sutskever</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Hinton</surname>
<given-names>G. E.</given-names>
</name>
</person-group> (<year>2012</year>). &#x201c;<article-title>Imagenet Classification with Deep Convolutional Neural Networks</article-title>,&#x201d; in <conf-name>Proceedings of the 25th International Conference on Neural Information Processing Systems</conf-name> (<publisher-loc>Red Hook,NY, USA</publisher-loc>: <publisher-name>Curran Associates Inc.</publisher-name>), <fpage>1097</fpage>&#x2013;<lpage>1105</lpage>. <comment>NIPS&#x2019;12</comment>.<volume>Vol. 1.</volume> </citation>
</ref>
<ref id="B37">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Le Cun</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Bottou</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Bengio</surname>
<given-names>Y.</given-names>
</name>
</person-group> (<year>1997</year>). &#x201c;<article-title>Reading Checks with Multilayer Graph Transformer Networks</article-title>,&#x201d; in <conf-name>1997 IEEE International Conference on Acoustics, Speech, and Signal Processing</conf-name>, <conf-loc>Munich, Germany</conf-loc>, <conf-date>April 21&#x2013;24, 1997</conf-date>, <fpage>151</fpage>&#x2013;<lpage>154</lpage>. <pub-id pub-id-type="doi">10.1109/ICASSP.1997.599580</pub-id> </citation>
</ref>
<ref id="B38">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Leivas Oliveira</surname>
<given-names>G.</given-names>
</name>
</person-group> (<year>2019</year>). <source>Encoder-decoder Methods for Semantic Segmentation: Efficiency and Robustness Aspects</source>. <publisher-loc>Freiburg im Breisgau, Germany</publisher-loc>: <publisher-name>Technische Fakult&#xe4;t, Albert-Ludwigs-Universit&#xe4;t Freiburg</publisher-name>. <comment>Ph.D. thesis</comment>. <pub-id pub-id-type="doi">10.6094/UNIFR/150065</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lin</surname>
<given-names>T. Y.</given-names>
</name>
<name>
<surname>Goyal</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Girshick</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Doll&#xe1;r</surname>
<given-names>P.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Focal Loss for Dense Object Detection</article-title>,&#x201d; in <conf-name>2017 IEEE International Conference on Computer Vision (ICCV)</conf-name>, <conf-loc>Venice, Italy</conf-loc>, <conf-date>October 22&#x2013;29, 2017</conf-date>, <fpage>2999</fpage>&#x2013;<lpage>3007</lpage>. <pub-id pub-id-type="doi">10.1109/ICCV.2017.324</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Love</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Neukirch</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Parnell</surname>
<given-names>C. E.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Analyzing Aia Flare Observations Using Convolutional Neural Networks</article-title>. <source>Front. Astron. Space Sci.</source> <volume>7</volume>. <pub-id pub-id-type="doi">10.3389/fspas.2020.00034</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Malherbe</surname>
<given-names>J.-M.</given-names>
</name>
<name>
<surname>Roudier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Stein</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Frank</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Dynamics of Trees of Fragmenting Granules in the Quiet Sun: Hinode/SOT Observations Compared to Numerical Simulation</article-title>. <source>Sol. Phys.</source> <volume>293</volume>, <fpage>4</fpage>. <pub-id pub-id-type="doi">10.1007/s11207-017-1225-x</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mart&#xed;nez Pillet</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Del Toro Iniesta</surname>
<given-names>J. C.</given-names>
</name>
<name>
<surname>&#xc1;lvarez-Herrero</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Domingo</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Bonet</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Gonz&#xe1;lez Fern&#xe1;ndez</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2011</year>). <article-title>The Imaging Magnetograph eXperiment (IMaX) for the Sunrise Balloon-Borne Solar Observatory</article-title>. <source>Sol. Phys.</source> <volume>268</volume>, <fpage>57</fpage>&#x2013;<lpage>102</lpage>. <pub-id pub-id-type="doi">10.1007/s11207-010-9644-y</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Muller</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Roudier</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>1984</year>). <article-title>Variability of the Quiet Photospheric Network</article-title>. <source>Sol. Phys.</source> <volume>94</volume>, <fpage>33</fpage>&#x2013;<lpage>47</lpage>. <pub-id pub-id-type="doi">10.1007/BF00154805</pub-id> </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Namba</surname>
<given-names>O.</given-names>
</name>
</person-group> (<year>1986</year>). <article-title>Evolution of &#x2019;exploding Granules</article-title>. <source>Astron. Astrophys.</source> <volume>161</volume>, <fpage>31</fpage>&#x2013;<lpage>38</lpage>. </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nordlund</surname>
<given-names>&#xc5;.</given-names>
</name>
<name>
<surname>Stein</surname>
<given-names>R. F.</given-names>
</name>
<name>
<surname>Asplund</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Solar Surface Convection</article-title>. <source>Living Rev. Sol. Phys.</source> <volume>6</volume>, <fpage>2</fpage>. <pub-id pub-id-type="doi">10.12942/lrsp-2009-2</pub-id> </citation>
</ref>
<ref id="B46">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Oksuz</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Cam</surname>
<given-names>B. C.</given-names>
</name>
<name>
<surname>Kalkan</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Akbas</surname>
<given-names>E.</given-names>
</name>
</person-group> (<year>2020</year>). &#x201c;<article-title>Imbalance Problems in Object Detection: A Review</article-title>,&#x201d; in <conf-name>IEEE Transactions on Pattern Analysis and Machine Intelligence</conf-name>, <volume>43</volume>, <fpage>3388</fpage>&#x2013;<lpage>3415</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2020.2981890</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Ol&#xe3; Bressan</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Marcato Junior</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Correa Martins</surname>
<given-names>J. A.</given-names>
</name>
<name>
<surname>Nunes Gon&#xe7;alves</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Matte Freitas</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Prado Osco</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2022</year>). <article-title>Semantic Segmentation With Labeling Uncertainty and Class Imbalance Applied to Vegetation Mapping</article-title>. <source>Internat. J. Appl. Earth Obs. Geoinf.</source> <volume>108</volume>, <fpage>102690</fpage>. <pub-id pub-id-type="doi">10.1016/j.jag.2022.102690</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Palacios</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Blanco Rodr&#xed;guez</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Vargas Dom&#xed;nguez</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Domingo</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Mart&#xed;nez Pillet</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Bonet</surname>
<given-names>J. A.</given-names>
</name>
<etal/>
</person-group> (<year>2012</year>). <article-title>Magnetic Field Emergence in Mesogranular-Sized Exploding Granules Observed with sunrise/IMaX Data</article-title>. <source>Astron. Astrophys.</source> <volume>537</volume>, <fpage>A21</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201117936</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Paszke</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gross</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Massa</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Lerer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Bradbury</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Chanan</surname>
<given-names>G.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). &#x201c;<article-title>Pytorch: An Imperative Style, High-Performance Deep Learning Library</article-title>,&#x201d; in <source>Advances in Neural Information Processing Systems</source>. Editors <person-group person-group-type="editor">
<name>
<surname>Wallach</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Larochelle</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Beygelzimer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>d&#x2019; Alch&#xe9;-Buc</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Fox</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Garnett</surname>
<given-names>R.</given-names>
</name>
</person-group> (<publisher-name>Curran Associates, Inc.</publisher-name>), <volume>32</volume>, <fpage>8024</fpage>&#x2013;<lpage>8035</lpage>. </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rempel</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Small-scale Dynamo Simulations: Magnetic Field Amplification in Exploding Granules and the Role of Deep and Shallow Recirculation</article-title>. <source>Astrophys. J.</source> <volume>859</volume>, <fpage>161</fpage>. <pub-id pub-id-type="doi">10.3847/1538-4357/aabba0</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Riethm&#xfc;ller</surname>
<given-names>T. L.</given-names>
</name>
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Zakharov</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Gandorfer</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2008</year>). <article-title>Brightness, Distribution, and Evolution of Sunspot Umbral Dots</article-title>. <source>Astron. Astrophys.</source> <volume>492</volume>, <fpage>233</fpage>&#x2013;<lpage>243</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361:200810701</pub-id> </citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rimmele</surname>
<given-names>T. R.</given-names>
</name>
<name>
<surname>Warner</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Keil</surname>
<given-names>S. L.</given-names>
</name>
<name>
<surname>Goode</surname>
<given-names>P. R.</given-names>
</name>
<name>
<surname>Kn&#xf6;lker</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kuhn</surname>
<given-names>J. R.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>The Daniel K. Inouye Solar Telescope - Observatory Overview</article-title>. <source>Sol. Phys.</source> <volume>295</volume>, <fpage>172</fpage>. <pub-id pub-id-type="doi">10.1007/s11207-020-01736-7</pub-id> </citation>
</ref>
<ref id="B53">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Ronneberger</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Fischer</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Brox</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>U-Net: Convolutional Networks for Biomedical Image Segmentation</article-title>,&#x201d; in <conf-name>Medical Image Computing and Computer-Assisted Intervention &#x2013; MICCAI 2015. MICCAI 2015 Lecture Notes in Computer Science</conf-name>. Editors <person-group person-group-type="editor">
<name>
<surname>Navab</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Hornegger</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wells</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Frangi</surname>
<given-names>A.</given-names>
</name>
</person-group> <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer</publisher-name>. <pub-id pub-id-type="doi">10.1007/978-3-319-24574-4_28</pub-id> </citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roudier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Muller</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2004</year>). <article-title>Relation between Families of Granules, Mesogranules and Photospheric Network</article-title>. <source>Astron. Astrophys.</source> <volume>419</volume>, <fpage>757</fpage>&#x2013;<lpage>762</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361:20035739</pub-id> </citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roudier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ligni&#xe8;res</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Rieutord</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Brandt</surname>
<given-names>P. N.</given-names>
</name>
<name>
<surname>Malherbe</surname>
<given-names>J. M.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Families of Fragmenting Granules and Their Relation to Meso- and Supergranular Flow Fields</article-title>. <source>Astron. Astrophys.</source> <volume>409</volume>, <fpage>299</fpage>&#x2013;<lpage>308</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361:20030988</pub-id> </citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roudier</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Malherbe</surname>
<given-names>J. M.</given-names>
</name>
<name>
<surname>Rieutord</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Frank</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Relation between Trees of Fragmenting Granules and Supergranulation Evolution</article-title>. <source>Astron. Astrophys.</source> <volume>590</volume>, <fpage>A121</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201628111</pub-id> </citation>
</ref>
<ref id="B57">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Schmidhuber</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Deep Learning in Neural Networks: An Overview</article-title>. <source>Neural Networks</source> <volume>61</volume>, <fpage>85</fpage>&#x2013;<lpage>117</lpage>. <pub-id pub-id-type="doi">10.1016/j.neunet.2014.09.003</pub-id> </citation>
</ref>
<ref id="B58">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Sebe</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Cohen</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Garg</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Huang</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2005</year>). <source>Machine Learning in Computer Vision</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name>. </citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shelhamer</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Long</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Darrell</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Fully Convolutional Networks for Semantic Segmentation</article-title>. <source>IEEE Trans. Pattern Anal. Mach. Intell.</source> <volume>39</volume>, <fpage>640</fpage>&#x2013;<lpage>651</lpage>. <pub-id pub-id-type="doi">10.1109/tpami.2016.2572683</pub-id> </citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Shorten</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Khoshgoftaar</surname>
<given-names>T. M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>A Survey on Image Data Augmentation for Deep Learning</article-title>. <source>J. Big Data</source> <volume>6</volume>, <fpage>60</fpage>. <pub-id pub-id-type="doi">10.1186/s40537-019-0197-0</pub-id> </citation>
</ref>
<ref id="B61">
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Simonyan</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Zisserman</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2015</year>). &#x201c;<article-title>Very Deep Convolutional Networks for Large-Scale Image Recognition</article-title>,&#x201d; in <conf-name>3rd International Conference on Learning Representations, ICLR 2015</conf-name>, <conf-loc>San Diego, CA</conf-loc>, <conf-date>May 7&#x2013;9, 2015 </conf-date>. </citation>
</ref>
<ref id="B62">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Singh Punn</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Agarwal</surname>
<given-names>S.</given-names>
</name>
</person-group>(<year>2021</year>). <source>BT-Unet: A Self-Supervised Learning Framework for Biomedical Image Segmentation Using Barlow Twins with U-Net Models</source>. <comment>arXiv e-prints</comment>. <comment>arXiv:2112.03916</comment>. </citation>
</ref>
<ref id="B71">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Barthol</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Danilovic</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Feller</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Gandorfer</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hirzberger</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Sunrise: Instrument, Mission, Data, and First Results</article-title>. <source>Geophysical Monograph Series</source> <volume>723</volume>, <fpage>L127</fpage>&#x2013;<lpage>L133</lpage>. <pub-id pub-id-type="doi">10.1088/2041-8205/723/2/l127</pub-id> </citation>
</ref>
<ref id="B63">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Solanki</surname>
<given-names>S. K.</given-names>
</name>
<name>
<surname>Riethm&#xfc;ller</surname>
<given-names>T. L.</given-names>
</name>
<name>
<surname>Barthol</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Danilovic</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Deutsch</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Doerr</surname>
<given-names>H. P.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>The Second Flight of the Sunrise Balloon-Borne Solar Observatory: Overview of Instrument Updates, the Flight, the Data, and First Results</article-title>. <source>Astrophys. J. Suppl. Ser.</source> <volume>229</volume>, <fpage>2</fpage>. <pub-id pub-id-type="doi">10.3847/1538-4365/229/1/2</pub-id> </citation>
</ref>
<ref id="B64">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Steiner</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Franz</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bello Gonz&#xe1;lez</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Nutto</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Rezaei</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Mart&#xed;nez Pillet</surname>
<given-names>V.</given-names>
</name>
<etal/>
</person-group> (<year>2010</year>). <article-title>Detection of Vortex Tubes in Solar Granulation from Observations with SUNRISE</article-title>. <source>Astrophys. J. Lett.</source> <volume>723</volume>, <fpage>L180</fpage>&#x2013;<lpage>L184</lpage>. <pub-id pub-id-type="doi">10.1088/2041-8205/723/2/L180</pub-id> </citation>
</ref>
<ref id="B65">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Stix</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2002</year>). <source>The Sun: An Introduction</source>. <publisher-loc>Germany</publisher-loc>: <publisher-name>Springer</publisher-name>. </citation>
</ref>
<ref id="B66">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Szeliski</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2011</year>). <source>Computer Vision: Algorithms and Applications</source>. <publisher-loc>New York</publisher-loc>: <publisher-name>Springer</publisher-name>. </citation>
</ref>
<ref id="B67">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Weiss</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Khoshgoftaar</surname>
<given-names>T. M.</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>D.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>A Survey of Transfer Learning</article-title>. <source>J. Big Data</source> <volume>3</volume>, <fpage>9</fpage>. <pub-id pub-id-type="doi">10.1186/s40537-016-0043-6</pub-id> </citation>
</ref>
<ref id="B68">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Xian</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Lampert</surname>
<given-names>C. H.</given-names>
</name>
<name>
<surname>Schiele</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Akata</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Zero-Shot Learning &#x2013; A Comprehensive Evaluation of the Good, the Bad and the Ugly</article-title>,&#x201d; in <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>, <volume>41</volume>, <fpage>2251</fpage>&#x2013;<lpage>2265</lpage>. <pub-id pub-id-type="doi">10.1109/TPAMI.2018.2857768</pub-id> </citation>
</ref>
<ref id="B69">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yanli</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Pengpeng</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>J-net: Asymmetric Encoder-Decoder for Medical Semantic Segmentation</article-title>. <source>Secur. Commun. Netw.</source> <volume>2021</volume>, <fpage>2139024</fpage>. <pub-id pub-id-type="doi">10.1155/2021/2139024</pub-id> </citation>
</ref>
<ref id="B70">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zou</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Weinacker</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Koch</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Towards Urban Scene Semantic Segmentation with Deep Learning from Lidar Point Clouds: A Case Study in Baden-W&#xfc;rttemberg, germany</article-title>. <source>Remote Sens.</source> <volume>13</volume>. <pub-id pub-id-type="doi">10.3390/rs13163220</pub-id> </citation>
</ref>
</ref-list>
</back>
</article>