<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article article-type="research-article" dtd-version="2.3" xml:lang="EN" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Artif. Intell.</journal-id>
<journal-title>Frontiers in Artificial Intelligence</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Artif. Intell.</abbrev-journal-title>
<issn pub-type="epub">2624-8212</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="publisher-id">673062</article-id>
<article-id pub-id-type="doi">10.3389/frai.2021.673062</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Artificial Intelligence</subject>
<subj-group>
<subject>Original Research</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Emulation of Cosmological Mass Maps with Conditional Generative Adversarial Networks</article-title>
<alt-title alt-title-type="left-running-head">Perraudin et&#x20;al.</alt-title>
<alt-title alt-title-type="right-running-head">GAN Emulation of Mass Maps</alt-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Perraudin</surname>
<given-names>Nathana&#xeb;l</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1301722/overview"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Marcon</surname>
<given-names>Sandro</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Lucchi</surname>
<given-names>Aurelien</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1325192/overview"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Kacprzak</surname>
<given-names>Tomasz</given-names>
</name>
<xref ref-type="aff" rid="aff3">
<sup>3</sup>
</xref>
<xref ref-type="corresp" rid="c001">&#x2a;</xref>
<uri xlink:href="https://loop.frontiersin.org/people/1248687/overview"/>
</contrib>
</contrib-group>
<aff id="aff1">
<label>
<sup>1</sup>
</label>Swiss Data Science Center, ETH Zurich, <addr-line>Zurich</addr-line>, <country>Switzerland</country>
</aff>
<aff id="aff2">
<label>
<sup>2</sup>
</label>Institute for Machine Learning, ETH Zurich, <addr-line>Zurich</addr-line>, <country>Switzerland</country>
</aff>
<aff id="aff3">
<label>
<sup>3</sup>
</label>Institute for Particle Physics and Astrophysics, ETH Zurich, <addr-line>Zurich</addr-line>, <country>Switzerland</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>
<bold>Edited by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/923100/overview">Maria Han Veiga</ext-link>, University of Michigan, United&#x20;States</p>
</fn>
<fn fn-type="edited-by">
<p>
<bold>Reviewed by:</bold> <ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/570108/overview">Bowei Chen</ext-link>, University of Glasgow, United&#x20;Kingdom</p>
<p>
<ext-link ext-link-type="uri" xlink:href="https://loop.frontiersin.org/people/172729/overview">Riccardo Zese</ext-link>, University of Ferrara, Italy</p>
</fn>
<corresp id="c001">&#x2a;Correspondence: Tomasz Kacprzak, <email>tomaszk@phys.ethz.ch</email>
</corresp>
<fn fn-type="other">
<p>This article was submitted to Machine Learning and Artificial Intelligence, a section of the journal Frontiers in Artificial Intelligence</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>04</day>
<month>06</month>
<year>2021</year>
</pub-date>
<pub-date pub-type="collection">
<year>2021</year>
</pub-date>
<volume>4</volume>
<elocation-id>673062</elocation-id>
<history>
<date date-type="received">
<day>26</day>
<month>02</month>
<year>2021</year>
</date>
<date date-type="accepted">
<day>06</day>
<month>05</month>
<year>2021</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2021 Perraudin, Marcon, Lucchi and Kacprzak.</copyright-statement>
<copyright-year>2021</copyright-year>
<copyright-holder>Perraudin, Marcon, Lucchi and Kacprzak</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these&#x20;terms.</p>
</license>
</permissions>
<abstract>
<p>Weak gravitational lensing mass maps play a crucial role in understanding the evolution of structures in the Universe and our ability to constrain cosmological models. The prediction of these mass maps is based on expensive N-body simulations, which can create a computational bottleneck for cosmological analyses. Simulation-based emulators of map summary statistics, such as the matter power spectrum and its covariance, are starting to play increasingly important role, as the analytical predictions are expected to reach their precision limits for upcoming experiments. Creating an emulator of the cosmological mass maps themselves, rather than their summary statistics, is a more challenging task. Modern deep generative models, such as Generative Adversarial Networks (GAN), have demonstrated their potential to achieve this goal. Most existing GAN approaches produce simulations for a fixed value of the cosmological parameters, which limits their practical applicability. We propose a novel conditional GAN model that is able to generate mass maps for any pair of matter density &#x3a9;<sub>
<italic>m</italic>
</sub> and matter clustering strength <italic>&#x3c3;</italic>
<sub>8</sub>, parameters which have the largest impact on the evolution of structures in the Universe, for a given source galaxy redshift distribution <italic>n</italic>(<italic>z</italic>). Our results show that our conditional GAN can interpolate efficiently within the space of simulated cosmologies, and generate maps anywhere inside this space with good visual quality high statistical accuracy. We perform an extensive quantitative comparison of the N-body and GAN -generated maps using a range of metrics: the pixel histograms, peak counts, power spectra, bispectra, Minkowski functionals, correlation matrices of the power spectra, the Multi-Scale Structural Similarity Index (MS-SSIM) and our equivalent of the Fr&#xe9;chet Inception Distance. We find a very good agreement on these metrics, with typical differences are &#x3c;5% at the center of the simulation grid, and slightly worse for cosmologies at the grid edges. The agreement for the bispectrum is slightly worse, on the &#x3c;20% level. This contribution is a step toward building emulators of mass maps directly, capturing both the cosmological signal and its variability. We make the code<xref ref-type="fn" rid="fn1">
<sup>1</sup>
</xref> and the data<xref ref-type="fn" rid="fn2">
<sup>2</sup>
</xref> publicly available.</p>
</abstract>
<kwd-group>
<kwd>generative models</kwd>
<kwd>cosmological simulations</kwd>
<kwd>cosmological emulators</kwd>
<kwd>N-body simulations</kwd>
<kwd>mass maps</kwd>
<kwd>generative adversarial network</kwd>
<kwd>fast cosmic web simulations</kwd>
<kwd>conditional GAN</kwd>
</kwd-group>
<contract-num rid="cn001">Swiss Data Science Center Grant: &#x201c;Deep Learning for Observational Cosmology&#x201d;</contract-num>
<contract-sponsor id="cn001">Eidgen&#xf6;ssische Technische Hochschule Z&#xfc;rich<named-content content-type="fundref-id">10.13039/501100003006</named-content>
</contract-sponsor>
<contract-sponsor id="cn002">Schweizerischer Nationalfonds zur F&#xf6;rderung der Wissenschaftlichen Forschung<named-content content-type="fundref-id">10.13039/501100001711</named-content>
</contract-sponsor>
</article-meta>
</front>
<body>
<sec id="s1">
<title>1 Introduction</title>
<p>The N-body technique simulates the evolution of the Universe from soon after the big bang, where the mass distribution was approximately a Gaussian random field, to today, where, under the action of gravity, it becomes highly non-Gaussian. The result of an N-body simulation consists of a 3D volume where the positions of particles represent the density of matter in specific regions. This 3-dimensional representation can then be projected in 2 dimensions by integrating the mass along the line of sight with a lensing kernel. The resulting images are called <italic>sky convergence maps</italic>, often referred to simply as the <italic>cosmological mass maps</italic>. These maps can be compared with real observations with the purpose of estimating the cosmological parameters and testing cosmological models. Their simulation, however, is a very challenging task: a single large N-body simulation can take from a few hours to several weeks on a supercomputer (<xref ref-type="bibr" rid="B51">Springel et&#x20;al., 2005</xref>; <xref ref-type="bibr" rid="B42">Potter et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B4">Collaboration et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B50">Sgier et&#x20;al., 2019</xref>).</p>
<p>One approach to overcome this challenge is to use simulation-based emulators of summary statistics of the maps. Emulators have so far focused on: (a) the power spectrum, which is commonly used in cosmology (<xref ref-type="bibr" rid="B24">Knabenhans et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B18">Heitmann et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B25">Knabenhans et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B1">Angulo et&#x20;al., 2020</xref>), (b) covariance matrices of 2-pt functions (<xref ref-type="bibr" rid="B50">Sgier et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B55">Taylor et&#x20;al., 2013</xref>; <xref ref-type="bibr" rid="B48">Sato et&#x20;al., 2011</xref>), and (c) non-Gaussian statistics of mass maps, which can be a source of significant additional cosmological information (<xref ref-type="bibr" rid="B41">Pires et&#x20;al., 2009</xref>; <xref ref-type="bibr" rid="B38">Petri et&#x20;al., 2013</xref>; <xref ref-type="bibr" rid="B60">Z&#xfc;rcher et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B8">Fluri et&#x20;al., 2018</xref>). These approaches, however, always considered a specific summary statistic, which limits the type of analysis that can be performed using the mass-map data. They typically do not simultaneously capture both the signal and its variation: the emulators interpolate the power spectrum across the cosmological parameter space, without considering the change in its covariance matrix, which is typically taken from the fiducial cosmology parameter set. This is a known source of potential error in the analysis (<xref ref-type="bibr" rid="B7">Eifler et&#x20;al., 2009</xref>) and was shown to have a large impact on the deep learning-based constraints (<xref ref-type="bibr" rid="B8">Fluri et&#x20;al., 2018</xref>). The solution proposed in this work address these problems simultaneously. We construct a map-level probabilistic emulator that generates the mass maps directly, and can accurately capture the signal and its variability. This emulator, built for a specific target survey dataset, would be of great practical use for innovative map-based cosmological analyses, additionally capturing the variation of the maps across the cosmological parameter&#x20;space.</p>
<p>With a similar goal, multiple contributions have leveraged the recent advances in the field of deep learning to aid the generation of cosmological simulations. In particular, recent works (<xref ref-type="bibr" rid="B32">Mustafa et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B45">Rodriguez et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B33">Nathana&#xeb;l et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B56">Tr&#xf6;ster et&#x20;al., 2019</xref>) have demonstrated the potential of Generative Adversarial Networks (GAN) (<xref ref-type="bibr" rid="B14">Goodfellow et&#x20;al., 2014</xref>) for production of N-body simulations. The work of (<xref ref-type="bibr" rid="B32">Mustafa et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B45">Rodriguez et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B33">Nathana&#xeb;l et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B56">Tr&#xf6;ster et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B13">Giusarma et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B17">He et&#x20;al., 2019</xref>) has shown deep generative models that can accurately model dark matter distributions and other related cosmological signals. However, a practical application of these approaches in an end-to-end cosmological analysis is yet to be demonstrated. In this work, we take an essential step toward the practical use of generative models by creating the first emulator of weak lensing mass maps as function of cosmological parameters. This step allows the generate mass maps with any parameters without the need to retrain the generative model. Our conditional GAN model generates convergence maps dependent on values of two parameters that have the largest impact on the evolution of the Large Scale Structure (LSS) of the Universe: &#x3a9;<sub>
<italic>m</italic>
</sub>, which controls the matter density as a fraction of total density, and <italic>&#x3c3;</italic>8, which controls the strength of matter density fluctuations (see (<xref ref-type="bibr" rid="B44">Refregier, 2003</xref>; <xref ref-type="bibr" rid="B22">Kilbinger, 2015</xref>) for reviews). Those are the only two parameters that can be effectively measured using the convergence maps data. After training, the conditional model can then interpolate to unseen values of <italic>&#x3c3;</italic>
<sub>8</sub> and &#x3a9;<sub>
<italic>m</italic>
</sub> by varying the distribution of the input latent variable. Other works (<xref ref-type="bibr" rid="B54">Tamosiunas et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B58">Villaescusa-Navarro et&#x20;al., 2020</xref>) have since also explored such models, although with the emphasis on generating various cosmological fields themselves, either in <sup>2</sup>D or&#x20;3D.</p>
<p>To assess that the GAN-generated maps are statistically very close to the originals, we perform an extensive quantitative comparison. We evaluate our GAN using both cosmological and image processing metrics: the power spectral density, mass map histogram, peak histogram, the bispectrum, Minkowski functionals, Multi-Scale Structural Similarity (MS-SSIM) (<xref ref-type="bibr" rid="B59">Wang et&#x20;al., 2003</xref>), and an adaptation of the Fr&#xe9;chet Inception Distance (FID) (<xref ref-type="bibr" rid="B19">Heusel et&#x20;al., 2017</xref>). We also compare the statistical consistency of a batch of generated maps by computing the correlation matrices of power spectra. Moreover, we assess the agreement as a function of cosmological parameters. This set of comparisons is the most exhaustive presentation of the capacity of generative models to learn the dark matter maps, to date. In this work we use the data generated by (<xref ref-type="bibr" rid="B9">Fluri et&#x20;al., 2019</xref>).</p>
<p>We build a sky convergence map dataset made of 57 different cosmologies (set of parameters) divided into a training set and a test set. The test set consists of 11 cosmological parameters sets was used to asses the capacity of the GAN to interpolate to unseen cosmologies.</p>
<p>This paper is structured as follows. In <xref ref-type="sec" rid="s2">Section 2</xref> we present a new type of generative adversarial network whose generated output can be conditioned on a set of parameters in the form of continuous values. <xref ref-type="sec" rid="s3">Section 3</xref> describes the simulation dataset used in this work. In <xref ref-type="sec" rid="s4">Section 4</xref> we describe the metrics used to evaluate the quality of the generative model. <xref ref-type="sec" rid="s5">Section 5</xref> shows the maps generated by our machine learning model, as well as compares its results to the original, simulated data. We summarize our findings and discuss the future prospects in <xref ref-type="sec" rid="s6">Section 6</xref>. <xref ref-type="app" rid="app1">
<bold>Appendix A</bold>
</xref> contains the architectures of the neural networks used in this&#x20;work.</p>
</sec>
<sec id="s2">
<title>2 Conditional Generative Adversarial Networks</title>
<p>A GAN consists of two neural networks, <italic>D</italic> and <italic>G</italic>, competing against each other in a zero-sum game. The task of the <italic>discriminator D</italic> is to distinguish real (training) data from fake (generated) data. Meanwhile, the <italic>generator G</italic> produces samples with the goal of deceiving the discriminator into believing that the generated data is real. Both networks are trained simultaneously and if the optimization process is carried out successfully, the generator will learn to produce the data distribution (<xref ref-type="bibr" rid="B14">Goodfellow et&#x20;al., 2014</xref>). Learning the optimal parameters of the discriminator and generator networks can be formulated as optimizing a min-max objective. Optimizing a GAN is a challenging task due to the fact that it consists of two networks competing against each other. In practice, one often observes unstable training behaviors which can be mitigated by relying on various types of regularization methods (<xref ref-type="bibr" rid="B46">Roth et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B15">Gulrajani et&#x20;al., 2017</xref>). In this paper, we rely on Wasserstein GANs (<xref ref-type="bibr" rid="B2">Arjovsky et&#x20;al., 2017</xref>) with the regularization approach suggested in (<xref ref-type="bibr" rid="B15">Gulrajani et&#x20;al., 2017</xref>). The model we use conditions both the generator and the discriminator on a given random variable <italic>y</italic>, yielding the following objective function,<disp-formula id="e1">
<mml:math id="m1">
<mml:mrow>
<mml:munder>
<mml:mrow>
<mml:mi>min</mml:mi>
</mml:mrow>
<mml:mi>G</mml:mi>
</mml:munder>
<mml:munder>
<mml:mrow>
<mml:mi>max</mml:mi>
</mml:mrow>
<mml:mi>D</mml:mi>
</mml:munder>
<mml:munder>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:munder>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>z</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3bb;</mml:mi>
<mml:munder>
<mml:mi mathvariant="double-struck">E</mml:mi>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
<mml:mo>&#x222a;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>g</mml:mi>
</mml:msub>
</mml:mrow>
</mml:munder>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mo>&#x2207;</mml:mo>
<mml:mi mathvariant="bold-italic">x</mml:mi>
</mml:msub>
<mml:mi>D</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
<label>(1)</label>
</disp-formula>where <inline-formula id="inf1">
<mml:math id="m2">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf2">
<mml:math id="m3">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>z</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are the data and latent variable distributions. The parameter <inline-formula id="inf3">
<mml:math id="m4">
<mml:mrow>
<mml:mi>&#x3bb;</mml:mi>
<mml:mo>&#x2265;</mml:mo>
<mml:mn>0</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> is the penalty coefficient of the regularization term that ensures that the gradient norm of the discriminator is close to 1. This ensures that the discriminator is 1-Lipschitz, which is a requirement for optimizing the Wasserstein distance (<xref ref-type="bibr" rid="B15">Gulrajani et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B2">Arjovsky et&#x20;al., 2017</xref>). The prior distribution of the latent variable, e.g., a uniform or a Gaussian distribution, defines implicitly the generator distribution <inline-formula id="inf4">
<mml:math id="m5">
<mml:mrow>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>g</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> by <inline-formula id="inf5">
<mml:math id="m6">
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">x</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>G</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">z</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>z</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:mi mathvariant="bold-italic">y</mml:mi>
<mml:mo>&#x223c;</mml:mo>
<mml:msub>
<mml:mi mathvariant="normal">&#x2119;</mml:mi>
<mml:mi>y</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>Practically, there exist many techniques and architectures to condition the generator and the discriminator (<xref ref-type="bibr" rid="B12">Gauthier, 2014</xref>; <xref ref-type="bibr" rid="B36">Perarnau et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B43">Reed et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B34">Odena et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B31">Miyato and Koyama, 2018</xref>). However, all the architectures in these works are conditioning on discrete parameters. We instead propose a different design that works specifically for continuous parameters and will be shown to have good performance in practice. We note that our conditioning technique could be used with other architectures as well. For simplicity we describe the case of a single parameter, but our technique was implemented for the case of two parameters. Our idea is to adapt the distribution of the latent vector according to the conditioning parameters using the function <inline-formula id="inf6">
<mml:math id="m7">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. Specifically, the function <italic>f</italic> simply rescales the norm of the latent vector according to the parameter <italic>y</italic>. Given the range <inline-formula id="inf7">
<mml:math id="m8">
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mi>a</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>b</mml:mi>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>, <italic>f</italic> reads:<disp-formula id="e2">
<mml:math id="m9">
<mml:mrow>
<mml:mrow>
<mml:mover accent="true">
<mml:mi>z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>z</mml:mi>
<mml:mo>,</mml:mo>
<mml:mi>y</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x2b;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
</mml:mfrac>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>y</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>a</mml:mi>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
<mml:mfrac>
<mml:mi>z</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mi>z</mml:mi>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(2)</label>
</disp-formula>Using this function, the length of the <italic>z</italic> vector is mapped to the interval <inline-formula id="inf8">
<mml:math id="m10">
<mml:mrow>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. In our case, we used <inline-formula id="inf9">
<mml:math id="m11">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.1</mml:mn>
<mml:msqrt>
<mml:mi>n</mml:mi>
</mml:msqrt>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf10">
<mml:math id="m12">
<mml:mrow>
<mml:msub>
<mml:mi>l</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:msqrt>
<mml:mi>n</mml:mi>
</mml:msqrt>
</mml:mrow>
</mml:math>
</inline-formula>, where <italic>n</italic> is the size of the latent vector. For the discriminator, the parameters are concatenated directly after the convolutional layers as in (<xref ref-type="bibr" rid="B43">Reed et&#x20;al., 2016</xref>). The relation between the features extracted from the convolutional layers and the parameters might in general be non-local. We therefore increase the complexity of the mapping functions of the discriminator and generator by adding some linear layers (as in a multi-layer perceptron) at the end of the discriminator and the beginning of the generator. The proposed model is sketched in <xref ref-type="fig" rid="F1">Figure&#x20;1</xref> and the architecture is described in more details in <xref ref-type="app" rid="app1">
<bold>Appendix A</bold>
</xref>. Specific parameters can be found in <xref ref-type="table" rid="T1">Table A1</xref>.</p>
<fig id="F1" position="float">
<label>FIGURE 1</label>
<caption>
<p>Sketch of the proposed GAN model, where <italic>z</italic> is a latent variable and <italic>y</italic> is a parameter vector (<italic>y</italic>
<sub>
<italic>r</italic>
</sub> &#x3d; real, <italic>y</italic>
<sub>
<italic>f</italic>
</sub> &#x3d; fake).</p>
</caption>
<graphic xlink:href="frai-04-673062-g001.tif"/>
</fig>
</sec>
<sec id="s3">
<title>3 Sky Convergence Maps Dataset</title>
<p>The data used in this work is the non-tomographic training and testing set introduced in (<xref ref-type="bibr" rid="B9">Fluri et&#x20;al., 2019</xref>), without noise and intrinsic alignments. The simulation grid consists of 57 different cosmologies in the standard cosmological model: a flat Universe with cold dark matter (&#x39b;CDM) (<xref ref-type="bibr" rid="B27">Lahav and Liddle, 2019</xref>). Each of these 57 configurations was run with different values of &#x3a9;<sub>
<italic>m</italic>
</sub> and <italic>&#x3c3;</italic>
<sub>8</sub>, resulting in the parameter grid shown in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. The output of the simulator consists of the particle positions in 3D space. The mass maps are obtained by the <italic>gravitational lensing</italic> technique (see (<xref ref-type="bibr" rid="B3">Bartelmann, 2010</xref>) for review). It consists of a tomographic projection of the particle densities along the radial (redshift) direction against the <italic>lensing kernel</italic>. This kernel is dependent on the relative distances between the observer and the lensed galaxies that are used to create the mass maps. The source galaxy redshift distribution <italic>n</italic>(<italic>z</italic>) used in this work is the non-tomographic distribution from (<xref ref-type="bibr" rid="B9">Fluri et&#x20;al., 2019</xref>). The projected matter distribution is pixelized into images of size 128 px &#xd7; 128 px, which corresponds to 5&#xb0; &#xd7; 5&#xb0; of the sky. Eventually, the resulting dataset consists of 57 sets of 12,000 sky convergence maps for a total of 684,000 samples. At training time, we randomly rotate and flip the input image to augment the dataset.</p>
<fig id="F2" position="float">
<label>FIGURE 2</label>
<caption>
<p>The cosmological parameter grid used in this work, from (<xref ref-type="bibr" rid="B9">Fluri et&#x20;al., 2019</xref>). The circles and diamonds show the training and the test sets, respectively. The total number of models was 57, of which 46 were used as the training set and 11 as the test set. The models labeled A, B, C, and D are investigated in more detail in <xref ref-type="sec" rid="s5">Section 5</xref>.</p>
</caption>
<graphic xlink:href="frai-04-673062-g002.tif"/>
</fig>
<p>The dataset is split into a training and test set in the following way: 11 cosmologies (132,000 samples) are selected for the test set, and the remaining 46 cosmologies (552,000 samples) are assigned to the training set, as depicted in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. This split is used to ensure that the model could interpolate to unseen cosmologies. At evaluation time, we use the cosmologies from the test set to validate the interpolation ability of our network. In the following sections, we show detailed summary statistics for the cosmologies marked with letters A, B, C, and D. We make the dataset publicly available.<xref ref-type="fn" rid="fn3">
<sup>3</sup>
</xref>
</p>
</sec>
<sec id="s4">
<title>4 Quantitative Comparison Metrics</title>
<p>We make a quantitative assessment of the quality of the generated maps using both cosmological summary statistics and similarity metrics used in computer vision. We focus on the following statistics:<list list-type="simple">
<list-item>
<p>1. the power spectral density <inline-formula id="inf72">
<mml:math id="m74">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, which describes how strongly the maps are correlated as a function of pixel separation <inline-formula id="inf73">
<mml:math id="m75">
<mml:mi>&#x2113;</mml:mi>
</mml:math>
</inline-formula>,</p>
</list-item>
<list-item>
<p>2. the distribution of mass map pixels <inline-formula id="inf74">
<mml:math id="m76">
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mtext>pixels</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, compared using Wasserstein-1 distance and histograms,</p>
</list-item>
<list-item>
<p>3. the distribution of mass map peaks <inline-formula id="inf75">
<mml:math id="m77">
<mml:mrow>
<mml:msub>
<mml:mi>N</mml:mi>
<mml:mrow>
<mml:mtext>peaks</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, which describes the distribution of values at the local maxima of the map, compared also compared using Wasserstein-1 distance and histograms,</p>
</list-item>
<list-item>
<p>4. the bispectrum <inline-formula id="inf76">
<mml:math id="m78">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, which describes the three-point correlation of the folded triangles of different&#x20;size,</p>
</list-item>
<list-item>
<p>5. Minkowski functionals, which are morphological measures of the map, and consist of three functions: <italic>V</italic>
<sub>0</sub>, which describes the&#x20;area of the islands after thresholding of the map at some density level, <italic>V</italic>
<sub>1</sub>, their perimeter, and <italic>V</italic>
<sub>2</sub>, their Euler characteristic (their number count minus the number of holes),</p>
</list-item>
<list-item>
<p>6. the Pearson&#x2019;s correlation matrices <inline-formula id="inf77">
<mml:math id="m79">
<mml:mrow>
<mml:msub>
<mml:mi>R</mml:mi>
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
<mml:mi>&#x2113;</mml:mi>
<mml:mtext>&#x27;</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> between the <inline-formula id="inf78">
<mml:math id="m80">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> of maps at different cosmologies,</p>
</list-item>
<list-item>
<p>7. the Multi-Scale Structural Similarity Index (MS-SSIM) (<xref ref-type="bibr" rid="B59">Wang et&#x20;al., 2003</xref>; <xref ref-type="bibr" rid="B34">Odena et&#x20;al., 2017</xref>), which is an image similarity measure commonly used in computer vision,</p>
</list-item>
<list-item>
<p>8. the Fr&#xe9;chet Distance between the output of a CNN regressor trained to predict &#x3a9;<sub>
<italic>m</italic>
</sub>, <italic>&#x3c3;</italic>
<sub>8</sub>, similarly to the Fr&#xe9;chet Inception Distance calculated using the Google Inception v3 network (<xref ref-type="bibr" rid="B20">Heusel et&#x20;al., 2017</xref>).</p>
</list-item>
</list>
</p>
<p>The mass map histograms and the peak counts are simple statistics used to compare the maps and constrain cosmological models (see <xref ref-type="bibr" rid="B11">Gatti et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B21">Kacprzak et&#x20;al., 2016</xref> for examples). These metrics, however, ignore the spatial information in the maps. The angular power spectrum <inline-formula id="inf79">
<mml:math id="m81">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> or its real-space equivalent, the angular correlation function, is the most common statistic for constraining cosmology with LSS (see (<xref ref-type="bibr" rid="B22">Kilbinger, 2015</xref>) for review). The 2-pt functions capture only the Gaussian part of the fluctuations. The 3-pt correlation function, or the bispectrum, probes higher order information and has also been used for constraining cosmological models (<xref ref-type="bibr" rid="B53">Takada and Jain, 2003</xref>; <xref ref-type="bibr" rid="B10">Fu et&#x20;al., 2014</xref>). Similarly, the Minkowski functionals have also been used for cosmological measurements (<xref ref-type="bibr" rid="B39">Petri et&#x20;al., 2015</xref>) as an alternative statistic that extracts topological information from the&#x20;maps.</p>
<p>The agreement between the pixel and peak values of N-body and GAN-generated images is quantified using the Wasserstein-1 distance <italic>W</italic>
<sub>1</sub>(<italic>P</italic>, <italic>Q</italic>). This distance corresponds to the <italic>optimal transport</italic> of probability mass to turn the distribution <italic>P</italic> into <italic>Q</italic>. As it is scale-dependent, we calculate it after normalizing the pixel values: we subtract the mean and divide by the standard deviation. We use mean and standard deviation of all N-body generated images for a given cosmology, for both samples. This way, the <italic>W</italic>
<sub>1</sub> distance is easily interpretable: for a Gaussian with <inline-formula id="inf80">
<mml:math id="m82">
<mml:mrow>
<mml:mi>&#x03BC;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0</mml:mn>
<mml:mo>,</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, a 1&#x3c3; shift of the mean corresponds to <inline-formula id="inf81">
<mml:math id="m83">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, and scaling its variance by <inline-formula id="inf82">
<mml:math id="m84">
<mml:mrow>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> lead to <inline-formula id="inf83">
<mml:math id="m85">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2248;</mml:mo>
<mml:mn>0.8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
<p>For <inline-formula id="inf84">
<mml:math id="m86">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf85">
<mml:math id="m87">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, and <inline-formula id="inf86">
<mml:math id="m88">
<mml:mrow>
<mml:msub>
<mml:mi>V</mml:mi>
<mml:mrow>
<mml:mn>0,1,2</mml:mn>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> we calculate the simple fractional difference between the original and generated samples, defined as <inline-formula id="inf87">
<mml:math id="m89">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>GAN</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. We quantify the agreement between correlation matrices by comparing their Frobenious norms <inline-formula id="inf88">
<mml:math id="m90">
<mml:mrow>
<mml:mo>&#x2225;</mml:mo>
<mml:mo>&#x22c5;</mml:mo>
<mml:msub>
<mml:mo>&#x2225;</mml:mo>
<mml:mi>F</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. For a N-dimensional, diagonal covariance matrix with elements <inline-formula id="inf89">
<mml:math id="m91">
<mml:mrow>
<mml:msubsup>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>i</mml:mi>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</inline-formula>, the Frobenious norm scales linearly with <inline-formula id="inf90">
<mml:math id="m92">
<mml:mrow>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mo>/</mml:mo>
<mml:mrow>
<mml:msqrt>
<mml:mi>N</mml:mi>
</mml:msqrt>
</mml:mrow>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. This way, it can be interpreted as a linear proxy for information content. We define the fractional difference between the Frobenious norm of GAN and N-body correlation matrices as:<disp-formula id="e3">
<mml:math id="m93">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mtext>R</mml:mtext>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mtext>R</mml:mtext>
<mml:mrow>
<mml:mtext>GAN</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>F</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mtext>R</mml:mtext>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>F</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2016;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mtext>R</mml:mtext>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>-</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>&#x2016;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mi>F</mml:mi>
</mml:msub>
</mml:mrow>
</mml:mfrac>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(3)</label>
</disp-formula>The Multi-Scale Structural Similarity Index (MS-SSIM) is useful in order to detect the problem commonly known as mode collapse, where the generator produces only a small subset of the training data distribution. Detecting this undesirable behavior is non-trivial as summary statistics can still agree during mode collapse. Taking inspiration from (<xref ref-type="bibr" rid="B34">Odena et&#x20;al., 2017</xref>), one solution is to leverage the MS-SSIM score from (<xref ref-type="bibr" rid="B59">Wang et&#x20;al., 2003</xref>) to quantify this effect. This metric was first proposed for prediction of similarity in human perception of images. Taking two images as inputs, it returns a value between 0 and 1, where 1 means &#x201c;identical&#x201d; and 0 means &#x201c;completely different.&#x201d; As the mass maps are stochastic and only similar in a statistical way, we are not interested in the similarity between a pair of specific images, but in the average similarity of a large set of images. We calculate the significance of the difference in the SSIM measures in the following way:<disp-formula id="e4">
<mml:math id="m94">
<mml:mrow>
<mml:msub>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mfrac>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x2329;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext>GAN</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>&#x232a;</mml:mo>
</mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mrow>
<mml:mo>&#x2329;</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>&#x232a;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext>GAN</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
<mml:mo>&#x2b;</mml:mo>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msup>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
<label>(4)</label>
</disp-formula>where <inline-formula id="inf91">
<mml:math id="m95">
<mml:mrow>
<mml:mo>&#x2329;</mml:mo>
<mml:mtext>SSIM</mml:mtext>
<mml:mo>&#x232a;</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> is the mean score, and <inline-formula id="inf92">
<mml:math id="m96">
<mml:mrow>
<mml:mi>&#x3c3;</mml:mi>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula> is the standard deviation. Large differences in the SSIM score indicate a significant difference in the samples generated by the GAN, thus pointing out to potential problems with the quality of the generated samples. On the other hand, a small difference will be&#x20;an indicator that the generative model preserve the data statistics.</p>
<p>Finally, we calculate an adaptation of the Fr&#xe9;chet Inception Distance (FID) (<xref ref-type="bibr" rid="B19">Heusel et&#x20;al., 2017</xref>) between N-body and GAN -generated images. The Inception Score (IS) (<xref ref-type="bibr" rid="B47">Salimans et&#x20;al., 2016</xref>) and FID have become standard measures for GANs. The idea consists to compare statistics of the output of the Google Inception-v3 network (<xref ref-type="bibr" rid="B52">Szegedy et&#x20;al., 2016</xref>) for the ImageNet dataset (<xref ref-type="bibr" rid="B5">Deng et&#x20;al., 2009</xref>). This has proven to be well correlated with human score. As the reference Inception network used for the FID was trained with the ImageNet dataset, its output statistics are meaningless for cosmological mass maps. To solve this challenge, we create our own reference network that is well suited for cosmological mass maps. This network is a CNN trained to perform a regression task and predict the true <italic>&#x3c3;</italic>
<sub>8</sub>, &#x3a9;<sub>
<italic>m</italic>
</sub> parameters, similarly to (<xref ref-type="bibr" rid="B8">Fluri et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B49">Schmelzle et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B16">Gupta et&#x20;al., 2018</xref>). Its parameters and detailed explanations of its construction can be found in <xref ref-type="table" rid="T2">Table A2</xref> and in <xref ref-type="app" rid="app2">Appendix B</xref>. The adapted FID score is obtained by comparing the regressor outputs for the N-body and GAN images. As regressor is composed of seven layers, this comparison depends on high order moments. Naturally, we expect that a well working conditional GAN should generate samples with similar output distribution to the one of the real samples. To estimate the distance between the two statistics distributions, we first approximate the network predictions with a normal distributions <inline-formula id="inf93">
<mml:math id="m97">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mtext>&#x3a3;</mml:mtext>
<mml:mi>r</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf94">
<mml:math id="m98">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>g</mml:mi>
</mml:msub>
<mml:mo>,</mml:mo>
<mml:msub>
<mml:mtext>&#x3a3;</mml:mtext>
<mml:mi>g</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, for the N-body and GAN -generated input, respectively. The FID is then calculated as:<disp-formula id="e5">
<mml:math id="m99">
<mml:mrow>
<mml:mtext>FID</mml:mtext>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>r</mml:mi>
</mml:msub>
<mml:mo>&#x2212;</mml:mo>
<mml:msub>
<mml:mi>&#x3bc;</mml:mi>
<mml:mi>g</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>&#x2b;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mo>&#x7c;</mml:mo>
<mml:mrow>
<mml:msubsup>
<mml:mtext>&#x3a3;</mml:mtext>
<mml:mi>r</mml:mi>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x2212;</mml:mo>
<mml:msubsup>
<mml:mtext>&#x3a3;</mml:mtext>
<mml:mi>g</mml:mi>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msubsup>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo>&#x7c;</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
</mml:msup>
<mml:mo>.</mml:mo>
</mml:mrow>
</mml:math>
<label>(5)</label>
</disp-formula>Note that this formula also correspond to the Wasserstein-1 distance between the two Gaussian distributions (<xref ref-type="bibr" rid="B6">Dowson and Landau, 1982</xref>). Eventually, before calculating FID, we normalize the network outputs for each true cosmology: we subtract the mean and divide by the standard deviation of the N-body sample. For the ease of interpretation, we report the square root of FID. This way, a 1<italic>&#x3c3;</italic> difference in the mean CNN predictions will correspond to <inline-formula id="inf95">
<mml:math id="m100">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>FID</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1. Similarly, a change of 1<italic>&#x3c3;</italic> in the covariance matrix also leads to <inline-formula id="inf96">
<mml:math id="m101">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>FID</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d;&#x20;1.</p>
</sec>
<sec id="s5">
<title>5 Results</title>
<p>We trained the GAN model described in <xref ref-type="sec" rid="s2">Section 2</xref> and <xref ref-type="app" rid="app1">
<bold>Appendix A</bold>
</xref>. We used <sc>RMSProp</sc> as an optimizer with an initial learning rate of <inline-formula id="inf129">
<mml:math id="m134">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and a batch size of 64. The discriminator was updated 5&#x20;times more than the generator. The gradient penalty was set to 10 and the negative slope of the LeakyRelu <inline-formula id="inf130">
<mml:math id="m135">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>. It took a week to train the model for 40 epochs on a GeForce GTX 1080 GPU. Similar to (<xref ref-type="bibr" rid="B43">Reed et&#x20;al., 2016</xref>; <xref ref-type="bibr" rid="B34">Odena et&#x20;al., 2017</xref>; <xref ref-type="bibr" rid="B31">Miyato and Koyama, 2018</xref>; <xref ref-type="bibr" rid="B30">Mirza and Osindero, 2014</xref>), we use batches composed of samples from different parameter sets. Note that the batches were composed of samples from different cosmologies from the <italic>training</italic> set. The summary statistics are computed using 5,000 real and fake samples for every pair of parameters of the <italic>test</italic> set. The peaks are extracted by searching for all pixels greater than their <inline-formula id="inf131">
<mml:math id="m136">
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> patch neighborhood, i.e. their 24 neighbors. Then, the histogram of the extracted peaks values is computed. We rely on <sc>LensTools</sc> (<xref ref-type="bibr" rid="B40">Petri, 2016</xref>) to compute the power spectra, bispectra and the Minkowski functionals. For the bispectrum, we use the <italic>folded</italic> configuration, with the ratio between one of the triangle sides and the base is set to the default value of 0.5. The SSIM is computed using the <sc>Scikit-Image</sc> packge<xref ref-type="fn" rid="fn4">
<sup>4</sup>
</xref> (<xref ref-type="bibr" rid="B57">Van der Walt et&#x20;al., 2014</xref>). We make our code is publicly available.<xref ref-type="fn" rid="fn5">
<sup>5</sup>
</xref>
</p>
<p>
<xref ref-type="fig" rid="F3">Figure&#x20;3</xref> shows images generated by the conditional GAN and as well as original ones, for several values of &#x3a9;<sub>
<italic>m</italic>
</sub> and <italic>&#x3c3;</italic>
<sub>8</sub> parameters. They are visually indistinguishable. Furthermore, the image structure evolves similarly with respect of the cosmological parameters change. As predicted by the theory, increasing &#x3a9;<sub>
<italic>m</italic>
</sub> results in convergence maps with additional mass and increasing <italic>&#x3c3;</italic>
<sub>8</sub> in images with higher variance in pixel intensities. In <xref ref-type="fig" rid="F4">Figure&#x20;4</xref> the same latent variable <italic>z</italic> is used to generate different cosmologies. The smooth transition from low to high mass density hints that the latent variable control the overall mass distribution and the conditioning parameter its two cosmological properties <italic>&#x3c3;</italic>
<sub>8</sub>, &#x3a9;<sub>
<italic>m</italic>
</sub>.</p>
<fig id="F3" position="float">
<label>FIGURE 3</label>
<caption>
<p>The original N-body images and GAN-generated images for four cosmological parameter&#x20;sets.</p>
</caption>
<graphic xlink:href="frai-04-673062-g003.tif"/>
</fig>
<fig id="F4" position="float">
<label>FIGURE 4</label>
<caption>
<p>Images generated with the same random seed but with different input cosmology parameters &#x3a9;<sub>
<italic>m</italic>
</sub> and <italic>&#x3c3;</italic>
<sub>8</sub>.</p>
</caption>
<graphic xlink:href="frai-04-673062-g004.tif"/>
</fig>
<p>
<xref ref-type="fig" rid="F5">Figure&#x20;5</xref> shows the histograms of pixels (top) and peaks (bottom) of the original maps simulated using N-body simulations (blue), and their GAN-generated equivalents (red), for the four models A,B,C,D shown in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. The peak counts were selected as maxima of the surrounding 24 neighbors. The solid line corresponds to the median of the histograms from 5,000 realisations, and the bands to 32% and 68% percentiles. The bottom part of each panel shows the fractional difference between the statistics, defined as <inline-formula id="inf132">
<mml:math id="m137">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mi>x</mml:mi>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mrow>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>GAN</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
<mml:mo>/</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>x</mml:mi>
<mml:mrow>
<mml:mtext>N</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>body</mml:mtext>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:mrow>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. The normlized Wasserstein-1 distance of the pixel values distribution (see <xref ref-type="sec" rid="s4">Section 4</xref>) is: <inline-formula id="inf133">
<mml:math id="m138">
<mml:mrow>
<mml:msubsup>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mtext>pixel</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.04</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.02</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.02</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.02</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for models A, B, C, and D, respectively. That indicates that the histograms differ on the level of &#x3c;5%. Similarly, the Wasserstein-1 distances of the peak value distribution for models A, B, C, and D is: <inline-formula id="inf134">
<mml:math id="m139">
<mml:mrow>
<mml:msubsup>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
<mml:mrow>
<mml:mtext>peak</mml:mtext>
</mml:mrow>
</mml:msubsup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.04</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.03</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.01</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.02.</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> The agreement here is also very good, on &#x3c;5%&#x20;level.</p>
<fig id="F5" position="float">
<label>FIGURE 5</label>
<caption>
<p>Comparison of histogram of pixel values <bold>(top)</bold> and peaks <bold>(bottom)</bold> between the original N-body and the GAN-generated maps. Models A, B, C, D correspond to the ones marked in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. The <italic>x</italic>-axis value <italic>&#x3ba;</italic> is the map pixel intensity. The solid line is the median histogram from 5,000 randomly selected maps. The bands correspond to 32 and 68% confidence limits of the ensemble of histograms. The lower panels show the fractional difference between the median histograms.</p>
</caption>
<graphic xlink:href="frai-04-673062-g005.tif"/>
</fig>
<p>The 2-pt and 3-pt statistics are shown in <xref ref-type="fig" rid="F6">Figure&#x20;6</xref>. The power spectra <inline-formula id="inf135">
<mml:math id="m140">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> overlap almost perfectly for all the cosmologies lying inside the parameter grid used for training. Again, the agreement is better than 5%. The agreement for the bispectrum <inline-formula id="inf136">
<mml:math id="m141">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> is good for models B and C, but worse for A and B; the GAN model seems to underestimate the strength of the 3-pt correlations for these models, which differ by &#x2248;20%. We note that the 3-pt signal is very weak and has a large variation, which may be difficult to model for the&#x20;GANs.</p>
<fig id="F6" position="float">
<label>FIGURE 6</label>
<caption>
<p>Comparison of the 2-pt and 3-pt functions between the original N-body maps and GAN-generated maps. The structure of this figure is the same as for <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>.</p>
</caption>
<graphic xlink:href="frai-04-673062-g006.tif"/>
</fig>
<p>The Minkowski functionals are presented in <xref ref-type="fig" rid="F7">Figure&#x20;7</xref>. They were calculated using <sc>Lenstools</sc> (<xref ref-type="bibr" rid="B40">Petri, 2016</xref>). The functional <italic>V</italic>
<sub>0</sub> (first line) corresponds to the area of the emerging &#x201c;islands,&#x201d; <italic>V</italic>
<sub>1</sub> (second line) to their circumference, and <italic>V</italic>
<sub>2</sub> (third line) to their Euler characteristic (their number count minus the number of holes in them). The value of threshold &#x3ba;, above which the functional values, i.e. the &#x201c;islands&#x201d; are calculated, is shown on the <italic>x</italic>-axis. Here the agreement is typically better than 10%, with some model D agreeing much better, to &#x2248;2%. The large differences in the fractional difference plots are due to instability close to value of <italic>V</italic>&#x20;&#x3d; 0. The confidence limits of the summary statistics shown in these figures overlap very well, which indicates that the variability of these statistics is also captured very well by the GAN system.</p>
<fig id="F7" position="float">
<label>FIGURE 7</label>
<caption>
<p>Comparison of the Minkowski functional between the original N-body maps and GAN-generated maps. The value of threshold <italic>&#x3ba;</italic>, above which the functional value is calculated, is shown on the <italic>x</italic>-axis. The functional <italic>V</italic>
<sub>0</sub> corresponds to the area of the emerging &#x201c;islands,&#x201d; <italic>V</italic>
<sub>1</sub> to their circumference, and <italic>V</italic>
<sub>2</sub> to their Euler characteristic (their number count minus the number of holes in them). The structure of this figure is the same as for <xref ref-type="fig" rid="F5">Figure&#x20;5</xref>.</p>
</caption>
<graphic xlink:href="frai-04-673062-g007.tif"/>
</fig>
<p>The Pearson&#x2019;s correlation matrices <bold>R</bold> of the power spectra are shown in <xref ref-type="fig" rid="F8">Figure&#x20;8</xref>. Those correlations were created from a coarsely-binned <inline-formula id="inf137">
<mml:math id="m142">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> in range <inline-formula id="inf138">
<mml:math id="m143">
<mml:mrow>
<mml:mi>&#x2113;</mml:mi>
<mml:mo>&#x2208;</mml:mo>
<mml:mrow>
<mml:mo>[</mml:mo>
<mml:mn>300,3000</mml:mn>
<mml:mo>]</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</inline-formula>. The upper and lower triangular parts of the matrix show the original N-body correlations and the GAN correlations, respectively. We calculate the Frobenious norms of these matrices and compare their ratios using <xref ref-type="disp-formula" rid="e3">Eq. 3</xref>. For the models A, B, C, D this difference is: <italic>f</italic>
<sub>R</sub> &#x3d; 0.02, 0.14, 0.06, 0.06. This agreement is overall very good, with model B being slightly worse. As the precision requirements for covariance matrices are not as strict as for the summary statistics, this level of agreement can be considered satisfactory for upcoming applications (<xref ref-type="bibr" rid="B55">Taylor et&#x20;al., 2013</xref>).</p>
<fig id="F8" position="float">
<label>FIGURE 8</label>
<caption>
<p>Pearson&#x2019;s correlation matrices for the four models highlighted in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. The upper triangular corresponds to the original N-body power spectra, while the lower triangular to the power spectra of GAN-generated images. The fractional difference of the Frobenious norms (<xref ref-type="disp-formula" rid="e3">Eq. 3</xref>) of these matrices is <italic>f</italic>
<sub>R</sub> &#x3d; 0.02, 0.14, 0.06, 0.06, for models A, B, C, and D, respectively.</p>
</caption>
<graphic xlink:href="frai-04-673062-g008.tif"/>
</fig>
<p>We calculate the mean and standard deviation of MS-SSIM score between 5,000 randomly selected images for each cosmology, both for GAN and original N-body maps. We test if&#x20;the mean SSIM score is consistent between the N-body and GAN data using <xref ref-type="disp-formula" rid="e4">Eq. 4</xref>. The SSIM difference significance for the&#x20;four models A, B, C, and D are: <inline-formula id="inf139">
<mml:math id="m144">
<mml:mrow>
<mml:msub>
<mml:mi>s</mml:mi>
<mml:mrow>
<mml:mtext>SSIM</mml:mtext>
</mml:mrow>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.08</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.23</mml:mn>
<mml:mo>,</mml:mo>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>0.27</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>0.42</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula>. This indicates very good statistical agreement for these models.</p>
<p>
<xref ref-type="fig" rid="F9">Figure&#x20;9</xref> shows the prediction of a regressor CNN trained on the N-body images with true <italic>&#x3c3;</italic>
<sub>8</sub>, &#x3a9;<sub>
<italic>m</italic>
</sub> values. For each category, we make the prediction with 500 randomly selected maps. The shaded areas show the 68 and 95% probability contours for the N-body image input (blue) and the GAN image input (red). The agreement is relatively good, but differences in the spread of these distributions is noticeable. The Fr&#xe9;chet Distance (FID) computed using the reference cosmological CNN, as described in <xref ref-type="sec" rid="s4">Section 4</xref>, is: <inline-formula id="inf140">
<mml:math id="m145">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>FID</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1.26</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.44</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.00</mml:mn>
<mml:mo>,</mml:mo>
<mml:mn>1.19</mml:mn>
<mml:mo>,</mml:mo>
</mml:mrow>
</mml:math>
</inline-formula> for models A,B,C, and D. This indicates a slight difference according to this metric and agree with the distributions in <xref ref-type="fig" rid="F9">Figure&#x20;9</xref>.</p>
<fig id="F9" position="float">
<label>FIGURE 9</label>
<caption>
<p>Predictions of a regressor CNN trained to predict the &#x3a9;<sub>
<italic>m</italic>
</sub>, <italic>&#x3c3;</italic>
<sub>8</sub> from input images. The details of this experiment are described in <xref ref-type="sec" rid="s4">Section 4</xref> and the network architecture in <xref ref-type="table" rid="T1">Table A1</xref> in <xref ref-type="app" rid="app1">Appedix A</xref>. The contours encircle the 68 and 95% samples for the N-body maps (blue) and GAN-generated maps from the test set. The black stars show the true values of the test set parameters.</p>
</caption>
<graphic xlink:href="frai-04-673062-g009.tif"/>
</fig>
<p>We compare the summary statistics as a function of cosmological parameters for both the training and the test set. We used the training and test sets displayed in <xref ref-type="fig" rid="F2">Figure&#x20;2</xref>. <xref ref-type="fig" rid="F10">Figure&#x20;10</xref> shows the six quantities as a function of cosmological parameters:<list list-type="simple">
<list-item>
<p>top left: significance of the difference <italic>s</italic>
<sub>SSIM</sub> in Multi-Scale Structural Similarity Index (<xref ref-type="disp-formula" rid="e4">Eq.&#x20;4</xref>),</p>
</list-item>
<list-item>
<p>top center: Fr&#xe9;chet distance using a CNN regressor (<xref ref-type="disp-formula" rid="e5">Eq. 5</xref>). Note that for a Gaussian distribution, a difference of <inline-formula id="inf141">
<mml:math id="m146">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mtext>FID</mml:mtext>
</mml:mrow>
<mml:mrow>
<mml:mn>1</mml:mn>
<mml:mo>/</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> &#x3d; 1 corresponds to either a 1<italic>&#x3c3;</italic> shift in the mean or 1<italic>&#x3c3;</italic> difference in standard deviation,</p>
</list-item>
<list-item>
<p>top right: normalized Wasserstein-1 distance of the pixel value distribution. For a Gaussian distribution, a 1<italic>&#x3c3;</italic> change in the mean corresponds to W<sub>1</sub> &#x3d; 1, and an increase in standard deviation of &#xd7;2 to <inline-formula id="inf142">
<mml:math id="m147">
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x2248;</mml:mo>
<mml:mn>0.8</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>,</p>
</list-item>
<list-item>
<p>bottom left: average fractional differences in the power spectrum <inline-formula id="inf143">
<mml:math id="m148">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>,</p>
</list-item>
<list-item>
<p>bottom center: fractional differences in the power bispectrum <inline-formula id="inf144">
<mml:math id="m149">
<mml:mrow>
<mml:msub>
<mml:mi>f</mml:mi>
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>,</p>
</list-item>
<list-item>
<p>bottom left: fractional difference in the Frobenious norm of the correlation matrices f<sub>R</sub> (<xref ref-type="disp-formula" rid="e3">Eq.&#x20;3</xref>).</p>
</list-item>
</list>
</p>
<fig id="F10" position="float">
<label>FIGURE 10</label>
<caption>
<p>Differences between summary statistics of the original N-body and the GAN-generated images. The left panel shows the significance of the difference in Multi-Scale Structural Similarity Index (MS-SSIM), defined in <xref ref-type="disp-formula" rid="e4">Eq. 4</xref>. The upper middle panel presents the Fr&#xe9;chet Distance, computed using a regressor CNN and <xref ref-type="disp-formula" rid="e5">Eq. 5</xref> (see <xref ref-type="sec" rid="s4">Section 4</xref>). The upper right panel shows the normalized Wasserstein-1 distance in the pixel value distributions (see <xref ref-type="sec" rid="s4">Section 4</xref>). The lower left panel shows the mean absolute fractional difference of the power spectra <inline-formula id="inf145">
<mml:math id="m150">
<mml:mrow>
<mml:msub>
<mml:mi>C</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The lower middle panel shows the fractional difference in the Frobenius norms of correlation matrices, defined in <xref ref-type="disp-formula" rid="e3">Eq. 3</xref>. The lower right panel present the mean absolute fractional difference of the bispectrum <inline-formula id="inf146">
<mml:math id="m151">
<mml:mrow>
<mml:msub>
<mml:mi>B</mml:mi>
<mml:mi>&#x2113;</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. The circles and squares indicates parameters from the training and the test&#x20;sets.</p>
</caption>
<graphic xlink:href="frai-04-673062-g010.tif"/>
</fig>
<p>Overall, we notice that the agreement between the N-body simulations and GAN-generated maps is the best in the center of the grid for both the training and test set. The fact that the differences in neighboring cosmologies are similar indicates that the GAN system can efficiently learn the latent interpolation of the maps. The agreement worsens at the edges of the grid. We observe the biggest deterioration in the realism of the GAN model for the high &#x3a9;<sub>
<italic>m</italic>
</sub> and low <italic>&#x3c3;</italic>
<sub>8</sub> parameters. This is most prominent for the correlation matrix and the SSIM differences. Conversely, the biggest difference for the bispectrum is present for low &#x3a9;<sub>
<italic>m</italic>
</sub> and high <italic>&#x3c3;</italic>
<sub>8</sub>.</p>
</sec>
<sec id="s6">
<title>6 Conclusion</title>
<p>We proposed a new conditional GAN model for continuous parameters where conditioning is done in the latent space. We demonstrated the ability of this model to generate sky convergence maps when conditioning on the cosmological parameters &#x3a9;<sub>
<italic>m</italic>
</sub> and <italic>&#x3c3;</italic>
<sub>8</sub>. Our model is able to produce samples that resemble samples from the test set with good statistical accuracy, which demonstrates its generalization abilities. The agreement of the low order summary statistics (pixel and peak histograms and power spectrum) is very good, typically on the &#x3c;5% level. Higher order statistics (Minkowski functionals, bispectrum) agree well, but with larger differences, generally around &#x2248;10%, and in some cases &#x2248;20%. The comparison of the Multi-Scale Structural Similarity Index (MS-SSIM) shows a good agreement in this metric, with the exception of the low <italic>&#x3c3;</italic>
<sub>8</sub> and high &#x3a9;<sub>
<italic>m</italic>
</sub> edge of the grid. Moreover, the GAN model is able to capture the variability in the conditioned dataset: we observe that the scatter of the summary statistics computed from an ensemble is very similar between the original and generated images. The investigation of the correlation matrices of the power spectra also shows a good agreement, with a quality deteriorating close to the edges of the grid, especially for low <italic>&#x3c3;</italic>
<sub>8</sub> and high &#x3a9;<sub>
<italic>m</italic>
</sub>. This is not unexpected, as the training set contains less information near the edges of the grid. More investigation is needed to more close inspect the behavior of the generative model in these areas. As generative models are rapidly growing in popularity in machine learning, we anticipate to be able to solve these problems in the near future.</p>
<p>Our results offer good prospects for GAN-based conditional models to be used as emulators of cosmology-dependent mass maps. As these models efficiently capture both the signal and its variability, the map-level emulators could potentially be used for cosmological analyses. They can accurately predict the power spectrum and its covariance, which is often unattainable in standard cosmological analyses (<xref ref-type="bibr" rid="B7">Eifler et&#x20;al., 2009</xref>). It can also be used for non-Gaussian analyses of lensing mass maps, such as, for example, in (<xref ref-type="bibr" rid="B60">Z&#xfc;rcher et&#x20;al., 2020</xref>; <xref ref-type="bibr" rid="B35">Parroni et&#x20;al., 2020</xref>). Further experiments will be needed, however, to bring the generative models to a level where they can be of practical use in a full, end-to-end cosmological analysis.</p>
<p>In this paper, we have demonstrated the ability of generative AI models to serve as emulators of cosmological mass maps for a given redshift distribution of source galaxies <italic>n</italic>(<italic>z</italic>). Generative models have also been shown to work directly on the full or sliced 3D matter density distributions (<xref ref-type="bibr" rid="B33">Nathana&#xeb;l et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B56">Tr&#xf6;ster et&#x20;al., 2019</xref>; <xref ref-type="bibr" rid="B58">Villaescusa-Navarro et&#x20;al., 2020</xref>). The three dimensional generation of cosmological fields proves to be particularly difficult. As most of the survey experiments publish their lensing catalogs and their corresponding redshift distributions, the generation of projected maps, as shown in this work, could be of direct practical use. Another challenge will be posed by the large sky area of the upcoming surveys and their spherical geometry. Spherical convolutional neural networks architectures have been proposed (<xref ref-type="bibr" rid="B37">Perraudin et&#x20;al., 2018</xref>; <xref ref-type="bibr" rid="B26">Krachmalnicoff and Tomasi, 2019</xref>; <xref ref-type="bibr" rid="B29">McEwen et&#x20;al., 2021</xref>). These architectures are expected to be easy to implement with generative models, which offers good prospect for the development of spherical mass map emulators.</p>
</sec>
</body>
<back>
<sec id="s7">
<title>Data Availability Statement</title>
<p>The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: <ext-link ext-link-type="uri" xlink:href="https://renkulab.io/projects/nathanael.perraudin/darkmattergan">https://renkulab.io/projects/nathanael.perraudin/darkmattergan</ext-link>, <ext-link ext-link-type="uri" xlink:href="https://zenodo.org/record/4646764">https://zenodo.org/record/4646764</ext-link>
</p>
</sec>
<sec id="s8">
<title>Author Contributions</title>
<p>NP experiment lead, experiment execution, code maintenance, paper writing SM experiment execution, implementation of conditional GANs AL machine learning advice, consulting, paper writing TK dataset creation, cosmology advice, consulting, paper writing, plotting, method evaluation.</p>
</sec>
<sec id="s9">
<title>Funding</title>
<p>This work was supported by a grant from the Swiss Data Science Center (SDCS) under project &#x201c;DLOC: Deep Learning for Observational Cosmology&#x201d; and Grant Number 200021&#x20;169130 from the Swiss National Science Foundation (SNSF).</p>
</sec>
<sec sec-type="COI-statement" id="s10">
<title>Conflict of Interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<ack>
<p>This work was supported by a Grant from the Swiss Data Science Center (SDCS) under project <italic>DLOC: Deep Learning for Observational Cosmology</italic> and Grant Number 200021_169130 from the Swiss National Science Foundation (SNSF). We thank Thomas Hofmann, Alexandre R&#xe9;fr&#xe9;gier, and Fernando Perez-Cruz for advice and helpful discussions. We thank the Cosmology Research Group of ETHZ and particularly Janis Fluri for giving us access to the dataset. Finally, we thank the two anonymous reviewers who provided extensive feedback that greatly improved the quality of this&#x20;paper.</p>
</ack>
<fn-group>
<fn id="fn1">
<label>1</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://renkulab.io/gitlab/nathanael.perraudin/darkmattergan">https://renkulab.io/gitlab/nathanael.perraudin/darkmattergan</ext-link>
</p>
</fn>
<fn id="fn2">
<label>2</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://zenodo.org/record/4646764">https://zenodo.org/record/4646764</ext-link>
</p>
</fn>
<fn id="fn3">
<label>3</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://zenodo.org/record/4646764">https://zenodo.org/record/4646764</ext-link>
</p>
</fn>
<fn id="fn4">
<label>4</label>
<p>
<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://scikit-image.org/">https://scikit-image.org/</ext-link>
</p>
</fn>
<fn id="fn5">
<label>5</label>
<p>
<ext-link ext-link-type="uri" xlink:href="https://renkulab.io/gitlab/nathanael.perraudin/darkmattergan">https://renkulab.io/gitlab/nathanael.perraudin/darkmattergan</ext-link>
</p>
</fn>
</fn-group>
<ref-list>
<title>References</title>
<ref id="B1">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Angulo</surname>
<given-names>R. E.</given-names>
</name>
<name>
<surname>Zennaro</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Contreras</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Aric&#xf2;</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Pellejero-Iba&#xf1;ez</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>St&#xfc;cker</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>The BACCO Simulation Project: Exploiting the Full Power of Large-Scale Structure for Cosmology</article-title>. <comment>arXiv e-prints, arXiv:2004.06245</comment>. </citation>
</ref>
<ref id="B2">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Arjovsky</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chintala</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Bottou</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Wasserstein Generative Adversarial Networks</article-title>. <source>Int. Conf. Mach. Learn.</source> <volume>14</volume>, <fpage>214</fpage>&#x2013;<lpage>223</lpage>. <pub-id pub-id-type="doi">10.1109/icpr.2018.8546264</pub-id> </citation>
</ref>
<ref id="B3">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bartelmann</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2010</year>). <article-title>Gravitational Lensing</article-title>. <source>Class. Quan. Grav.</source> <volume>27</volume>, <fpage>233001</fpage>. <pub-id pub-id-type="doi">10.1088/0264-9381/27/23/233001</pub-id> </citation>
</ref>
<ref id="B4">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Collaboration</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Knabenhans</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Stadel</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Marelli</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Potter</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Teyssier</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Euclid Preparation: Ii. The Euclidemulator&#x2013;A Tool to Compute the Cosmology Dependence of the Nonlinear Matter Power Spectrum</article-title>. <source>Monthly Notices R. Astronomical Soc.</source> <volume>484</volume>, <fpage>5509</fpage>&#x2013;<lpage>5529</lpage>. </citation>
</ref>
<ref id="B5">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Deng</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Dong</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Socher</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>L. J.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Fei-Fei</surname>
<given-names>L.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Imagenet: A Large-Scale Hierarchical Image Database</article-title>. <source>IEEE Conf. Computer Vis. Pattern Recognit.</source> <volume>33</volume>, <fpage>248</fpage>&#x2013;<lpage>255</lpage>. <pub-id pub-id-type="doi">10.1109/cvprw.2009.5206848</pub-id> </citation>
</ref>
<ref id="B6">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dowson</surname>
<given-names>D. C.</given-names>
</name>
<name>
<surname>Landau</surname>
<given-names>B. V.</given-names>
</name>
</person-group> (<year>1982</year>). <article-title>The Fr&#xe9;chet Distance between Multivariate normal Distributions</article-title>. <source>J.&#x20;Multivariate Anal.</source> <volume>12</volume>, <fpage>450</fpage>&#x2013;<lpage>455</lpage>. <pub-id pub-id-type="doi">10.1016/0047-259X(82)90077-X</pub-id> </citation>
</ref>
<ref id="B7">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Eifler</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Schneider</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Hartlap</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Dependence of Cosmic Shear Covariances on Cosmology</article-title>. <source>A&#x26;A</source> <volume>502</volume>, <fpage>721</fpage>&#x2013;<lpage>731</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/200811276</pub-id> </citation>
</ref>
<ref id="B8">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fluri</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Refregier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hofmann</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Cosmological Constraints from Noisy Convergence Maps through Deep Learning</article-title>. <comment>arXiv preprint arXiv:1807.08732</comment>. </citation>
</ref>
<ref id="B9">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fluri</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Refregier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hofmann</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Cosmological Constraints with Deep Learning from KiDS-450 Weak Lensing Maps</article-title>. <comment>arXiv e-prints, arXiv:1906.03156</comment>. </citation>
</ref>
<ref id="B10">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Fu</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kilbinger</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Erben</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Heymans</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Hildebrandt</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Hoekstra</surname>
<given-names>H.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>CFHTLenS: Cosmological Constraints from a Combination of Cosmic Shear Two-point and Three-point Correlations</article-title>. <source>Monthly Notices R. Astronomical Soc.</source> <volume>441</volume>, <fpage>2725</fpage>&#x2013;<lpage>2743</lpage>. <pub-id pub-id-type="doi">10.1093/mnras/stu754</pub-id> </citation>
</ref>
<ref id="B11">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gatti</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Chang</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Friedrich</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Bacon</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Crocce</surname>
<given-names>M.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Dark Energy Survey Year 3 Results: Cosmology with Moments of Weak Lensing Mass Maps - Validation on Simulations</article-title>. <source>MNRAS</source> <volume>498</volume>, <fpage>4060</fpage>&#x2013;<lpage>4087</lpage>. <pub-id pub-id-type="doi">10.1093/mnras/staa2680</pub-id> </citation>
</ref>
<ref id="B12">
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Gauthier</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). &#x201c;<article-title>Conditional Generative Adversarial Nets for Convolutional Face Generation</article-title>,&#x201d; in <source>Class Project for Stanford CS231N: Convolutional Neural Networks for Visual Recognition</source> (<publisher-loc>London</publisher-loc>: <publisher-name>Winter Semester</publisher-name>), <fpage>2</fpage>. </citation>
</ref>
<ref id="B13">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Giusarma</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Reyes Hurtado</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Villaescusa-Navarro</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>He</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ho</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hahn</surname>
<given-names>C.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Learning Neutrino Effects in Cosmology with Convolutional Neural Networks</article-title>. <comment>arXiv e-prints, arXiv:1910.04255</comment>. </citation>
</ref>
<ref id="B14">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Mirza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Warde-Farley</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Ozair</surname>
<given-names>S.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Generative Adversarial Nets</article-title>. <source>Adv. Neural Inform. Process. Syst.</source> <volume>11</volume>, <fpage>2672</fpage>&#x2013;<lpage>2680</lpage>. <pub-id pub-id-type="doi">10.3156/jsoft.29.5_177_2</pub-id> </citation>
</ref>
<ref id="B15">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gulrajani</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Ahmed</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Arjovsky</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Dumoulin</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Courville</surname>
<given-names>A. C.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Improved Training of Wasserstein Gans</article-title>. <source>Adv. Neural Inform. Process. Syst.</source> <volume>12</volume>, <fpage>5767</fpage>&#x2013;<lpage>5777</lpage>. <pub-id pub-id-type="doi">10.1109/ismsit.2019.8932868</pub-id> </citation>
</ref>
<ref id="B16">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gupta</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Matilla</surname>
<given-names>J.&#x20;M. Z.</given-names>
</name>
<name>
<surname>Hsu</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Haiman</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Non-gaussian Information from Weak Lensing Data via Deep Learning</article-title>. <source>Phys. Rev. D</source> <volume>97</volume>, <fpage>103515</fpage>. <pub-id pub-id-type="doi">10.1103/physrevd.97.103515</pub-id> </citation>
</ref>
<ref id="B17">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>He</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Feng</surname>
<given-names>Y.</given-names>
</name>
<name>
<surname>Ho</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Ravanbakhsh</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>W.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Learning to Predict the Cosmological Structure Formation</article-title>. <source>Proc. Natl. Acad. Sci. USA</source> <volume>116</volume>, <fpage>13825</fpage>&#x2013;<lpage>13832</lpage>. <pub-id pub-id-type="doi">10.1073/pnas.1821458116</pub-id> </citation>
</ref>
<ref id="B18">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Heitmann</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Bingham</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Lawrence</surname>
<given-names>E.</given-names>
</name>
<name>
<surname>Bergner</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Habib</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Higdon</surname>
<given-names>D.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>The Mira-Titan Universe: Precision Predictions for Dark Energy Surveys</article-title>. <source>ApJ</source> <volume>820</volume>, <fpage>108</fpage>. <pub-id pub-id-type="doi">10.3847/0004-637X/820/2/108</pub-id> </citation>
</ref>
<ref id="B19">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Heusel</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ramsauer</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Unterthiner</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Nessler</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Hochreiter</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017a</year>). <article-title>Gans Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium</article-title>. <source>Adv. Neural Inform. Process. Syst.</source> <volume>7</volume>, <fpage>6626</fpage>&#x2013;<lpage>6637</lpage>. <pub-id pub-id-type="doi">10.1057/9781137294678.0453</pub-id> </citation>
</ref>
<ref id="B20">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Heusel</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Ramsauer</surname>
<given-names>H.</given-names>
</name>
<name>
<surname>Unterthiner</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Nessler</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Hochreiter</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2017b</year>). <article-title>GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium</article-title>. <comment>arXiv e-prints, arXiv:1706.08500</comment>. </citation>
</ref>
<ref id="B21">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Kirk</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Friedrich</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Refregier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Marian</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2016</year>). <article-title>Cosmology Constraints from Shear Peak Statistics in Dark Energy Survey Science Verification Data</article-title>. <source>Mon. Not. R. Astron. Soc.</source> <volume>463</volume>, <fpage>3653</fpage>&#x2013;<lpage>3673</lpage>. <pub-id pub-id-type="doi">10.1093/mnras/stw2070</pub-id> </citation>
</ref>
<ref id="B22">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kilbinger</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Cosmology with Cosmic Shear Observations: a Review</article-title>. <source>Rep. Prog. Phys.</source> <volume>78</volume>, <fpage>086901</fpage>. <pub-id pub-id-type="doi">10.1088/0034-4885/78/8/086901</pub-id> </citation>
</ref>
<ref id="B23">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Kingma</surname>
<given-names>D. P.</given-names>
</name>
<name>
<surname>Ba</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Adam: A Method for Stochastic Optimization</article-title>. <comment>arXiv preprint arXiv:1412.6980</comment> </citation>
</ref>
<ref id="B24">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Knabenhans</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Stadel</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Marelli</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Potter</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Teyssier</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Legrand</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2019</year>). <article-title>Euclid Preparation: II. The EUCLIDEMULATOR&#x2014;a Tool to Compute the Cosmology Dependence of the Nonlinear Matter Power Spectrum</article-title>. <source>Monthly Notices R. Astronomical Soc.</source> <volume>484</volume>, <fpage>5509</fpage>&#x2013;<lpage>5529</lpage>. <pub-id pub-id-type="doi">10.1093/mnras/stz197</pub-id> </citation>
</ref>
<ref id="B25">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Knabenhans</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Stadel</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Potter</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Dakin</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Hannestad</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Tram</surname>
<given-names>T.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>Euclid Preparation: IX. EuclidEmulator2 &#x2013; Power Spectrum Emulation with Massive Neutrinos and Self-Consistent Dark Energy Perturbations</article-title>. <comment>arXiv e-prints, arXiv:2010.11288</comment>. </citation>
</ref>
<ref id="B26">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Krachmalnicoff</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Tomasi</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Convolutional Neural Networks on the HEALPix Sphere: a Pixel-Based Algorithm and its Application to CMB Data Analysis</article-title>. <source>A&#x26;A</source> <volume>628</volume>, <fpage>A129</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201935211</pub-id> </citation>
</ref>
<ref id="B27">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Lahav</surname>
<given-names>O.</given-names>
</name>
<name>
<surname>Liddle</surname>
<given-names>A. R.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>The Cosmological Parameters</article-title>. <comment>arXiv e-prints, arXiv:1912.03687</comment>. </citation>
</ref>
<ref id="B28">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Maas</surname>
<given-names>A. L.</given-names>
</name>
<name>
<surname>Hannun</surname>
<given-names>A. Y.</given-names>
</name>
<name>
<surname>Ng</surname>
<given-names>A. Y.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Rectifier Nonlinearities Improve Neural Network Acoustic Models</article-title>. <source>Proc. Icml (Citeseer)</source> <volume>30</volume>, <fpage>3</fpage>. <pub-id pub-id-type="doi">10.21437/interspeech.2016-1230</pub-id> </citation>
</ref>
<ref id="B29">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>McEwen</surname>
<given-names>J.&#x20;D.</given-names>
</name>
<name>
<surname>Wallis</surname>
<given-names>C. G. R.</given-names>
</name>
<name>
<surname>Mavor-Parker</surname>
<given-names>A. N.</given-names>
</name>
</person-group> (<year>2021</year>). <article-title>Scattering Networks on the Sphere for Scalable and Rotationally Equivariant Spherical CNNs</article-title>. <comment>arXiv e-prints, arXiv:2102.02828</comment>. </citation>
</ref>
<ref id="B30">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mirza</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Osindero</surname>
<given-names>S.</given-names>
</name>
</person-group> (<year>2014</year>). <article-title>Conditional Generative Adversarial Nets</article-title>. <comment>arXiv preprint arXiv:1411.1784</comment>. </citation>
</ref>
<ref id="B31">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Miyato</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Koyama</surname>
<given-names>M.</given-names>
</name>
</person-group> (<year>2018</year>). &#x201c;<article-title>cGANs with Projection Discriminator</article-title>,&#x201d; in <conf-name>International Conference on Learning Representations</conf-name>. </citation>
</ref>
<ref id="B32">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mustafa</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Bard</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Bhimji</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Al-Rfou</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Luki&#x107;</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Creating Virtual Universes Using Generative Adversarial Networks</article-title>. <comment>arXiv preprint arXiv:1706.02390</comment>. </citation>
</ref>
<ref id="B33">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Nathana&#xeb;l</surname>
<given-names>P.</given-names>
</name>
<name>
<surname>Ankit</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Hofmann</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>R&#xe9;fr&#xe9;gier</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Cosmological N-Body Simulations: A challenge for Scalable Generative Models</article-title>. <comment>arXiv preprint arXiv:1908.05519</comment>. </citation>
</ref>
<ref id="B34">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Odena</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Olah</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Shlens</surname>
<given-names>J.</given-names>
</name>
</person-group> (<year>2017</year>). &#x201c;<article-title>Conditional Image Synthesis with Auxiliary Classifier gans</article-title>,&#x201d; in <conf-name>Proceedings of the 34th International Conference on Machine Learning</conf-name>, <volume>70</volume>, <fpage>2642</fpage>&#x2013;<lpage>2651</lpage>. </citation>
</ref>
<ref id="B35">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Parroni</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Cardone</surname>
<given-names>V. F.</given-names>
</name>
<name>
<surname>Maoli</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Scaramella</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Going Deep with Minkowski Functionals of Convergence Maps</article-title>. <source>A&#x26;A</source> <volume>633</volume>, <fpage>A71</fpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/201935988</pub-id> </citation>
</ref>
<ref id="B36">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perarnau</surname>
<given-names>G.</given-names>
</name>
<name>
<surname>Van De Weijer</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Raducanu</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>&#xc1;lvarez</surname>
<given-names>J.&#x20;M.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Invertible Conditional gans for Image Editing</article-title>. <comment>arXiv preprint arXiv:1611.06355</comment>. </citation>
</ref>
<ref id="B37">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Perraudin</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Defferrard</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Sgier</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2018</year>). <article-title>Deepsphere: Efficient Spherical Convolutional Neural Network with Healpix Sampling for Cosmological Applications</article-title>. <comment>arXiv preprint arXiv:1810.12186</comment>. </citation>
</ref>
<ref id="B38">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Petri</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Haiman</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Hui</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>May</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Kratochvil</surname>
<given-names>J.&#x20;M.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Cosmology with Minkowski Functionals and Moments of the Weak Lensing Convergence Field</article-title>. <source>Phys. Rev. D</source> <volume>88</volume>, <fpage>123002</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevD.88.123002</pub-id> </citation>
</ref>
<ref id="B39">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Petri</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Haiman</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>May</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hui</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Kratochvil</surname>
<given-names>J.&#x20;M.</given-names>
</name>
</person-group> (<year>2015</year>). <article-title>Emulating the CFHTLenS Weak Lensing Data: Cosmological Constraints from Moments and Minkowski Functionals</article-title>. <source>Phys. Rev. D</source> <volume>91</volume>, <fpage>103511</fpage>. <pub-id pub-id-type="doi">10.1103/PhysRevD.91.103511</pub-id> </citation>
</ref>
<ref id="B40">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Petri</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Mocking the Weak Lensing Universe: The LensTools Python Computing Package</article-title>. <source>Astron. Comput.</source> <volume>17</volume>, <fpage>73</fpage>&#x2013;<lpage>79</lpage>. <pub-id pub-id-type="doi">10.1016/j.ascom.2016.06.001</pub-id> </citation>
</ref>
<ref id="B41">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pires</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Starck</surname>
<given-names>J.-L.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>R&#xe9;fr&#xe9;gier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Teyssier</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2009</year>). <article-title>Cosmological Model Discrimination with Weak Lensing</article-title>. <source>A&#x26;A</source> <volume>505</volume>, <fpage>969</fpage>&#x2013;<lpage>979</lpage>. <pub-id pub-id-type="doi">10.1051/0004-6361/200811459</pub-id> </citation>
</ref>
<ref id="B42">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Potter</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Stadel</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Teyssier</surname>
<given-names>R.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>PKDGRAV3: beyond Trillion Particle Cosmological Simulations for the Next Era of Galaxy Surveys</article-title>. <source>Comput. Astrophys.</source> <volume>4</volume>, <fpage>2</fpage>. <pub-id pub-id-type="doi">10.1186/s40668-017-0021-1</pub-id> </citation>
</ref>
<ref id="B43">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Reed</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Akata</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Yan</surname>
<given-names>X.</given-names>
</name>
<name>
<surname>Logeswaran</surname>
<given-names>L.</given-names>
</name>
<name>
<surname>Schiele</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Lee</surname>
<given-names>H.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Generative Adversarial Text to Image Synthesis</article-title>. <comment>arXiv preprint arXiv:1605.05396</comment>. </citation>
</ref>
<ref id="B44">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Refregier</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Weak Gravitational Lensing by Large-Scale Structure</article-title>. <source>Annu. Rev. Astron. Astrophys.</source> <volume>41</volume>, <fpage>645</fpage>&#x2013;<lpage>668</lpage>. <pub-id pub-id-type="doi">10.1146/annurev.astro.41.111302.102207</pub-id> </citation>
</ref>
<ref id="B45">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rodriguez</surname>
<given-names>A. C.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sgier</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Fluri</surname>
<given-names>J.</given-names>
</name>
<etal/>
</person-group> (<year>2018</year>). <article-title>Fast Cosmic Web Simulations with Generative Adversarial Networks</article-title>. <comment>arXiv preprint arXiv:1801.09070</comment>. </citation>
</ref>
<ref id="B46">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Roth</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Nowozin</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Hofmann</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2017</year>). <article-title>Stabilizing Training of Generative Adversarial Networks through Regularization</article-title>. <source>Adv. Neural Inform. Process. Syst.</source> <volume>13</volume>, <fpage>2018</fpage>&#x2013;<lpage>2028</lpage>. <pub-id pub-id-type="doi">10.21203/rs.2.22269/v1</pub-id> </citation>
</ref>
<ref id="B47">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Salimans</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Goodfellow</surname>
<given-names>I.</given-names>
</name>
<name>
<surname>Zaremba</surname>
<given-names>W.</given-names>
</name>
<name>
<surname>Cheung</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Radford</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Chen</surname>
<given-names>X.</given-names>
</name>
</person-group> (<year>2016</year>). <article-title>Improved Techniques for Training Gans</article-title>. <source>Adv. Neural Inform. Process. Syst.</source> <volume>4</volume> <fpage>2234</fpage>&#x2013;<lpage>2242</lpage>. <pub-id pub-id-type="doi">10.1117/12.2513139.6013937645001</pub-id> </citation>
</ref>
<ref id="B48">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sato</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Takada</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Hamana</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Matsubara</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2011</year>). <article-title>Simulations of Wide-Field Weak-Lensing Surveys. II. Covariance Matrix of Real-Space Correlation Functions</article-title>. <source>ApJ</source> <volume>734</volume>, <fpage>76</fpage>. <pub-id pub-id-type="doi">10.1088/0004-637X/734/2/76</pub-id> </citation>
</ref>
<ref id="B49">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Schmelzle</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Lucchi</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Sgier</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>R&#xe9;fr&#xe9;gier</surname>
<given-names>A.</given-names>
</name>
<etal/>
</person-group> (<year>2017</year>). <article-title>Cosmological Model Discrimination with Deep Learning</article-title>. <comment>arXiv preprint arXiv:1707.05167</comment>. </citation>
</ref>
<ref id="B50">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Sgier</surname>
<given-names>R. J.</given-names>
</name>
<name>
<surname>R&#xe9;fr&#xe9;gier</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Amara</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Nicola</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Fast Generation of Covariance Matrices for Weak Lensing</article-title>. <source>J.&#x20;Cosmol. Astropart. Phys.</source> <volume>2019</volume>, <fpage>044</fpage>. <pub-id pub-id-type="doi">10.1088/1475-7516/2019/01/044</pub-id> </citation>
</ref>
<ref id="B51">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Springel</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>White</surname>
<given-names>S. D. M.</given-names>
</name>
<name>
<surname>Jenkins</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Frenk</surname>
<given-names>C. S.</given-names>
</name>
<name>
<surname>Yoshida</surname>
<given-names>N.</given-names>
</name>
<name>
<surname>Gao</surname>
<given-names>L.</given-names>
</name>
<etal/>
</person-group> (<year>2005</year>). <article-title>Simulations of the Formation, Evolution and Clustering of Galaxies and Quasars</article-title>. <source>Nature</source> <volume>435</volume>, <fpage>629</fpage>&#x2013;<lpage>636</lpage>. <pub-id pub-id-type="doi">10.1038/nature03597</pub-id> </citation>
</ref>
<ref id="B52">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Szegedy</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Vanhoucke</surname>
<given-names>V.</given-names>
</name>
<name>
<surname>Ioffe</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Shlens</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Wojna</surname>
<given-names>Z.</given-names>
</name>
</person-group> (<year>2016</year>). &#x201c;<article-title>Rethinking the Inception Architecture for Computer Vision</article-title>,&#x201d; in <conf-name>Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</conf-name>, <fpage>2818</fpage>&#x2013;<lpage>2826</lpage>. </citation>
</ref>
<ref id="B53">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Takada</surname>
<given-names>M.</given-names>
</name>
<name>
<surname>Jain</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>The Three-point Correlation Function in Cosmology</article-title>. <source>MNRAS</source> <volume>340</volume>, <fpage>580</fpage>&#x2013;<lpage>608</lpage>. <pub-id pub-id-type="doi">10.1046/j.1365-8711.2003.06321.x</pub-id> </citation>
</ref>
<ref id="B54">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tamosiunas</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Winther</surname>
<given-names>H. A.</given-names>
</name>
<name>
<surname>Koyama</surname>
<given-names>K.</given-names>
</name>
<name>
<surname>Bacon</surname>
<given-names>D. J.</given-names>
</name>
<name>
<surname>Nichol</surname>
<given-names>R. C.</given-names>
</name>
<name>
<surname>Mawdsley</surname>
<given-names>B.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Investigating Cosmological GAN Emulators Using Latent Space Interpolation</article-title>. <comment>arXiv e-prints, arXiv:2004.10223</comment>. </citation>
</ref>
<ref id="B55">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Taylor</surname>
<given-names>A.</given-names>
</name>
<name>
<surname>Joachimi</surname>
<given-names>B.</given-names>
</name>
<name>
<surname>Kitching</surname>
<given-names>T.</given-names>
</name>
</person-group> (<year>2013</year>). <article-title>Putting the Precision in Precision Cosmology: How Accurate Should Your Data Covariance Matrix Be?</article-title>. <source>Month. Notices R. Astronomical Soc.</source> <volume>432</volume>, <fpage>1928</fpage>&#x2013;<lpage>1946</lpage>. <pub-id pub-id-type="doi">10.1093/mnras/stt270</pub-id> </citation>
</ref>
<ref id="B56">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tr&#xf6;ster</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Ferguson</surname>
<given-names>C.</given-names>
</name>
<name>
<surname>Harnois-D&#xe9;raps</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>McCarthy</surname>
<given-names>I. G.</given-names>
</name>
</person-group> (<year>2019</year>). <article-title>Painting with Baryons: Augmenting N-Body Simulations with Gas Using Deep Generative Models</article-title>. <source>MNRAS</source> <volume>487</volume>, <fpage>L24</fpage>&#x2013;<lpage>L29</lpage>. <pub-id pub-id-type="doi">10.1093/mnrasl/slz075</pub-id> </citation>
</ref>
<ref id="B57">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Van der Walt</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Sch&#xf6;nberger</surname>
<given-names>J.&#x20;L.</given-names>
</name>
<name>
<surname>Nunez-Iglesias</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Boulogne</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Warner</surname>
<given-names>J.&#x20;D.</given-names>
</name>
<name>
<surname>Yager</surname>
<given-names>N.</given-names>
</name>
<etal/>
</person-group> (<year>2014</year>). <article-title>Scikit-Image: Image Processing in python</article-title>. <source>PeerJ</source> <volume>2</volume>, <fpage>e453</fpage>. <pub-id pub-id-type="doi">10.7717/peerj.453</pub-id> </citation>
</ref>
<ref id="B58">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Villaescusa-Navarro</surname>
<given-names>F.</given-names>
</name>
<name>
<surname>Angl&#xe9;s-Alc&#xe1;zar</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Genel</surname>
<given-names>S.</given-names>
</name>
<name>
<surname>Spergel</surname>
<given-names>D. N.</given-names>
</name>
<name>
<surname>Somerville</surname>
<given-names>R. S.</given-names>
</name>
<name>
<surname>Dave</surname>
<given-names>R.</given-names>
</name>
<etal/>
</person-group> (<year>2020</year>). <article-title>The CAMELS Project: Cosmology and Astrophysics with MachinE Learning Simulations</article-title>. <comment>arXiv e-prints, arXiv:2010.00619</comment>. </citation>
</ref>
<ref id="B59">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname>
<given-names>Z.</given-names>
</name>
<name>
<surname>Simoncelli</surname>
<given-names>E. P.</given-names>
</name>
<name>
<surname>Bovik</surname>
<given-names>A. C.</given-names>
</name>
</person-group> (<year>2003</year>). <article-title>Multiscale Structural Similarity for Image Quality Assessment. <italic>The Thrity-Seventh Asilomar Conference On Signals</italic>
</article-title>. <source>Syst. Comput.</source> <volume>2</volume>, <fpage>1398</fpage>&#x2013;<lpage>1402</lpage>. <pub-id pub-id-type="doi">10.1109/acssc.2003.1292181</pub-id> </citation>
</ref>
<ref id="B60">
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Z&#xfc;rcher</surname>
<given-names>D.</given-names>
</name>
<name>
<surname>Fluri</surname>
<given-names>J.</given-names>
</name>
<name>
<surname>Sgier</surname>
<given-names>R.</given-names>
</name>
<name>
<surname>Kacprzak</surname>
<given-names>T.</given-names>
</name>
<name>
<surname>Refregier</surname>
<given-names>A.</given-names>
</name>
</person-group> (<year>2020</year>). <article-title>Cosmological Forecast for Non-gaussian Statistics in Large-Scale Weak Lensing Surveys</article-title>. <comment>arXiv e-prints, arXiv:2006.12506</comment>. </citation>
</ref>
</ref-list>
<app-group>
<app id="app1">
<title>Appendix A. Generative Adversarial Network Architecture</title>
<p>
<italic>Model</italic> <xref ref-type="table" rid="T1">Table A1</xref> summarizes the architecture of the GAN system, i.e. the generator and the discriminator. From the latent variable <italic>z</italic> and the cosmological parameters <inline-formula id="inf147">
<mml:math id="m152">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf148">
<mml:math id="m153">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x3a9;</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>, the generator starts by computing <inline-formula id="inf149">
<mml:math id="m154">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> using <xref ref-type="disp-formula" rid="e2">Eq. 2</xref>. As a second step, <inline-formula id="inf150">
<mml:math id="m155">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula> is transformed with three linear layers. i.e. a Multi Layer Perceptron (MLP), that outputs a 32768 tensor (<inline-formula id="inf151">
<mml:math id="m156">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula id="inf152">
<mml:math id="m157">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>). The data is then reshaped to <inline-formula id="inf153">
<mml:math id="m158">
<mml:mrow>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> (<inline-formula id="inf154">
<mml:math id="m159">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) and further transformed with four deconvolutional layers with stride 2 and kernel sizes of <inline-formula id="inf155">
<mml:math id="m160">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> or <inline-formula id="inf156">
<mml:math id="m161">
<mml:mrow>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> (<inline-formula id="inf157">
<mml:math id="m162">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> to <inline-formula id="inf158">
<mml:math id="m163">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>7</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>). The last generator layer consists of a deconvolution with stride 1 and kernel size <inline-formula id="inf159">
<mml:math id="m164">
<mml:mrow>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and it is intended to generate fine-grained details (<inline-formula id="inf160">
<mml:math id="m165">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>). The discriminator is symmetric to the generator with two exceptions. First the parameters <inline-formula id="inf161">
<mml:math id="m166">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf162">
<mml:math id="m167">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x3a9;</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> are concatenated in <inline-formula id="inf163">
<mml:math id="m168">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>5</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> just before the first linear layer. Second, an extra linear layer is added at the end of the discriminator (<inline-formula id="inf164">
<mml:math id="m169">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>9</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>) in order to recover a single output. All layers are separated by a LeakyRelu activation function with the parameter <inline-formula id="inf165">
<mml:math id="m170">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> (<xref ref-type="bibr" rid="B28">Maas et&#x20;al., 2013</xref>).</p>
<table-wrap id="T1" position="float">
<label>APPENDIX TABLE A1</label>
<caption>
<p>Conditional GAN architecture.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Layer</th>
<th align="center">Operation</th>
<th align="center">Activation</th>
<th align="center">Dimension</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td colspan="4" align="left">
<italic>Generator</italic>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf11">
<mml:math id="m13">
<mml:mrow>
<mml:mover accent="true">
<mml:mi>z</mml:mi>
<mml:mo>&#x5e;</mml:mo>
</mml:mover>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">
<xref ref-type="disp-formula" rid="e2">Eq. 2</xref>
</td>
<td align="left"/>
<td align="center">
<inline-formula id="inf12">
<mml:math id="m14">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf13">
<mml:math id="m15">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf14">
<mml:math id="m16">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf15">
<mml:math id="m17">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf16">
<mml:math id="m18">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf17">
<mml:math id="m19">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf18">
<mml:math id="m20">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32768</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf19">
<mml:math id="m21">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Reshape</td>
<td align="left"/>
<td align="center">
<inline-formula id="inf20">
<mml:math id="m22">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf21">
<mml:math id="m23">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Deconv (<inline-formula id="inf22">
<mml:math id="m24">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf23">
<mml:math id="m25">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf24">
<mml:math id="m26">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf25">
<mml:math id="m27">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>5</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Deconv (<inline-formula id="inf26">
<mml:math id="m28">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf27">
<mml:math id="m29">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf28">
<mml:math id="m30">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf29">
<mml:math id="m31">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>6</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Deconv (<inline-formula id="inf30">
<mml:math id="m32">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf31">
<mml:math id="m33">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf32">
<mml:math id="m34">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf33">
<mml:math id="m35">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>7</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Deconv (<inline-formula id="inf34">
<mml:math id="m36">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf35">
<mml:math id="m37">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf36">
<mml:math id="m38">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf37">
<mml:math id="m39">
<mml:mrow>
<mml:msub>
<mml:mi>g</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Deconv (<inline-formula id="inf38">
<mml:math id="m40">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf39">
<mml:math id="m41">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">Relu</td>
<td align="center">
<inline-formula id="inf40">
<mml:math id="m42">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td colspan="4" align="left">
<italic>Discriminator</italic>
</td>
</tr>
<tr>
<td align="left">
<italic>X</italic>
</td>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf41">
<mml:math id="m43">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf42">
<mml:math id="m44">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf43">
<mml:math id="m45">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf44">
<mml:math id="m46">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf45">
<mml:math id="m47">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf46">
<mml:math id="m48">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf47">
<mml:math id="m49">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf48">
<mml:math id="m50">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf49">
<mml:math id="m51">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf50">
<mml:math id="m52">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf51">
<mml:math id="m53">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf52">
<mml:math id="m54">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf53">
<mml:math id="m55">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf54">
<mml:math id="m56">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf55">
<mml:math id="m57">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf56">
<mml:math id="m58">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf57">
<mml:math id="m59">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf58">
<mml:math id="m60">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf59">
<mml:math id="m61">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf60">
<mml:math id="m62">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf61">
<mml:math id="m63">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf62">
<mml:math id="m64">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>5</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Reshape &#x2b; concatenate</td>
<td align="left"/>
<td align="center">
<inline-formula id="inf63">
<mml:math id="m65">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32770</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf64">
<mml:math id="m66">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>6</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf65">
<mml:math id="m67">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf66">
<mml:math id="m68">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>7</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf67">
<mml:math id="m69">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf68">
<mml:math id="m70">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf69">
<mml:math id="m71">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf70">
<mml:math id="m72">
<mml:mrow>
<mml:msub>
<mml:mi>d</mml:mi>
<mml:mn>9</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf71">
<mml:math id="m73">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>d<sub>5</sub> is a layer that reshapes the tensor to a vector and then concatenates the conditioning parameters to it. Here b is the batch size, k the convolutional kernel size and s the stride. The number of filters (convolution layer) and the number of neurons (linear layers) is shown in&#x20;blue.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>
<italic>Training</italic> The cosmological dataset described in <xref ref-type="sec" rid="s3">Section 3</xref> is used to train the GAN, where the batches are composed of samples from different cosmologies. We select a Wasserstein loss, with a gradient penalty of 10 (<xref ref-type="bibr" rid="B2">Arjovsky et&#x20;al., 2017</xref>). We use RMSProp as an optimizer with an initial learning rate of <inline-formula id="inf166">
<mml:math id="m171">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, and a batch size of 64. The discriminator is updated 5&#x20;times more often than the generator. The model is trained for <inline-formula id="inf167">
<mml:math id="m172">
<mml:mrow>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> epochs on a GeForce GTX 1080 GPU, which takes around 170&#xa0;h.</p>
</app>
<app id="app2">
<title>Appendix B. Regressor training</title>
<p>Given real and generated images, the general idea of the Frechet Inception Distance (FID) is to compute the distance between some of their complex statistics. For natural images, these statistics are given by the last layer, i.e. the logits, of a pre-trained Inception-V3 network (<xref ref-type="bibr" rid="B52">Szegedy et&#x20;al., 2016</xref>). As these statistics are meaningless for our cosmological data, we build new ones using a carefully designed regressor. Given an image, the regressor is trained to predict the two parameters <inline-formula id="inf168">
<mml:math id="m173">
<mml:mrow>
<mml:msub>
<mml:mtext>&#x3a9;</mml:mtext>
<mml:mi>m</mml:mi>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf169">
<mml:math id="m174">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>. We provide the regressor weights with the code to make our FID metric reusable.</p>
<p>
<italic>Data</italic> Naturally, we use the training dataset described in <xref ref-type="sec" rid="s3">Section 3</xref>, i.e. 46 different cosmologies composed by 12000 images each. This training dataset is further randomly split into a regressor training set (<inline-formula id="inf170">
<mml:math id="m175">
<mml:mrow>
<mml:mn>80</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula>) and a restressor test set (<inline-formula id="inf171">
<mml:math id="m176">
<mml:mrow>
<mml:mn>20</mml:mn>
<mml:mtext>%</mml:mtext>
</mml:mrow>
</mml:math>
</inline-formula>).</p>
<p>
<italic>Model</italic> The architecture of the regressor is described in <xref ref-type="table" rid="T2">Table A2</xref>. It shares the same structure as the GAN discriminator. It consists of a four convolutional layers followed by three linear layers with leaky relu non-linearity. The last layer is a linear layer with two outputs and it is responsible for producing the predicted parameters. We select the LeakyRelu activation functions for better gradient propagation.</p>
<table-wrap id="T2" position="float">
<label>APPENDIX TABLE A2</label>
<caption>
<p>Architecture of the regressor.</p>
</caption>
<table>
<thead valign="top">
<tr>
<th align="left">Layer</th>
<th align="center">Operation</th>
<th align="center">Activation</th>
<th align="center">Dimension</th>
</tr>
</thead>
<tbody valign="top">
<tr>
<td align="left">
<italic>X</italic>
</td>
<td align="left"/>
<td align="left"/>
<td align="center">
<inline-formula id="inf97">
<mml:math id="m102">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf98">
<mml:math id="m103">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>0</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf99">
<mml:math id="m104">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>7</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>7</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf100">
<mml:math id="m105">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf101">
<mml:math id="m106">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf102">
<mml:math id="m107">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf103">
<mml:math id="m108">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf104">
<mml:math id="m109">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf105">
<mml:math id="m110">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>64</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf106">
<mml:math id="m111">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf107">
<mml:math id="m112">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf108">
<mml:math id="m113">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf109">
<mml:math id="m114">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf110">
<mml:math id="m115">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>3</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf111">
<mml:math id="m116">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>5</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf112">
<mml:math id="m117">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf113">
<mml:math id="m118">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>16</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf114">
<mml:math id="m119">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>4</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">conv (<inline-formula id="inf115">
<mml:math id="m120">
<mml:mrow>
<mml:mi>k</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>3</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>3</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf116">
<mml:math id="m121">
<mml:mrow>
<mml:mi>s</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>)</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf117">
<mml:math id="m122">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>8</mml:mn>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf118">
<mml:math id="m123">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>5</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Reshape</td>
<td align="left"/>
<td align="center">
<inline-formula id="inf119">
<mml:math id="m124">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>32768</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf120">
<mml:math id="m125">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>6</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf121">
<mml:math id="m126">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>512</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf122">
<mml:math id="m127">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>7</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf123">
<mml:math id="m128">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>256</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf124">
<mml:math id="m129">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>8</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">LeakyRelu</td>
<td align="center">
<inline-formula id="inf125">
<mml:math id="m130">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>128</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
<tr>
<td align="left">
<inline-formula id="inf126">
<mml:math id="m131">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>9</mml:mn>
</mml:msub>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
<td align="center">Linear</td>
<td align="center">Linear</td>
<td align="center">
<inline-formula id="inf127">
<mml:math id="m132">
<mml:mrow>
<mml:mi>b</mml:mi>
<mml:mo>&#xd7;</mml:mo>
<mml:mn>2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>
</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>Here b is the batch size, k the convolutional kernel size and s the stride. The number of filters (convolution layer) and the number of neurons (linear layers) is shown in blue. The LeakyRelu activation uses the parameter <inline-formula id="inf128">
<mml:math id="m133">
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.2</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>
<italic>Training</italic> We use the mean squared error between the predicted and true parameters as a loss function. The model was trained for 20 epochs using an Adam (<xref ref-type="bibr" rid="B23">Kingma and Ba, 2014</xref>) optimizer with an initial learning rate of <inline-formula id="inf172">
<mml:math id="m177">
<mml:mrow>
<mml:mn>3</mml:mn>
<mml:mo>&#x22c5;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>5</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf173">
<mml:math id="m178">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.9</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula>, <inline-formula id="inf174">
<mml:math id="m179">
<mml:mrow>
<mml:msub>
<mml:mi>&#x3b2;</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:mo>&#x3d;</mml:mo>
<mml:mn>0.999</mml:mn>
</mml:mrow>
</mml:math>
</inline-formula> and <inline-formula id="inf175">
<mml:math id="m180">
<mml:mrow>
<mml:mi mathvariant="italic">&#x3f5;</mml:mi>
<mml:mo>&#x3d;</mml:mo>
<mml:msup>
<mml:mrow>
<mml:mn>10</mml:mn>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>8</mml:mn>
</mml:mrow>
</mml:msup>
</mml:mrow>
</mml:math>
</inline-formula> and a batch size of 64. The mean squared error evaluated on the test set corresponds to 8.93<italic>e</italic>
<sup>&#x2212;5</sup>, which is low enough for the purpose of computing the&#x20;FID.</p>
</app>
</app-group>
</back>
</article>