<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.3 20070202//EN" "journalpublishing.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="brief-report" dtd-version="2.3" xml:lang="EN">
<front>
<journal-meta>
<journal-id journal-id-type="publisher-id">Front. Psychiatry</journal-id>
<journal-title>Frontiers in Psychiatry</journal-title>
<abbrev-journal-title abbrev-type="pubmed">Front. Psychiatry</abbrev-journal-title>
<issn pub-type="epub">1664-0640</issn>
<publisher>
<publisher-name>Frontiers Media S.A.</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.3389/fpsyt.2024.1395243</article-id>
<article-categories>
<subj-group subj-group-type="heading">
<subject>Psychiatry</subject>
<subj-group>
<subject>Brief Research Report</subject>
</subj-group>
</subj-group>
</article-categories>
<title-group>
<article-title>Generation and discrimination of autism MRI images based on autoencoder</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname>Shi</surname>
<given-names>Yuxin</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2671963"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
<role content-type="https://credit.niso.org/contributor-roles/investigation/"/>
<role content-type="https://credit.niso.org/contributor-roles/data-curation/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Gong</surname>
<given-names>Yongli</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Guan</surname>
<given-names>Yurong</given-names>
</name>
<xref ref-type="aff" rid="aff2">
<sup>2</sup>
</xref>
<xref ref-type="author-notes" rid="fn001">
<sup>*</sup>
</xref>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Tang</surname>
<given-names>Jiawei</given-names>
</name>
<xref ref-type="aff" rid="aff1">
<sup>1</sup>
</xref>
<uri xlink:href="https://loop.frontiersin.org/people/2861894/overview"/>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/"/>
<role content-type="https://credit.niso.org/contributor-roles/methodology/"/>
</contrib>
</contrib-group>
<aff id="aff1">
<sup>1</sup>
<institution>Computer School, Hubei Polytechnic University</institution>, <addr-line>Huangshi</addr-line>, <country>China</country>
</aff>
<aff id="aff2">
<sup>2</sup>
<institution>College of Computer Science, Huanggang Normal University</institution>, <addr-line>Huanggang</addr-line>, <country>China</country>
</aff>
<author-notes>
<fn fn-type="edited-by">
<p>Edited by: Chang Cai, University of California, San Francisco, United States</p>
</fn>
<fn fn-type="edited-by">
<p>Reviewed by: Guangshuai Wang, Central China Normal University, China</p>
<p>Zhihong Sun, Naval Engineering University, China</p>
</fn>
<fn fn-type="corresp" id="fn001">
<p>*Correspondence: Yongli Gong, <email xlink:href="mailto:2468323287@qq.com">2468323287@qq.com</email>; Yurong Guan, <email xlink:href="mailto:jsjgyr@hgnu.edu.cn">jsjgyr@hgnu.edu.cn</email>
</p>
</fn>
</author-notes>
<pub-date pub-type="epub">
<day>14</day>
<month>10</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>15</volume>
<elocation-id>1395243</elocation-id>
<history>
<date date-type="received">
<day>03</day>
<month>03</month>
<year>2024</year>
</date>
<date date-type="accepted">
<day>16</day>
<month>09</month>
<year>2024</year>
</date>
</history>
<permissions>
<copyright-statement>Copyright &#xa9; 2024 Shi, Gong, Guan and Tang</copyright-statement>
<copyright-year>2024</copyright-year>
<copyright-holder>Shi, Gong, Guan and Tang</copyright-holder>
<license xlink:href="http://creativecommons.org/licenses/by/4.0/">
<p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.</p>
</license>
</permissions>
<abstract>
<p>This study aims to explore an autoencoder-based method for generating brain MRI images of patients with Autism Spectrum Disorder (ASD) and non-ASD individuals, and to discriminate ASD based on the generated images. Initially, we introduce the research background of ASD and related work, as well as the application of deep learning in the field of medical imaging. Subsequently, we detail the architecture and training process of the proposed autoencoder model, and present the results of generating MRI images for ASD and non-ASD patients. Following this, we designed an ASD classifier based on the generated images and elucidated its structure and training methods. Finally, through analysis and discussion of experimental results, we validated the effectiveness of the proposed method and explored future research directions and potential clinical applications. This research offers new insights and methodologies for addressing challenges in ASD studies using deep learning technology, potentially contributing to the automated diagnosis and research of ASD.</p>
</abstract>
<kwd-group>
<kwd>deep autoencoder</kwd>
<kwd>sMRI</kwd>
<kwd>image generation</kwd>
<kwd>image classification</kwd>
<kwd>ASD</kwd>
</kwd-group>
<counts>
<fig-count count="3"/>
<table-count count="1"/>
<equation-count count="6"/>
<ref-count count="26"/>
<page-count count="8"/>
<word-count count="4259"/>
</counts>
<custom-meta-wrap>
<custom-meta>
<meta-name>section-in-acceptance</meta-name>
<meta-value>Computational Psychiatry</meta-value>
</custom-meta>
</custom-meta-wrap>
</article-meta>
</front>
<body>
<sec id="s1" sec-type="intro">
<label>1</label>
<title>Introduction</title>
<p>Autism Spectrum Disorder (ASD) is a neurodevelopmental disorder characterized by deficits in social interaction and communication, as well as restricted behaviors. Over the past decades, with the advancement of medical imaging technology, researchers have begun utilizing Magnetic Resonance Imaging (MRI) to study the differences in brain structure and functional connectivity in individuals with ASD. However, due to the high heterogeneity and complexity of ASD, relying solely on expert manual analysis of MRI images has its limitations. Thus, the use of deep learning technologies for the analysis and diagnosis of ASD has become increasingly important.</p>
<p>As machine learning and deep learning continue to evolve, significant progress has been made in the analysis and diagnosis of ASD. For instance, Support Vector Machines (SVM) have been widely used and studied in brain medical image classification. Early research by Katuwal et&#xa0;al. (<xref ref-type="bibr" rid="B1">1</xref>) explored SVM classifiers using MRI data to detect ASD, demonstrating that classification accuracy increases with the severity of autism. Beyond SVM, other machine learning methods have also achieved notable successes in ASD analysis and diagnostics. Bhaumik et&#xa0;al. (<xref ref-type="bibr" rid="B2">2</xref>) used&#xa0;Partial Least Squares Regression (PLS) based on features obtained from 42 bilateral Brodmann areas to distinguish between ASD patients and healthy controls. Donato et&#xa0;al. (<xref ref-type="bibr" rid="B3">3</xref>) employed Bayesian methods for reverse inference studies, finding probabilistic evidence of gray matter changes selective to ASD in specific brain regions.</p>
<p>Compared to machine learning, deep learning can learn high-level features from data without the need for expert identification (<xref ref-type="bibr" rid="B4">4</xref>). The applications of deep learning in ASD diagnostics and analysis can be broadly divided into three categories: methods based on Convolutional Neural Networks (CNNs), Generative Adversarial Networks (GANs), and autoencoders. CNN methods often employ various CNN variants or increase the depth of CNNs to handle the high dimensionality of MRI data, making them suitable for image classification tasks in deep learning. Zhao et&#xa0;al. (<xref ref-type="bibr" rid="B5">5</xref>) proposed a 3D CNN framework to capture the overlapping spatial brain network patterns between ASD and healthy controls, proving it to be effective. Dolz et&#xa0;al. (<xref ref-type="bibr" rid="B6">6</xref>) used a 3D CNN ensemble model with internally annotated images for 3D segmentation of infant T1-weighted and T2-weighted MRI images, achieving notable performance. Recently, Yin et&#xa0;al. (<xref ref-type="bibr" rid="B7">7</xref>) combined autoencoders with Deep Neural Networks (DNN), achieving 82.4% AUC and 79.2% accuracy. While CNNs effectively leverage spatial structural information across the entire brain, they have limitations in handling spatial coordinate transformation-related issues (<xref ref-type="bibr" rid="B8">8</xref>). Therefore, in ASD image generation tasks, integrating CNNs with other models is preferred.</p>
<p>GANs, as generative models, utilize game theory methods to generate images, thereby enhancing ASD datasets. Yang et&#xa0;al. (<xref ref-type="bibr" rid="B9">9</xref>) successfully improved classifier performance by using FC-GAN to augment ASD datasets. Recently, Devika et&#xa0;al. (<xref ref-type="bibr" rid="B10">10</xref>) employed a GAN-based encoder-decoder framework using consecutive sMRI slices as input, where the error between the reconstructed and actual adjacent slice stacks was used to identify ASD samples as outliers, yielding effective results. GANs learn the distribution of real data to generate lifelike images, and their generators can capture multiple data patterns, adding diversity to the generated images. However, training GAN models can be challenging due to the need to achieve Nash equilibrium (<xref ref-type="bibr" rid="B11">11</xref>).</p>
<p>Meanwhile, the accuracy of MRI image diagnostics depends on the quality of the images, which is often compromised by noise and artifacts (<xref ref-type="bibr" rid="B12">12</xref>). Thus, MRI image generation methods that also assist in noise reduction can significantly enhance image quality. For image denoising, autoencoders are generally more suitable than GANs.</p>
<p>For high-dimensional data image reconstruction (<xref ref-type="bibr" rid="B13">13</xref>) and noise removal (<xref ref-type="bibr" rid="B14">14</xref>), autoencoders are an excellent choice. As a deep learning model, autoencoders learn data features to reconstruct data. Their advantages include the ability to learn data representations that capture essential features of the data while minimizing redundancy, reducing the dimensionality of the data to make it more manageable, and removing noise from images to enhance clarity and accuracy (<xref ref-type="bibr" rid="B15">15</xref>). Autoencoders minimize the difference between input and reconstructed output during training, attempting to remove noise or unnecessary information from input images, thus typically providing a more stable model training process. Given the high dimensionality of neuroimaging data and the complexity involved in manually creating features from such data, the applications of autoencoders in brain disease research are of significant relevance (<xref ref-type="bibr" rid="B16">16</xref>).</p>
<p>Autoencoders have been used for MRI data generation and reconstruction, yielding positive research results. For example, Pinaya et&#xa0;al. (<xref ref-type="bibr" rid="B17">17</xref>) first trained a deep autoencoder model using data from healthy subjects, then estimated the overall and regional neuroanatomical deviations in ASD patients using the ASD dataset. This model provides a flexible framework for assessing overall and regional neuroanatomical deviations in neuropsychiatric populations. Li et&#xa0;al. (<xref ref-type="bibr" rid="B18">18</xref>), based on transfer learning techniques, proposed a novel deep neural network framework aimed at improving the ASD classification problem in situations with limited training samples. This framework first trained an SSAE using healthy subject data, which was then transferred to a new classification task for target subjects to achieve more accurate classification. Eslami et&#xa0;al. (<xref ref-type="bibr" rid="B19">19</xref>) proposed an Auto-ASD-Network model that combines the advantages of autoencoders and SVM, aimed at classifying ASD patients from healthy individuals using fMRI data, significantly enhancing classification results. Most recently, Mostafa (<xref ref-type="bibr" rid="B20">20</xref>) proposed using a CAE model to reconstruct images of ASD and healthy control groups. This model first processes 3D T1 images through slice analysis and classification, then inputs the slice images into an autoencoder for image reconstruction, using various similarity metrics to measure the resemblance between actual and reconstructed images, achieving high accuracy. These studies show that autoencoders can extract key features from MRI data, aiding medical experts in better understanding the brain structures and functions of ASD patients, providing support for early diagnosis and personalized treatment of the disorder. These studies demonstrate that, within ASD research, autoencoders are capable of extracting crucial features from MRI data, aiding medical experts in better understanding the brain structure and function of ASD patients. This facilitates early diagnosis and personalized treatment of the disorder.</p>    <p>Given this, we will detail the architecture and training process of the proposed autoencoder model, as well as how to use this model to generate MRI images of ASD and non-ASD patients. Next, we will introduce the structure and training methods of our designed ASD classifier, and describe in detail how to use the generated MRI images for ASD discrimination. Finally, we will analyze the experimental results and discuss future research directions and potential clinical applications. Specific research contributions include:</p>
<list list-type="order">
<list-item>
<p>Autoencoder-generated MRI images of ASD and non-ASD patients: By designing and training the autoencoder model, this paper successfully generates MRI images for ASD and non-ASD patients. This method not only expands the dataset and increases sample diversity but also helps understand the characteristic brain structure features and differences in ASD patients. This is significant for ASD research in situations with limited sample sizes and high data heterogeneity.</p>
</list-item>
<list-item>
<p>ASD discrimination method based on generated images: Using the generated MRI images, this paper proposes a novel ASD classifier for discriminating patients with ASD. Compared to traditional methods based on original MRI images, this image generation-based approach offers more universality and effectively addresses issues of data sparsity and label imbalance. By combining autoencoder generation and deep learning classification techniques, this paper provides a new approach and method for the automation of ASD diagnosis.</p>
</list-item>
</list>
</sec>
<sec id="s2">
<label>2</label>
<title>Method</title>
<p>Our algorithm primarily operates by learning an autoencoder that takes MRI images as input. This autoencoder is then used for reconstruction, and the reconstructed MRI images are fed into a convolutional neural network, where a classifier determines the clinical validity of the images. This section analyzes both the autoencoder model (see Section 2.1) and the classifier model (see Section 2.2) proposed by our method.</p>
<sec id="s2_1">
<label>2.1</label>
<title>Autoencoder model architecture</title>
<sec id="s2_1_1">
<label>2.1.1</label>
<title>Model architecture</title>
<p>The autoencoder is composed of an encoder and a decoder. <xref ref-type="fig" rid="f1">
<bold>Figure&#xa0;1</bold>
</xref> displays the structure of this one-dimensional autoencoder, which utilizes minimal encoders, decoders, and hidden layers to simplify the model while achieving high reconstruction performance. The autoencoder starts with an input layer, followed by the encoder, which transforms the input data into feature vectors through two fully connected layers. Specifically, the encoder contains two ReLU-activated dense layers with 4 and 2 units, respectively, which compress the input data into a compact feature representation. An intermediate fully connected layer is situated between the encoder and decoder. One fully connected layer is located between the encoder and decoder (model hyperparameters optimized using Bayesian hyperparameter optimization algorithm). The intermediate hidden layer adopts the settings from the literature (<xref ref-type="bibr" rid="B21">21</xref>), and the data serves as input for the decoder. The decoder mirrors the encoder&#x2019;s structure, consisting of two ReLU-activated dense layers with 2 and 4 units, respectively. The final layer of the decoder features a tanh activation function and has an output size of (8000000), which corresponds to the flattened three-dimensional image data. This symmetry between the encoder and decoder layers facilitates accurate reconstruction of the original input data.</p>
<fig id="f1" position="float">
<label>Figure&#xa0;1</label>
<caption>
<p>One-dimensional autoencoder architecture.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-15-1395243-g001.tif"/>
</fig>
<p>The overall model adopts a symmetrical structure, with the output of each layer serving as the input for the next, thereby facilitating layer-by-layer training and effective learning of data representations. For training the autoencoder, we utilized the Adam optimizer, with a learning rate of 0.001 and an epsilon value of 1e-08. The Adam optimizer is well-suited for training large-scale and high-dimensional data due to its adaptive learning rate capabilities, which adjust based on gradient statistics and improve convergence during the training process.</p>
</sec>
<sec id="s2_1_2">
<label>2.1.2</label>
<title>Algorithm overview</title>
<p>Assuming the input data is y and the output data is x, the encoding and decoding process of the autoencoder can be described as (<xref ref-type="bibr" rid="B22">22</xref>):</p>
<disp-formula id="eq1">
<label>(1)</label>
<mml:math display="block" id="M1">
<mml:mrow>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>e</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mi>x</mml:mi>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq2">
<label>(2)</label>
<mml:math display="block" id="M2">
<mml:mrow>
<mml:mtext>y</mml:mtext>
<mml:mo>=</mml:mo>
<mml:msub>
<mml:mi>&#x3c3;</mml:mi>
<mml:mi>d</mml:mi>
</mml:msub>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:msub>
<mml:mi>W</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
<mml:msub>
<mml:mi>h</mml:mi>
<mml:mn>1</mml:mn>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>b</mml:mi>
<mml:mn>2</mml:mn>
</mml:msub>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Where <italic>W</italic>
<sub>1</sub> and <italic>b</italic>
<sub>1</sub> represent the encoding weights and biases, <italic>W</italic>
<sub>2</sub> and <italic>b</italic>
<sub>2</sub> represent the decoding weights and biases, <italic>&#x3c3;<sub>e</sub>
</italic> is a non-linear transformation, and <italic>&#x3c3;<sub>d</sub>
</italic> is the same non-linear transformation used during the encoding process.</p>
</sec>
<sec id="s2_1_3">
<label>2.1.3</label>
<title>Loss function</title>
<p>The loss function is calculated using mean squared error loss:</p>
<disp-formula id="eq3">
<label>(3)</label>
<mml:math display="block" id="M3">
<mml:mrow>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mtext>b</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>L</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>x</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mtext>y</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mo>&#x2211;</mml:mo>
<mml:msubsup>
<mml:mrow>
<mml:mrow>
<mml:mo>|</mml:mo>
<mml:mo>|</mml:mo>
<mml:mrow>
<mml:mtext>y</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mtext>x</mml:mtext>
</mml:mrow>
<mml:mo>|</mml:mo>
<mml:mo>|</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mn>2</mml:mn>
<mml:mn>2</mml:mn>
</mml:msubsup>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Using gradient descent, the backpropagation algorithm updates parameters to minimize the error function. The update formulas for the weights <italic>W</italic> and biases <italic>b</italic> are as follows, where <italic>v</italic> is the learning rate:</p>
<disp-formula id="eq4">
<label>(4)</label>
<mml:math display="block" id="M4">
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mtext>W</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mtext>b</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:mtext>W</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<disp-formula id="eq5">
<label>(5)</label>
<mml:math display="block" id="M5">
<mml:mrow>
<mml:mtext>b</mml:mtext>
<mml:mo>=</mml:mo>
<mml:mtext>b</mml:mtext>
<mml:mo>&#x2212;</mml:mo>
<mml:mi>v</mml:mi>
<mml:mfrac>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:mi>J</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mtext>W</mml:mtext>
<mml:mo>,</mml:mo>
<mml:mtext>b</mml:mtext>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
</mml:mrow>
<mml:mrow>
<mml:mo>&#x2202;</mml:mo>
<mml:mtext>b</mml:mtext>
</mml:mrow>
</mml:mfrac>
</mml:mrow>
</mml:math>
</disp-formula>
<p>For images, especially medical images that contain detailed pathological information, it is necessary to extract this information from the images as a whole. Since autoencoders learn holistic representations and can effectively extract high-level abstract features, they have the capability to generate brain MRI images as needed.</p>
</sec>
</sec>
<sec id="s2_2">
<label>2.2</label>
<title>CNN-based MRI classification discrimination model architecture</title>
<p>In this study, MRI identification is a binary classification problem, aimed at distinguishing whether a three-dimensional MRI belongs to a diseased (label: 1) or healthy (label: 0) category. An MRI dataset is reshaped into an array Tm&#xd7;a&#xd7;b (in this study, 128&#xd7;128&#xd7;128), serving as the input for the 3D CNN.</p>
<p>The designed convolutional neural network structure of our algorithm (partially detailed in <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Materials</bold>
</xref>) consists of convolutional layers, pooling layers, activation function layers, and fully connected layers, each performing specific functions. It begins with initial layers employing 3x3x3 filters: 2 filters in the first layer and 8 filters in subsequent layers with a stride of 2 for downsampling. Each convolutional layer uses ReLU activation. Batch normalization is applied for stability after select convolutional layers. Average pooling with a 2x2x2 kernel and stride of 2 follows, reducing spatial dimensions while retaining important features. Deeper layers increase complexity with 16, 32, 64, and 128 filters, maintaining spatial resolution. All use ReLU activation and batch normalization. Output from convolutional layers is flattened for fully connected layers. Two dense layers follow with 128 and 64 neurons, using ReLU and tanh activation. Dropout with a 0.5 rate is applied after each dense layer to prevent overfitting. The final layer consists of 2 neurons with softmax activation for binary classification. Adam optimizer minimizes binary cross-entropy loss, assessing accuracy metrics. The final output of the 3D CNN is the probability values for the two classification labels, summing to one.</p>
<p>This study uses the Adam optimizer to optimize the 3D CNN, with a learning rate set at 0.00001 and includes a learning rate decay callback function, ReduceLROnPlateau, which dynamically reduces the learning rate. The initialization strategy follows the settings described in the referenced literature (<xref ref-type="bibr" rid="B23">23</xref>). Post-training, the model reports performance on the validation set (or new MRI).</p>
</sec>
</sec>
<sec id="s3" sec-type="results">
<label>3</label>
<title>Results</title>
<p>This section delineates the experimental framework and dataset used (refer to Section 3.1). We carried out a series of experiments to assess the efficacy of the proposed method through both qualitative (Section 3.2) and quantitative analyses (Section 3.3) of the results. Our testing setup was configured on a Windows 10 system, employing Python 3.6 and TensorFlow 2.5, with GPU acceleration provided by an NVIDIA GeForce GTX 1650 graphics card, which has 4.0GB of GPU memory. The source code and implementation details are publicly available at (<ext-link ext-link-type="uri" xlink:href="https://github.com/Lyrae17514/Generation-and-Discrimination-of-Autism-MRI-Images-Based-on-Autoencoder">https://github.com/Lyrae17514/Generation-and-Discrimination-of-Autism-MRI-Images-Based-on-Autoencoder</ext-link>).</p>
<sec id="s3_1">
<label>3.1</label>
<title>Data description</title>
<p>The ABIDE I dataset, consisting of data from 17 international sites and encompassing 1112 subjects, was utilized in this study. Specifically, we employed data from 871 ABIDE I subjects that underwent quality checks, aligning with the dataset used by Mostafa et&#xa0;al. (<xref ref-type="bibr" rid="B20">20</xref>). This subset includes 403 subjects diagnosed with Autism Spectrum Disorder (ASD) and 468 from a typical control group (<xref ref-type="bibr" rid="B24">24</xref>). The subjects ranged from 7 to 64 years in age, and our study focused exclusively on the T1 dataset.</p>
</sec>
<sec id="s3_2">
<label>3.2</label>
<title>Qualitative analysis</title>
<sec id="s3_2_1">
<label>3.2.1</label>
<title>Generative performance evaluation experiment</title>
<p>This experiment aimed to appraise the image reconstruction capabilities of the autoencoder. We assessed the image reconstruction efficacy of the autoencoder by inputting images of various sizes and comparing the output to the original images. Given the inclusion of critical pathological information within medical images, it is crucial to enhance the fidelity between the reconstructed and original images to preserve essential pathological structures. <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2A</bold>
</xref> exhibits a comparative analysis of reconstructed images with dimensions of 64<sup>3</sup>, 128<sup>3</sup>, and 200<sup>3</sup> against the originals, clearly demonstrating the autoencoder&#x2019;s ability to accurately replicate details. <xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2B</bold>
</xref> further presents various performance metrics associated with the reconstructed images.</p>
<fig id="f2" position="float">
<label>Figure&#xa0;2</label>
<caption>
<p>Visualization of the reconstructed image performance of the model. <bold>(A)</bold> Visualization of the generated image results. <bold>(B)</bold> Image reconstruction performance of ASD and HC. ASD represents the use of data from patients with ASD, while HC represents the healthy control group.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-15-1395243-g002.tif"/>
</fig>
<p>The data for this experiment comprised 3D data sourced from ANTS. The autoencoder employed mean squared error (MSE) as the loss function. Given the constraints imposed by the limited GPU memory capacity, our training methodology had to be specifically tailored to handle high-dimensional MRI data within these hardware limitations. The MRI data sets, characterized by their substantial memory footprint, necessitated a training approach that accommodates the GPU&#x2019;s memory constraints. Consequently, the generator function was designed to dynamically select a single MRI image from the dataset at a time, resulting in a batch size of 1 for each training iteration. However, it also poses challenges such as increased training variability and slower convergence rates. To mitigate these issues and enhance training efficiency, the model employs the Adam optimizer, with a learning rate set at 0.001, known for its adaptive learning rate adjustments, which help in navigating the stochastic nature of the training process and improving convergence over extensive epochs. For instance, when generating MRI images of size 200<sup>3</sup>, the autoencoder can achieve its best SSIM score upon reaching 68000 epochs. When the autoencoder processes a 3D matrix of size 128<sup>3</sup>, it is flattened into a 1D array of length (2097152) as input. Both the encoding and decoding phases featured two fully connected layers, with neuron configurations of 4 and 2 (encoding) and 2 and 4 (decoding) respectively.</p>
<p>In the comparison between generated image results and real images (<xref ref-type="fig" rid="f2">
<bold>Figure&#xa0;2</bold>
</xref>), the generated images show no significant distortions or artifacts, effectively removing noise. This indicates that the model fits the data well without overfitting. Specifically, this means that the model demonstrates good adaptability to the training data. The reconstruction performance is high for images of various sizes, demonstrating the model&#x2019;s robustness. Although some details in the generated images might appear slightly blurry, the overall quality remains high, likely reflecting the intrinsic characteristics of the autoencoder. Moreover, high values of Structural Similarity Index (SSIM) and Peak Signal-to-Noise Ratio (PSNR), coupled with low Mean Absolute Error (MAE), corroborate the autoencoder&#x2019;s ability to generate high-quality brain MRI images.</p>
</sec>
<sec id="s3_2_2">
<label>3.2.2</label>
<title>Classification performance evaluation experiment</title>
<p>The ROC curve is used to evaluate the classification and detection results. The curve closer to the top-left corner indicates a higher true positive rate and lower false positive rate, suggesting better classification ability of the model. The area under the ROC curve (AUC) is used to measure the performance. The classifier is validated on the validation set, and the ROC curve is plotted. The model has an average accuracy of 58.85%, with an accuracy of 63% in detecting healthy samples. The calculated AUC value is 0.60, indicating relatively good performance of the binary classifier with some predictive capability. Overall, this classifier can effectively screen ASD patients and healthy controls to a certain extent.</p>
</sec>
</sec>
<sec id="s3_3">
<label>3.3</label>
<title>Quantitative analysis</title>
<p>In this section, we evaluate the reconstruction effectiveness of the quantized autoencoder and compare it with other models using the Structural Similarity Index (SSIM) (<xref ref-type="bibr" rid="B25">25</xref>) and Peak Signal-to-Noise Ratio (PSNR) to assess the generated images. <xref ref-type="table" rid="T1">
<bold>Table&#xa0;1</bold>
</xref> presents the quantitative results of these evaluation metrics, with some model results referenced from Mostafa (<xref ref-type="bibr" rid="B20">20</xref>). The model studied in this paper has relatively high SSIM scores, indicating that the distribution of the generated samples closely resembles the distribution of real data. For the 128<sup>2</sup> resolution, the model in this study achieved excellent SSIM scores, and it also performed well at a resolution of 256<sup>2</sup>. The generated images have high SSIM and PSNR values, as well as low MAE (Mean Absolute Error), showcasing excellent image quality and accurate pixel matching.</p>
<table-wrap id="T1" position="float">
<label>Table&#xa0;1</label>
<caption>
<p>Comparison of quantitative results.</p>
</caption>
<table frame="hsides">
<thead>
<tr>
<th valign="top" align="center"/>
<th valign="middle" align="center">SSIM&#x2191;</th>
<th valign="middle" align="center">PSNR&#x2191;</th>
<th valign="middle" align="center">MSE&#x2193;</th>
</tr>
</thead>
<tbody>
<tr>
<td valign="bottom" align="center">CAE</td>
<td valign="bottom" align="center">0.22</td>
<td valign="bottom" align="center">0.79</td>
<td valign="bottom" align="center">0.21</td>
</tr>
<tr>
<td valign="top" align="center">128<sup>2</sup> Ours</td>
<td valign="top" align="center">0.58</td>
<td valign="top" align="center">19.42</td>
<td valign="top" align="center">0.01</td>
</tr>
<tr>
<td valign="top" align="center">256<sup>2</sup> Ours</td>
<td valign="top" align="center">0.62</td>
<td valign="top" align="center">19.91</td>
<td valign="top" align="center">0.01</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p>The upward arrow indicates that a higher score is better, meaning the image is closer to the real image. Conversely, the downward arrow signifies that a lower score is preferable, indicating that the image error is smaller.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p>To ensure a fair comparison with Mostafa&#x2019;s proposed CAE model, we maintained consistency in data processing, using T1-weighted MRI slice images of the healthy control group for autoencoder reconstruction (<xref ref-type="bibr" rid="B26">26</xref>). When generating images at sizes of 256<sup>2</sup> and 128<sup>2</sup>, the autoencoder adjusted the size of the input and output images and only modified the structure of the intermediate hidden layer of the autoencoder.</p>
<p>Compared to Mostafa&#x2019;s proposed CAE model, the autoencoder proposed in this paper exhibits superior SSIM scores at a size of 128<sup>2</sup>. When generating larger images at a size of 256<sup>2</sup>, the PSNR metric still performs relatively well. The higher SSIM and PSNR scores, along with the lower MSE, indicate that the generated images preserve more structural details and have lower noise levels. This suggests that our autoencoder produces reconstructed images with higher clarity and quality while preserving the original information. Therefore, this experiment demonstrates that our autoencoder model generates images that are closer to the original images, exhibiting good quality and excellent generalization capabilities.</p>
</sec>
</sec>
<sec id="s4" sec-type="discussion">
<label>4</label>
<title>Discussion</title>
<p>First, this section analyzed the computational complexity of the proposed method (see section 4.1). Second, the generator interpretation was discussed in section 4.2. Third, the influence of optimizers on performance was discussed (see Section 4.3).</p>
<sec id="s4_1">
<label>4.1</label>
<title>Computational complexity</title>
<p>The autoencoder proposed in this study is based on a fully connected neural network structure. The complexity of this model can be estimated by calculating the number of parameters, which is proportional to the number of connection weights between the input and output layers, as well as the number of bias terms. The time complexity calculation is as follows:</p>
<disp-formula id="eq6">
<label>(6)</label>
<mml:math display="block" id="M6">
<mml:mrow>
<mml:mi>O</mml:mi>
<mml:mrow>
<mml:mo stretchy="false">(</mml:mo>
<mml:mrow>
<mml:mi>U</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:mrow>
<mml:mo stretchy="false">)</mml:mo>
</mml:mrow>
<mml:mo>=</mml:mo>
<mml:mi>O</mml:mi>
<mml:mrow>
<mml:mo>(</mml:mo>
<mml:mrow>
<mml:mstyle displaystyle="true">
<mml:munderover>
<mml:mo>&#x2211;</mml:mo>
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>=</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
<mml:mi>d</mml:mi>
</mml:munderover>
</mml:mstyle>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mrow>
<mml:mi>&#x3b1;</mml:mi>
<mml:mo>&#x2212;</mml:mo>
<mml:mn>1</mml:mn>
</mml:mrow>
</mml:msub>
<mml:mo>&#xb7;</mml:mo>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:msub>
<mml:mo>+</mml:mo>
<mml:msub>
<mml:mi>n</mml:mi>
<mml:mi>&#x3b1;</mml:mi>
</mml:msub>
</mml:mrow>
<mml:mo>)</mml:mo>
</mml:mrow>
</mml:mrow>
</mml:math>
</disp-formula>
<p>Where <italic>n<sub>&#x3b1;-1</sub>
</italic> and <italic>n<sub>&#x3b1;</sub>
</italic> represent the number of neurons in the (<italic>&#x3b1;</italic>-1) and <italic>&#x3b1;</italic> layers, respectively. Assuming <italic>&#x3b1;</italic> represents the number of layers and <italic>U</italic> represents the number of neurons per layer, the time complexity of the autoencoder model is as described above.</p>
</sec>
<sec id="s4_2">
<label>4.2</label>
<title>Interpretation of autoencoders for MRI generation</title>
<p>This set of experiments aims to elucidate the performance of autoencoders in generating MRI images. The generator proficiently learns the correlations and distribution patterns among various features from the training dataset, successfully capturing the essential characteristics of the images and reconstructing them based on these identified patterns. We visualize color feature mapping of some brain regions to reflect the feature correlations between the generated images and the original images.</p>
<p>We train the model using ASD patient data and provide color mapping of generated image slices and real sample slices to demonstrate the model&#x2019;s performance. Both generated and real sample slices are aligned in the same direction. The resultant grayscale images are then transformed into color-mapped images, where different colors denote the intensities of various features.</p>
<p>
<xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref> presents 2D brain MRI feature mapping slices, displaying partial Brodmann areas. In <xref ref-type="fig" rid="f3">
<bold>Figure&#xa0;3</bold>
</xref>, A represents the color mapping of the generated image slice, while B represents the color mapping of the real sample slice, with yellow representing higher intensity and blue representing lower intensity. The similar color distributions indicate a high degree of feature correspondence and an enhancement of information during the generation process. Notably, in parts of Brodmann areas 7, Brodmann areas 9, and Brodmann areas 18, there is an increase in feature intensity, signifying enhanced feature information, suggesting that the autoencoder primarily generates images by correctly extracting and processing features from these areas. This result is consistent with partial research findings of Bhaumik (<xref ref-type="bibr" rid="B2">2</xref>) and Donato (<xref ref-type="bibr" rid="B3">3</xref>). In summary, the autoencoder model is capable of accurately extracting and reflecting the features hidden within MRI images for image generation.</p>
<fig id="f3" position="float">
<label>Figure&#xa0;3</label>
<caption>
<p>Visualizations of feature mapping, including <bold>(A)</bold> generated samples and <bold>(B)</bold> real samples.</p>
</caption>
<graphic mimetype="image" mime-subtype="tiff" xlink:href="fpsyt-15-1395243-g003.tif"/>
</fig>
</sec>
<sec id="s4_3">
<label>4.3</label>
<title>The influence of optimizer on performance</title>
<p>This subsection compares different optimization methods in the autoencoder, including momentum SGD, RMSprop, Adagrad, and Adam. Corresponding image have been added to the <xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Materials</bold>
</xref>. In this study, RMSprop and Adam display superior performance, while Adagrad and momentum SGD are somewhat less effective. Adagrad and SGD have relatively fixed learning rates. Fixed learning rates may fail to effectively train models when there are significant differences in learning rate requirements across layers or parameters, particularly in complex optimization scenarios, leading to poor convergence or slow training speeds that impact performance. In contrast, RMSprop and Adam provide smooth and stable parameter updates, which help in avoiding drastic fluctuations in learning rates and thereby facilitate more effective training. RMSprop and Adam optimizers are generally more suitable for handling complex, non-convex optimization problems due to their ability to adapt to varying learning rate requirements, sparse gradients, and complex loss surfaces. SGD and Adagrad optimization methods did not succeed in achieving fast training of neural networks with complex structures as anticipated. The primary reason is likely the different adaptability of these optimizers in handling specific model gradient updates. In contrast, RMSprop and Adam, with their adaptive learning rate adjustment capabilities, typically perform more effectively in dealing with complex model structures and gradient characteristics.</p>
</sec>
</sec>
<sec id="s5">
<label>5</label>
<title>Conclusion and future work</title>
<p>This study proposes a method based on autoencoders for generating brain magnetic resonance imaging (MRI) images of patients with Autism Spectrum Disorder (ASD) and non-ASD individuals, and for discriminating ASD based on the generated images. Through experimental validation, we found that the proposed autoencoder model can effectively generate diverse MRI images of both ASD and non-ASD patients, capturing the structural differences in the brains of the patients. Additionally, our designed ASD classifier based on the generated images demonstrates good performance in the task of ASD discrimination, showing better robustness and generalizability compared to traditional classification methods based on original MRI images. These results provide new insights and approaches for the application of deep learning technologies in ASD research, with the potential to contribute to the automation of ASD diagnosis and research. However, generating MRI images of ASD and non-ASD patients using autoencoders faces certain limitations, such as restricted image sizes due to the computational complexity of MRI data, currently preventing the flexible generation of MRI images in various sizes. Future work may enhance model performance, expand datasets, and explore additional methods for medical image analysis based on automated image generation, thereby promoting early diagnosis and personalized treatment of ASD. Additionally, while this study assesses the clinical validity of the generated images using classifiers, future research should integrate pathophysiology to more deeply analyze the clinical application potential of the model, thus enhancing its supportive role in ASD treatment.</p>
</sec>
</body>
<back>
<sec id="s6" sec-type="data-availability">
<title>Data availability statement</title>
<p>The original contributions presented in the study are included in the article/<xref ref-type="supplementary-material" rid="SM1">
<bold>Supplementary Material</bold>
</xref>. Further inquiries can be directed to the corresponding authors.</p>
</sec>
<sec id="s7" sec-type="ethics-statement">
<title>Ethics statement</title>
<p>The studies involving humans were approved by International Neuroimaging Datasharing Initiative (INDI). The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants&#x2019; legal guardians/next of kin. Written informed consent was obtained from the individual(s), and minor(s)&#x2019; legal guardian/next of kin, for the publication of any potentially identifiable images or data included in this article.</p>
</sec>
<sec id="s8" sec-type="author-contributions">
<title>Author contributions</title>
<p>YS: Writing &#x2013; review &amp; editing, Writing &#x2013; original draft, Methodology, Investigation, Data curation. YLG: Writing &#x2013; review &amp; editing. YRG: Writing &#x2013; review &amp; editing. JT: Writing &#x2013; review &amp; editing, Methodology.</p>
</sec>
<sec id="s9" sec-type="funding-information">
<title>Funding</title>
<p>The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the Hubei Provincial Natural Science Foundation Project (2022CFB524); and in part by the Teaching and Research Reform Project of Hubei Provincial Department of Education (2022438); and in part by the Natural Science Foundation of Hubei Province (2021CFB316), the Preliminary support project of Hubei Social Science Foundation (21ZD137), and the Hundreds of Schools Unite with Hundreds of Counties-University Serving Rural Revitalization Science and Technology Support Action Plan (BXLBX0847); and in part by Research Project at the University Level of Hubei Polytechnic University (20xjz05Q).</p>
</sec>
<sec id="s10" sec-type="COI-statement">
<title>Conflict of interest</title>
<p>The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.</p>
</sec>
<sec id="s11" sec-type="disclaimer">
<title>Publisher&#x2019;s note</title>
<p>All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.</p>
</sec>
<sec id="s12" sec-type="supplementary-material">
<title>Supplementary material</title>
<p>The Supplementary Material for this article can be found online at: <ext-link ext-link-type="uri" xlink:href="https://www.frontiersin.org/articles/10.3389/fpsyt.2024.1395243/full#supplementary-material">https://www.frontiersin.org/articles/10.3389/fpsyt.2024.1395243/full#supplementary-material</ext-link>
</p>
<supplementary-material xlink:href="DataSheet1.pdf" id="SM1" mimetype="application/pdf"/>
</sec>
<ref-list>
<title>References</title>
<ref id="B1">
<label>1</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Katuwal</surname> <given-names>GJ</given-names>
</name>
<name>
<surname>Cahill</surname> <given-names>ND</given-names>
</name>
<name>
<surname>Baum</surname> <given-names>SA</given-names>
</name>
<name>
<surname>Michael</surname> <given-names>AM</given-names>
</name>
</person-group>. (<year>2015</year>). <article-title>The predictive power of structural mri in autism diagnosis</article-title>, in: <conf-name>2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</conf-name>, pp. <page-range>4270&#x2013;3</page-range>.</citation>
</ref>
<ref id="B2">
<label>2</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bhaumik</surname> <given-names>R</given-names>
</name>
<name>
<surname>Pradhan</surname> <given-names>A</given-names>
</name>
<name>
<surname>Das</surname> <given-names>S</given-names>
</name>
<name>
<surname>Bhaumik</surname> <given-names>DK</given-names>
</name>
</person-group>. <article-title>Predicting autism spectrum disorder using domain-adaptive cross-site evaluation</article-title>. <source>Neuroinformatics</source>. (<year>2018</year>) <volume>16</volume>:<fpage>197</fpage>&#x2013;<lpage>205</lpage>. doi:&#xa0;<pub-id pub-id-type="doi">10.1007/s12021-018-9366-0</pub-id>
</citation>
</ref>
<ref id="B3">
<label>3</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liloia</surname> <given-names>D</given-names>
</name>
<name>
<surname>Cauda</surname> <given-names>F</given-names>
</name>
<name>
<surname>Uddin</surname> <given-names>LQ</given-names>
</name>
<name>
<surname>Manuello</surname> <given-names>J</given-names>
</name>
<name>
<surname>Mancuso</surname> <given-names>L</given-names>
</name>
<name>
<surname>Keller</surname> <given-names>R</given-names>
</name>
<etal/>
</person-group>. <article-title>Revealing the selectivity of neuroanatomical alteration in autism spectrum disorder via reverse inference</article-title>. <source>Biol Psychiatry: Cogn Neurosci Neuroimaging</source>. (<year>2018</year>) <volume>8</volume>:<page-range>1075&#x2013;83</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.bpsc.2022.01.007</pub-id>
</citation>
</ref>
<ref id="B4">
<label>4</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname> <given-names>L</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>M</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>M</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>D</given-names>
</name>
</person-group>. <article-title>A survey on deep learning for neuroimaging-based brain disorder analysis</article-title>. <source>Front Neurosci</source>. (<year>2020</year>) <volume>14</volume>:<elocation-id>779</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fnins.2020.00779</pub-id>
</citation>
</ref>
<ref id="B5">
<label>5</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Zhao</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Ge</surname> <given-names>F</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>S</given-names>
</name>
<name>
<surname>Liu</surname> <given-names>T</given-names>
</name>
</person-group>. <article-title>3d deep convolutional neural network revealed the value of brain network overlap in differentiating autism spectrum disorder from healthy controls</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Frangi</surname> <given-names>AF</given-names>
</name>
<name>
<surname>Schnabel</surname> <given-names>JA</given-names>
</name>
<name>
<surname>Davatzikos</surname> <given-names>C</given-names>
</name>
<name>
<surname>Alberola-L&#xf3;pez</surname> <given-names>C</given-names>
</name>
<name>
<surname>Fichtinger</surname> <given-names>G</given-names>
</name>
</person-group>, editors. <source>Medical image computing and computer assisted intervention &#x2013; MICCAI 2018</source>. <publisher-loc>Cham</publisher-loc>: <publisher-name>Springer International Publishing</publisher-name> (<year>2018</year>). p. <page-range>172&#x2013;80</page-range>.</citation>
</ref>
<ref id="B6">
<label>6</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dolz</surname> <given-names>J</given-names>
</name>
<name>
<surname>Desrosiers</surname> <given-names>C</given-names>
</name>
<name>
<surname>Wang</surname> <given-names>L</given-names>
</name>
<name>
<surname>Yuan</surname> <given-names>J</given-names>
</name>
<name>
<surname>Shen</surname> <given-names>D</given-names>
</name>
<name>
<surname>Ben Ayed</surname> <given-names>I</given-names>
</name>
</person-group>. <article-title>Deep cnn ensembles and suggestive annotations for infant brain mri segmentation</article-title>. <source>Computerized Med Imaging Graphics</source>. (<year>2020</year>) <volume>79</volume>:<elocation-id>101660</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.compmedimag.2019.101660</pub-id>
</citation>
</ref>
<ref id="B7">
<label>7</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yin</surname> <given-names>W</given-names>
</name>
<name>
<surname>Mostafa</surname> <given-names>S</given-names>
</name>
</person-group>. <article-title>Wu Fx. Diagnosis of autism spectrum disorder based on functional brain networks with deep learning</article-title>. <source>J Comput Biol</source>. (<year>2021</year>) <volume>28</volume>:<page-range>146&#x2013;65</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1089/cmb.2020.0252</pub-id>
</citation>
</ref>
<ref id="B8">
<label>8</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Liu</surname> <given-names>R</given-names>
</name>
<name>
<surname>Lehman</surname> <given-names>J</given-names>
</name>
<name>
<surname>Molino</surname> <given-names>P</given-names>
</name>
<name>
<surname>Such</surname> <given-names>FP</given-names>
</name>
<name>
<surname>Frank</surname> <given-names>E</given-names>
</name>
<name>
<surname>Sergeev</surname> <given-names>A</given-names>
</name>
<etal/>
</person-group>. <article-title>An intriguing failing of convolutional neural networks and the coordconv solution</article-title>. <source>ArXiv</source>. (<year>2018</year>) <page-range>9628&#x2013;39</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.48550/arXiv.1807.03247</pub-id>
</citation>
</ref>
<ref id="B9">
<label>9</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yang</surname> <given-names>G</given-names>
</name>
<name>
<surname>Yu</surname> <given-names>S</given-names>
</name>
<name>
<surname>Dong</surname> <given-names>H</given-names>
</name>
<name>
<surname>Slabaugh</surname> <given-names>G</given-names>
</name>
<name>
<surname>Dragotti</surname> <given-names>PL</given-names>
</name>
<name>
<surname>Ye</surname> <given-names>X</given-names>
</name>
<etal/>
</person-group>. <article-title>Dagan: Deep de-aliasing generative adversarial networks for fast compressed sensing mri reconstruction</article-title>. <source>IEEE Trans Med Imaging</source>. (<year>2018</year>) <volume>37</volume>:<page-range>1310&#x2013;21</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TMI.2017.2785879</pub-id>
</citation>
</ref>
<ref id="B10">
<label>10</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Devika</surname> <given-names>K</given-names>
</name>
<name>
<surname>Mahapatra</surname> <given-names>D</given-names>
</name>
<name>
<surname>Subramanian</surname> <given-names>R</given-names>
</name>
<name>
<surname>Oruganti</surname> <given-names>VRM</given-names>
</name>
</person-group>. <article-title>Outlier-based autism detection using longitudinal structural mri</article-title>. <source>IEEE Access</source>. (<year>2022</year>) <volume>10</volume>:<page-range>27794&#x2013;808</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/ACCESS.2022.3157613</pub-id>
</citation>
</ref>
<ref id="B11">
<label>11</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Goodfellow</surname> <given-names>I</given-names>
</name>
<name>
<surname>Pouget-Abadie</surname> <given-names>J</given-names>
</name>
<name>
<surname>Mirza</surname> <given-names>M</given-names>
</name>
<name>
<surname>Xu</surname> <given-names>B</given-names>
</name>
<name>
<surname>Warde-Farley</surname> <given-names>D</given-names>
</name>
<name>
<surname>Ozair</surname> <given-names>S</given-names>
</name>
<etal/>
</person-group>. <article-title>Generative adversarial nets</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Ghahramani</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Welling</surname> <given-names>M</given-names>
</name>
<name>
<surname>Cortes</surname> <given-names>C</given-names>
</name>
<name>
<surname>Lawrence</surname> <given-names>N</given-names>
</name>
<name>
<surname>Weinberger</surname> <given-names>K</given-names>
</name>
</person-group>, editors. <source>Advances in neural information processing systems</source>. <publisher-name>Curran Associates, Inc</publisher-name> (<year>2014</year>).</citation>
</ref>
<ref id="B12">
<label>12</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mishro</surname> <given-names>PK</given-names>
</name>
<name>
<surname>Agrawal</surname> <given-names>S</given-names>
</name>
<name>
<surname>Panda</surname> <given-names>R</given-names>
</name>
<name>
<surname>Abraham</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>A survey on state-of-the-art denoising techniques for brain magnetic resonance images</article-title>. <source>IEEE Rev Biomed Eng</source>. (<year>2022</year>) <volume>15</volume>:<page-range>184&#x2013;99</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/RBME.2021.3055556</pub-id>
</citation>
</ref>
<ref id="B13">
<label>13</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Yeh</surname> <given-names>RA</given-names>
</name>
<name>
<surname>Chen</surname> <given-names>C</given-names>
</name>
<name>
<surname>Lim</surname> <given-names>TY</given-names>
</name>
<name>
<surname>Schwing</surname> <given-names>AG</given-names>
</name>
<name>
<surname>Hasegawa-Johnson</surname> <given-names>M</given-names>
</name>
<name>
<surname>Do</surname> <given-names>MN</given-names>
</name>
</person-group>. (<year>2017</year>). <article-title>Semantic image inpainting with deep generative models</article-title>, in: <conf-name>2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</conf-name>, pp. <page-range>6882&#x2013;90</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/CVPR.2017.728</pub-id>
</citation>
</ref>
<ref id="B14">
<label>14</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Creswell</surname> <given-names>A</given-names>
</name>
<name>
<surname>Bharath</surname> <given-names>AA</given-names>
</name>
</person-group>. <article-title>Denoising adversarial autoencoders</article-title>. <source>IEEE Trans Neural Networks Learn Syst</source>. (<year>2019</year>) <volume>30</volume>:<page-range>968&#x2013;84</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TNNLS.2018.2852738</pub-id>
</citation>
</ref>
<ref id="B15">
<label>15</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname> <given-names>P</given-names>
</name>
<name>
<surname>Pei</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Li</surname> <given-names>J</given-names>
</name>
</person-group>. <article-title>A comprehensive survey on design and application of autoencoder in deep learning</article-title>. <source>Appl Soft Computing</source>. (<year>2023</year>) <volume>138</volume>:<elocation-id>110176</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.asoc.2023.110176</pub-id>
</citation>
</ref>
<ref id="B16">
<label>16</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Lopez Pinaya</surname> <given-names>WH</given-names>
</name>
<name>
<surname>Vieira</surname> <given-names>S</given-names>
</name>
<name>
<surname>Garcia-Dias</surname> <given-names>R</given-names>
</name>
<name>
<surname>Mechelli</surname> <given-names>A</given-names>
</name>
</person-group>. <article-title>Chapter 11 - autoencoders</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>Mechelli</surname> <given-names>A</given-names>
</name>
<name>
<surname>Vieira</surname> <given-names>S</given-names>
</name>
</person-group>, editors. <source>Machine learning</source>. <publisher-name>Academic Press</publisher-name> (<year>2020</year>). p. <fpage>193</fpage>&#x2013;<lpage>208</lpage>.</citation>
</ref>
<ref id="B17">
<label>17</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pinaya</surname> <given-names>WHL</given-names>
</name>
<name>
<surname>Mechelli</surname> <given-names>A</given-names>
</name>
<name>
<surname>Sato</surname> <given-names>JR</given-names>
</name>
</person-group>. <article-title>Using deep autoencoders to identify abnormal brain structural patterns in neuropsychiatric disorders: A large-scale multi-sample study</article-title>. <source>Hum Brain Mapp</source>. (<year>2019</year>) <volume>40</volume>:<page-range>944&#x2013;54</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1002/hbm.24423</pub-id>
</citation>
</ref>
<ref id="B18">
<label>18</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Li</surname> <given-names>H</given-names>
</name>
<name>
<surname>Parikh</surname> <given-names>NA</given-names>
</name>
<name>
<surname>He</surname> <given-names>L</given-names>
</name>
</person-group>. <article-title>A novel transfer learning approach to enhance deep neural network classification of brain functional connectomes</article-title>. <source>Front Neurosci</source>. (<year>2018</year>) <volume>12</volume>:<elocation-id>491</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fnins.2018.00491</pub-id>
</citation>
</ref>
<ref id="B19">
<label>19</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Eslami</surname> <given-names>T</given-names>
</name>
<name>
<surname>Saeed</surname> <given-names>F</given-names>
</name>
</person-group>. <source>Auto-asd-network: A technique based on deep learning and support vector machines for diagnosing autism spectrum disorder using fmri data</source>. <publisher-loc>Niagara Falls, NY, USA</publisher-loc>: <publisher-name>Association for Computing Machinery</publisher-name> (<year>2019</year>) p. <page-range>646&#x2013;51</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1145/3307339.3343482</pub-id>
</citation>
</ref>
<ref id="B20">
<label>20</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Mostafa</surname> <given-names>S</given-names>
</name>
<name>
<surname>Wu</surname> <given-names>FX</given-names>
</name>
</person-group>. <article-title>Chapter 3 - diagnosis of autism spectrum disorder with convolutional autoencoder and structural mri images</article-title>. In: <person-group person-group-type="editor">
<name>
<surname>El-Baz</surname> <given-names>AS</given-names>
</name>
<name>
<surname>Suri</surname> <given-names>JS</given-names>
</name>
</person-group>, editors. <source>Neural engineering techniques for autism spectrum disorder</source>. <publisher-name>Academic Press</publisher-name> (<year>2021</year>). p. <fpage>23</fpage>&#x2013;<lpage>38</lpage>.</citation>
</ref>
<ref id="B21">
<label>21</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>He</surname> <given-names>K</given-names>
</name>
<name>
<surname>Zhang</surname> <given-names>X</given-names>
</name>
<name>
<surname>Ren</surname> <given-names>S</given-names>
</name>
<name>
<surname>Sun</surname> <given-names>J</given-names>
</name>
</person-group>. (<year>2015</year>). <article-title>Delving deep into rectifiers: Surpassing human-level performance on imagenet classification</article-title>, in: <conf-name>Proceedings of the IEEE International Conference on Computer Vision (ICCV)</conf-name>, .</citation>
</ref>
<ref id="B22">
<label>22</label>
<citation citation-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Rifai</surname> <given-names>S</given-names>
</name>
<name>
<surname>Vincent</surname> <given-names>P</given-names>
</name>
<name>
<surname>Muller</surname> <given-names>X</given-names>
</name>
<name>
<surname>Glorot</surname> <given-names>X</given-names>
</name>
<name>
<surname>Bengio</surname> <given-names>Y</given-names>
</name>
</person-group>. (<year>2011</year>). <article-title>Contractive auto-encoders: Explicit invariance during feature extraction</article-title>, in: <conf-name>Proceedings of the 28th International Conference on International Conference on Machine Learning</conf-name>, <publisher-loc>Bellevue, Washington, USA</publisher-loc> pp. <page-range>833&#x2013;40</page-range>. <publisher-name>Omnipress</publisher-name>.</citation>
</ref>
<ref id="B23">
<label>23</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hengjin</surname> <given-names>K</given-names>
</name>
<name>
<surname>Cang</surname> <given-names>C</given-names>
</name>
<name>
<surname>Fengqin</surname> <given-names>W</given-names>
</name>
<name>
<surname>Fang</surname> <given-names>H</given-names>
</name>
<name>
<surname>Jiawei</surname> <given-names>T</given-names>
</name>
<name>
<surname>Yuxin</surname> <given-names>S</given-names>
</name>
</person-group>. <article-title>Interpretation of frequency channel-based cnn on depression identification</article-title>. <source>Front Comput Neurosci</source>. (<year>2021</year>) <volume>15</volume>:<elocation-id>773147</elocation-id>. doi:&#xa0;<pub-id pub-id-type="doi">10.3389/fncom.2021.773147</pub-id>
</citation>
</ref>
<ref id="B24">
<label>24</label>
<citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname>Craddock</surname> <given-names>C</given-names>
</name>
<name>
<surname>Benhajali</surname> <given-names>Y</given-names>
</name>
<name>
<surname>Chu</surname> <given-names>C</given-names>
</name>
<name>
<surname>Chouinard</surname> <given-names>F</given-names>
</name>
<name>
<surname>Evans</surname> <given-names>A</given-names>
</name>
<name>
<surname>Jakab</surname> <given-names>A</given-names>
</name>
<etal/>
</person-group>. <article-title>The Neuro Bureau Preprocessing Initiative: open sharing of preprocessed neuroimaging data and derivatives</article-title>. In <source>Neuroinformatics 2013</source>, <publisher-loc>Stockholm, Sweden</publisher-loc>. (<year>2013</year>). doi:&#xa0;<pub-id pub-id-type="doi">10.3389/conf.fninf.2013.09.00041</pub-id>.</citation>
</ref>
<ref id="B25">
<label>25</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Wang</surname> <given-names>Z</given-names>
</name>
<name>
<surname>Bovik</surname> <given-names>A</given-names>
</name>
<name>
<surname>Sheikh</surname> <given-names>H</given-names>
</name>
<name>
<surname>Simoncelli</surname> <given-names>E</given-names>
</name>
</person-group>. <article-title>Image quality assessment: from error visibility to structural similarity</article-title>. <source>IEEE Trans Image Process</source>. (<year>2004</year>) <volume>13</volume>:<page-range>600&#x2013;12</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1109/TIP.2003.819861</pub-id>
</citation>
</ref>
<ref id="B26">
<label>26</label>
<citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname>Reuter</surname> <given-names>M</given-names>
</name>
<name>
<surname>Schmansky</surname> <given-names>NJ</given-names>
</name>
<name>
<surname>Rosas</surname> <given-names>HD</given-names>
</name>
<name>
<surname>Fischl</surname> <given-names>B</given-names>
</name>
</person-group>. <article-title>Within-subject template estimation for unbiased longitudinal image analysis</article-title>. <source>NeuroImage</source>. (<year>2012</year>) <volume>61</volume>:<page-range>1402&#x2013;18</page-range>. doi:&#xa0;<pub-id pub-id-type="doi">10.1016/j.neuroimage.2012.02.084</pub-id>
</citation>
</ref>
</ref-list>
</back>
</article>